|
10 | 10 | { |
11 | 11 | "cell_type": "markdown", |
12 | 12 | "metadata": {}, |
13 | | - "source": [ |
14 | | - "### Setup" |
15 | | - ] |
| 13 | + "source": "### Setup" |
16 | 14 | }, |
17 | 15 | { |
18 | 16 | "cell_type": "code", |
19 | | - "execution_count": null, |
20 | 17 | "metadata": {}, |
21 | | - "outputs": [], |
22 | 18 | "source": [ |
23 | 19 | "# Import necessary packages\n", |
24 | 20 | "import os\n", |
|
27 | 23 | "from predictionguard import PredictionGuard\n", |
28 | 24 | "\n", |
29 | 25 | "\n", |
30 | | - "# Set PG API Key\n", |
| 26 | + "\n", |
| 27 | + "# Set your Prediction Guard token and url as an environmental variable.\n", |
31 | 28 | "os.environ[\"PREDICTIONGUARD_API_KEY\"] = \"<api key>\"\n", |
| 29 | + "os.environ[\"PREDICTIONGUARD_URL\"] = \"<url>\"\n", |
32 | 30 | "\n", |
33 | | - "# Initialize PG client\n", |
34 | | - "client = PredictionGuard()" |
35 | | - ] |
| 31 | + "# Or set your Prediction Guard token and url when initializing the PredictionGuard class.\n", |
| 32 | + "client = PredictionGuard(\n", |
| 33 | + " api_key=\"Bg-98uZ5mJPEwFQJE8UN9MuG6KG2SK9gJILyw3nYPFA\",\n", |
| 34 | + " url=\"<url>\"\n", |
| 35 | + ")" |
| 36 | + ], |
| 37 | + "outputs": [], |
| 38 | + "execution_count": null |
36 | 39 | }, |
37 | 40 | { |
38 | 41 | "cell_type": "markdown", |
|
43 | 46 | }, |
44 | 47 | { |
45 | 48 | "cell_type": "code", |
46 | | - "execution_count": null, |
47 | 49 | "metadata": {}, |
48 | | - "outputs": [], |
49 | 50 | "source": [ |
50 | 51 | "response = client.completions.create(\n", |
51 | | - " model=\"Hermes-2-Pro-Mistral-7B\",\n", |
52 | | - " messages=\"Tell me a joke.\",\n", |
53 | | - " max_tokens=500\n", |
| 52 | + " model=\"Hermes-3-Llama-3.1-8B\",\n", |
| 53 | + " prompt=\"Tell me a joke.\",\n", |
| 54 | + " max_tokens=100\n", |
54 | 55 | ")\n", |
55 | 56 | "\n", |
56 | 57 | "print(json.dumps(\n", |
|
59 | 60 | " indent=4,\n", |
60 | 61 | " separators=(',', ': ')\n", |
61 | 62 | "))" |
62 | | - ] |
| 63 | + ], |
| 64 | + "outputs": [], |
| 65 | + "execution_count": null |
| 66 | + }, |
| 67 | + { |
| 68 | + "metadata": {}, |
| 69 | + "cell_type": "markdown", |
| 70 | + "source": "### Streaming Completions" |
| 71 | + }, |
| 72 | + { |
| 73 | + "metadata": {}, |
| 74 | + "cell_type": "code", |
| 75 | + "source": [ |
| 76 | + "for res in client.completions.create(\n", |
| 77 | + " model=\"Hermes-3-Llama-3.1-8B\",\n", |
| 78 | + " prompt=\"Tell me a joke.\",\n", |
| 79 | + " max_tokens=100,\n", |
| 80 | + " stream=True\n", |
| 81 | + "):\n", |
| 82 | + " # Use 'end' parameter in print function to avoid new lines.\n", |
| 83 | + " print(res[\"data\"][\"choices\"][0][\"text\"], end=\"\")" |
| 84 | + ], |
| 85 | + "outputs": [], |
| 86 | + "execution_count": null |
63 | 87 | }, |
64 | 88 | { |
65 | 89 | "cell_type": "markdown", |
|
70 | 94 | }, |
71 | 95 | { |
72 | 96 | "cell_type": "code", |
73 | | - "execution_count": null, |
74 | 97 | "metadata": {}, |
75 | | - "outputs": [], |
76 | 98 | "source": [ |
77 | 99 | "model_list = client.completions.list_models()\n", |
78 | 100 | "\n", |
79 | 101 | "print(model_list)" |
80 | | - ] |
| 102 | + ], |
| 103 | + "outputs": [], |
| 104 | + "execution_count": null |
81 | 105 | } |
82 | 106 | ], |
83 | 107 | "metadata": { |
|
0 commit comments