Upload notebook.ipynb (#3)
Browse files- Upload notebook.ipynb (c4c5dfc9b5df75035aef17b062e2ddbf486e7ded)
- notebook.ipynb +40 -0
notebook.ipynb
CHANGED
|
@@ -89,6 +89,46 @@
|
|
| 89 |
},
|
| 90 |
"execution_count": null,
|
| 91 |
"outputs": []
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 92 |
}
|
| 93 |
]
|
| 94 |
}
|
|
|
|
| 89 |
},
|
| 90 |
"execution_count": null,
|
| 91 |
"outputs": []
|
| 92 |
+
},
|
| 93 |
+
{
|
| 94 |
+
"cell_type": "markdown",
|
| 95 |
+
"source": [
|
| 96 |
+
"## Too big to run in Colab?\n",
|
| 97 |
+
"\n",
|
| 98 |
+
"Try using Inference Providers for serverless usage of these models"
|
| 99 |
+
],
|
| 100 |
+
"metadata": {
|
| 101 |
+
"id": "-r9Z_OnzRSFn"
|
| 102 |
+
}
|
| 103 |
+
},
|
| 104 |
+
{
|
| 105 |
+
"cell_type": "code",
|
| 106 |
+
"source": [
|
| 107 |
+
"import os\n",
|
| 108 |
+
"from huggingface_hub import InferenceClient\n",
|
| 109 |
+
"\n",
|
| 110 |
+
"client = InferenceClient(\n",
|
| 111 |
+
" provider=\"auto\",\n",
|
| 112 |
+
" api_key=os.environ[\"HF_TOKEN\"],\n",
|
| 113 |
+
")\n",
|
| 114 |
+
"\n",
|
| 115 |
+
"completion = client.chat.completions.create(\n",
|
| 116 |
+
" model=\"Qwen/Qwen3-4B\",\n",
|
| 117 |
+
" messages=[\n",
|
| 118 |
+
" {\n",
|
| 119 |
+
" \"role\": \"user\",\n",
|
| 120 |
+
" \"content\": \"What is the capital of France?\"\n",
|
| 121 |
+
" }\n",
|
| 122 |
+
" ],\n",
|
| 123 |
+
")\n",
|
| 124 |
+
"\n",
|
| 125 |
+
"print(completion.choices[0].message)"
|
| 126 |
+
],
|
| 127 |
+
"metadata": {
|
| 128 |
+
"id": "ZtpuYegeRYPg"
|
| 129 |
+
},
|
| 130 |
+
"execution_count": null,
|
| 131 |
+
"outputs": []
|
| 132 |
}
|
| 133 |
]
|
| 134 |
}
|