|
33 | 33 | "github_token = os.environ[\"GITHUB_TOKEN\"]\n", |
34 | 34 | "\n", |
35 | 35 | "# We can use some defaults for the other two variables\n", |
36 | | - "endpoint = \"https://models.inference.ai.azure.com\"\n", |
37 | | - "model_name = \"mistral-large\"" |
| 36 | + "endpoint = \"https://models.github.ai/inference\"\n", |
| 37 | + "model_name = \"mistral-ai/mistral-large\"" |
38 | 38 | ] |
39 | 39 | }, |
40 | 40 | { |
|
100 | 100 | "from mistralai.models.chat_completion import ChatMessage\n", |
101 | 101 | "\n", |
102 | 102 | "\n", |
103 | | - "def run_mistral(user_message, model=\"mistral-small\"):\n", |
| 103 | + "def run_mistral(user_message, model=\"mistral-ai/mistral-small\"):\n", |
104 | 104 | " client = MistralClient(api_key=github_token, endpoint=endpoint)\n", |
105 | 105 | " messages = [ChatMessage(role=\"user\", content=user_message)]\n", |
106 | 106 | " chat_response = client.chat(\n", |
|
225 | 225 | "from mistralai.models.chat_completion import ChatMessage\n", |
226 | 226 | "\n", |
227 | 227 | "\n", |
228 | | - "def run_mistral(user_message, model=\"mistral-small\"):\n", |
| 228 | + "def run_mistral(user_message, model=\"mistral-ai/mistral-small\"):\n", |
229 | 229 | " client = MistralClient(api_key=github_token, endpoint=endpoint)\n", |
230 | 230 | " messages = [ChatMessage(role=\"user\", content=user_message)]\n", |
231 | 231 | " chat_response = client.chat(model=model, messages=messages)\n", |
|
379 | 379 | "from mistralai.models.chat_completion import ChatMessage\n", |
380 | 380 | "\n", |
381 | 381 | "\n", |
382 | | - "def run_mistral(user_message, model=\"mistral-small\", is_json=False):\n", |
| 382 | + "def run_mistral(user_message, model=\"mistral-ai/mistral-small\", is_json=False):\n", |
383 | 383 | " client = MistralClient(api_key=github_token, endpoint=endpoint)\n", |
384 | 384 | " messages = [ChatMessage(role=\"user\", content=user_message)]\n", |
385 | 385 | "\n", |
|
502 | 502 | " scoring_prompt.format(\n", |
503 | 503 | " news=news, summary=summary, metric=i[\"metric\"], rubrics=i[\"rubrics\"]\n", |
504 | 504 | " ),\n", |
505 | | - " model=\"mistral-small\",\n", |
| 505 | + " model=\"mistral-ai/mistral-small\",\n", |
506 | 506 | " is_json=True,\n", |
507 | 507 | " )\n", |
508 | 508 | " print(eval_output)" |
|
511 | 511 | ], |
512 | 512 | "metadata": { |
513 | 513 | "kernelspec": { |
514 | | - "display_name": "Python 3 (ipykernel)", |
| 514 | + "display_name": "Python 3", |
515 | 515 | "language": "python", |
516 | 516 | "name": "python3" |
517 | 517 | }, |
|
0 commit comments