
In this notebook, you’ll see how Portkey streamlines LLM evaluation for the Top 10 LMSYS Models, giving you valuable insights into cost, performance, and accuracy metrics. Let’s dive in!
Video Guide
The notebook comes with a video guide that you can follow alongSetting up Portkey
To get started, install the necessary packages:Defining the Top 10 LMSYS Models
Let’s define the list of Top 10 LMSYS models and their corresponding providers.Python
Add Providers to Model Catalog
ALL the providers above are integrated with Portkey - add them to Model Catalog to get provider slugs for streamlined API key management.| Provider | Link to get API Key | Payment Mode |
|---|---|---|
| openai | https://platform.openai.com/ | Wallet Top Up |
| anthropic | https://console.anthropic.com/ | Wallet Top Up |
| https://aistudio.google.com/ | Free to Use | |
| cohere | https://dashboard.cohere.com/ | Free Credits |
| together-ai | https://api.together.ai/ | Free Credits |
| reka-ai | https://platform.reka.ai/ | Wallet Top Up |
| zhipu | https://open.bigmodel.cn/ | Free to Use |
Python
Running the Models with Portkey
Now, let’s create a function to run the Top 10 LMSYS models using OpenAI SDK with Portkey Gateway:OpenAI Python
Comparing Model Outputs
To display the model outputs in a tabular format for easy comparison, we define the print_model_outputs function:Python
Example: Evaluating LLMs for a Specific Task
Let’s run the notebook with a specific prompt to showcase the differences in responses from various LLMs: On Portkey, you will be able to see the logs for all models:
Python

