Skip to main content

You will need Portkey and Together AI API keys to get started

pip install -qU portkey-ai openai

With OpenAI Client

OpenAI Python
from openai import OpenAI
from portkey_ai import PORTKEY_GATEWAY_URL, createHeaders

openai = OpenAI(
    api_key='TOGETHER_API_KEY',  # Grab from https://api.together.xyz/
    base_url=PORTKEY_GATEWAY_URL,
    default_headers=createHeaders(
        provider="together-ai",
        api_key='PORTKEY_API_KEY'  # Grab from https://app.portkey.ai/
    )
)

response = openai.chat.completions.create(
    model="meta-llama/Llama-3-8b-chat-hf",
    messages=[{"role": "user", "content": "What's a fractal?"}],
    max_tokens=500
)

print(response.choices[0].message.content)

With Portkey Client

Add your Together API key in Model Catalog and access models using your provider slug
Python
from portkey_ai import Portkey

portkey = Portkey(api_key="PORTKEY_API_KEY")

response = portkey.chat.completions.create(
    model="@together-prod/meta-llama/Llama-3-8b-chat-hf",  # @provider-slug/model
    messages=[{"role": "user", "content": "Who are you?"}],
    max_tokens=500
)

print(response.choices[0].message.content)

Monitoring your Requests

Using Portkey you can monitor your Llama 3 requests and track tokens, cost, latency, and more.