Skip to main content
Portkey provides a robust and secure gateway to integrate various Large Language Models (LLMs) into applications, including Together AI’s hosted models. With Portkey, take advantage of features like fast AI gateway access, observability, prompt management, and more, while securely managing API keys through Model Catalog.

Quick Start

Get Together AI working in 3 steps:
from portkey_ai import Portkey

# 1. Install: pip install portkey-ai
# 2. Add @together-ai provider in model catalog
# 3. Use it:

portkey = Portkey(api_key="PORTKEY_API_KEY")

response = portkey.chat.completions.create(
    model="@together-ai/meta-llama/Meta-Llama-3.1-8B-Instruct-Turbo",
    messages=[{"role": "user", "content": "Say this is a test"}]
)

print(response.choices[0].message.content)
Tip: You can also set provider="@together-ai" in Portkey() and use just model="meta-llama/Meta-Llama-3.1-8B-Instruct-Turbo" in the request.

Add Provider in Model Catalog

  1. Go to Model Catalog → Add Provider
  2. Select Together AI
  3. Choose existing credentials or create new by entering your Together AI API key
  4. Name your provider (e.g., together-ai-prod)

Complete Setup Guide →

See all setup options, code examples, and detailed instructions

Reasoning / Thinking Support

Together AI supports reasoning models that expose their internal chain of thought. Use the reasoning_effort parameter to control reasoning behavior, and set strict_open_ai_compliance=False to receive the thinking content in content_blocks.
from portkey_ai import Portkey

portkey = Portkey(
    api_key="PORTKEY_API_KEY",
    strict_open_ai_compliance=False
)

response = portkey.chat.completions.create(
    model="@together-ai/deepseek-ai/DeepSeek-R1",
    messages=[{"role": "user", "content": "Solve step by step: What is 23 * 47?"}],
    reasoning_effort="medium"
)

# Access thinking content from content_blocks
for block in response.choices[0].message.content_blocks:
    if block.get("type") == "thinking":
        print("Thinking:", block["thinking"])
    elif block.get("type") == "text":
        print("Response:", block["text"])
The reasoning response includes content_blocks with both the model’s thinking process and the final answer. Streaming is also supported and returns reasoning chunks in the content_blocks field of the stream delta.

Thinking Mode Documentation

Learn more about thinking/reasoning support across providers

Managing Together AI Prompts

Manage all prompt templates to Together AI in the Prompt Library. All current Together AI models are supported, and you can easily test different prompts. Use the portkey.prompts.completions.create interface to use the prompt in an application.

Next Steps

For complete SDK documentation:

SDK Reference

Complete Portkey SDK documentation
Last modified on February 14, 2026