Skip to main content

How to Connect Claude API in Chatbox & Cherry Studio (2026 Tested)

A step-by-step guide to configuring Claude API in both Chatbox and Cherry Studio via ClaudeAPI.com — accessible from anywhere, no VPN required. Includes real benchmark data across 5 models.

ToolsTool IntegrationsTechnical TutorialCherry StudioChatboxEst. read8min
2026.04.03 published
How to Connect Claude API in Chatbox & Cherry Studio (2026 Tested)

How to Connect Claude API in Chatbox & Cherry Studio (2026 Tested)

Not everyone wants to write API code — sometimes you just want a solid desktop client where you can chat with Claude, draft documents, refactor code, and analyze files.

Chatbox and Cherry Studio are two of the most popular AI desktop clients that both support custom API endpoints. This guide walks you through connecting them to Claude’s full model lineup via ClaudeAPI.com — accessible from anywhere, no VPN required, and ready in under 3 minutes.

📌 All data in this article was tested in April 2026 using the OpenAI-compatible endpoint at ClaudeAPI.com — direct connection, no proxy needed.


Step 0: Get Your Claude API Key

Before configuring any client, you’ll need a working API key.

Get one at ClaudeAPI.com in 3 steps:

1.Create an account — Sign up with your email at ClaudeAPI.com

2.Top up your balance — Multiple payment methods supported; new accounts receive a $5 free credit to get started

3.Generate your key — Go to Dashboard → Token Management → Add New Token → Copy & save it

💡 For a detailed walkthrough, see:Claude API Pricing & Model Selection Guide (2026)

Once you have your key, note down these connection details:

Purpose Base URL Notes
OpenAI-compatible / Anthropic https://claudeapi.com Works for both Chatbox and Cherry Studio

💡 Both Chatbox and Cherry Studio connect via the OpenAI-compatible format. Just set the Base URL to https://claudeapi.com — direct access, no proxy required.


Claude Model Benchmarks (All 5 Models Verified)

Before diving into client setup, we ran real tests across all available Claude models using the OpenAI-compatible endpoint — so you know exactly what speed and token usage to expect.

Test Conditions

  • Test Date:2026
  • Network:Direct connection, no VPN
  • API Format:OpenAI-compatible (claudeapi.com)
  • SDK:Python OpenAI SDK
  • Test Prompt"Briefly introduce the advantages of Python"(max_tokens=200)

Speed & Token Usage — 5 Models Benchmarked

Model Model ID Response Time Input Tokens Output Tokens Status
Claude Sonnet 4.6 claude-sonnet-4-6 8.35s 22 200 ✅ Pass
Claude Opus 4.6 claude-opus-4-6 7.01s 22 200 ✅ Pass
Claude Haiku 4.5 claude-haiku-4-5-20251001 4.03s 22 180 ✅ Pass过
Claude Sonnet 4.5 claude-sonnet-4-5-20250929 4.63s 22 148 ✅ Pass
Claude Sonnet 4 claude-sonnet-4-20250514 6.85s 22 200 ✅ Pass过

💡 Haiku 4.5 is the fastest (4.03s) — ideal for simple tasks and high-volume batch processing. Sonnet 4.6 offers the best price-to-performance ratio for everyday use. For detailed pricing, see the Claude API Pricing & Model Selection Guide (2026)

Streaming Output Test

Tested with stream=True

Metric Result
Model claude-sonnet-4-6
Prompt "Describe yourself in one sentence."
Time to First Token 5.38s
Total Response Time 7.16s
Chunks Received 13
Response “I’m Claude, an AI assistant made by Anthropic, designed to be helpful, harmless, and honest.”

Streaming works perfectly — in Chatbox and Cherry Studio, you’ll see the response appear word by word in real time.

Multi-Turn Conversation Test

Tested a 3-turn conversation with a System Prompt:

Turn User Message Claude’s Response TIime
1 “What is Python’s GIL? One sentence.” “The GIL (Global Interpreter Lock) is a mutex in CPython that ensures only one thread executes Python bytecode at a time, preventing memory safety issues from concurrent thread access.” 11.43s
2 “How do you work around it? One sentence.。” “Use multiprocessing instead of threading, or use C extensions / asyncio for I/O-bound scenarios.” 3.08s
3 “Which is better for CPU-bound tasks — multiprocessing or threading?” “multiprocessing — each process has its own GIL, enabling true parallel execution across multiple cores.” 7.90s

Total time for 3 turns: 22.41s. Context was maintained throughout — responses were accurate and consistent.

Long-Form Generation Test

Prompt Result
Prompt “Write a 300-word short essay: Why should developers learn AI?”
max_tokens 1024
Response Time 42.22s
Input Tokens 35
Output Tokens 476
Generated Length 478 characters

Connecting Chatbox to Claude API

What is Chatbox?

[ChatBox]is a free, open-source AI desktop client available on Windows, macOS, Linux, iOS, Android, and Web. Key features include:

  • 🗨️ Multi-model chat (GPT, Claude, Gemini, and more)
  • 📄 Document & image conversations (PDF, Word, code files)
  • 💻 Code highlighting and preview
  • 📝 Markdown / LaTeX rendering
  • 🔒 Local data storage — your conversations stay on your device
  • 🔌 MCP service integration

Download & Install Chatbox

Head to chatboxai.app and grab the installer for your platform:

Platform Version
Windows 64-bit installer
macOS Intel / Apple Silicon
Linux 64-bit / ARM64
iOS / Android App Store / Google Play
Web web.chatboxai.app

Configuring Claude API in Chatbox — Step by Step

Step 1: Open Settings

Launch Chatbox → click the ⚙️ Settingsicon in the bottom-left corner.

Step 2: Select a Model Provider

In the Settings panel, find “AI Model Provider” and choose:

Add Model Provider (OpenAI API Compatible)

Step 3: Fill In Your API Configuration

Field Value Notes
API Host / Base URL https://claudeapi.com
API Key Your sk- Token Get it from your ClaudeAPI.com dashboard
Model claude-sonnet-4-6 Recommended for everyday use

Step 4: Add Your Model

Click “+ Add” to add at least one model.

You can optionally check the capability flags for that model (vision, function calling, etc.) — if left unchecked, it defaults to text-only mode.

Step 5: Save and Test

Click Save, go back to the chat window, and send a message.

Try this prompt:

Hi, introduce yourself in one sentence.

Expected response:

“Hi there! Great to meet you! What can I help you with today? 😊”

— claudeapi (claude-sonnet-4-6) · 80 tokens (actual test result)

Switching Claude Models in Chatbox

You can switch models anytime from the chat interface or settings — just update the model name:

Model Model ID Avg. Response Time
Claude Sonnet 4.6(⭐ Recommended ) claude-sonnet-4-6 ~8s
Claude Opus 4.6(🧠 Best reasoning ) claude-opus-4-6 ~7s
Claude Haiku 4.5(🧠 Best reasoning) claude-haiku-4-5-20251001 ~4s
Claude Sonnet 4.5 claude-sonnet-4-5-20250929 ~5s
Claude Sonnet 4 claude-sonnet-4-20250514 ~7s

💡 Quick model-picking guide:

  • Everyday chat → Sonnet 4.6
  • Deep analysis & complex reasoning → Opus 4.6
  • Simple tasks where speed matters → Haiku 4.5

Connecting Cherry Studio to Claude API

What is Cherry Studio?

[Cherry Studio] is an open-source AI desktop client for Windows, macOS, and Linux. It stands out with:

  • 🤖 300+ AI models supported, compatible with 50+ AI providers

  • 🧩 300+ built-in AI assistants ready to use

  • 📚 Knowledge base, AI image generation, and translation tools

  • 🖥️ Local model support via Ollama and LM Studio

  • 🆓 Completely free and open-source

Download Cherry Studio

Grab the installer for your platform at cherry-ai.com/download .

Configuring Claude API in Cherry Studio — Step by Step

Step 1: Open Settings

Launch Cherry Studio → click the ⚙️ Settings icon in the left sidebar.

Step 2: Add a Custom Provider

In the Settings panel, navigate to “Model Services” or “Provider Management”:

Click “Add Provider” Select OpenAI or Anthropic as the provider type Set the provider name to something like ClaudeAPI **Step 3: Fill In Your API Configuration

Field Value Notes
API Address / Base URL https://claudeapi.com
API Key Your sk- Token Get it from your (https://claudeapi.com) dashboard

Step 4: Add Your Models

Once the provider is configured, manually add the models you want to use:

Under the ClaudeAPI provider, click “Add Model” and enter the model IDs one by one:

  • claude-opus-4-6
  • claude-haiku-4-5-20251001
  • claude-sonnet-4-5-20250929
  • claude-sonnet-4-20250514

Step 5: Start Chatting

  1. Go back to the chat interface

2.Switch to the ClaudeAPI provider in the model selector

3.Select claude-sonnet-4-6

4.Send a test message

💡 Pro tip: Cherry Studio lets you switch between models mid-conversation — great for comparing response quality on the fly.

Want to see more advanced Cherry Studio workflows? Check out our guide: How to Use Claude API in Cherry Studio for SEO — A Complete Beginner’s Guide


Chatbox vs. Cherry Studio: Which One Should You Pick?

Feature ChatBox Cherry Studio
Open-source & free ✅ Yes ✅ Yes
Platform support Win/Mac/Linux/iOS/Android/Web Win/Mac/Linux
Mobile app ✅ iOS & Android ❌ Desktop only
Web version ✅ web.chatboxai.app ❌ None
Built-in assistants A few 300+ presets
Knowledge base ✅ Desktop ✅ Yes
文Document chat ✅ PDF/Word/Images ✅ Yes
MCP integration ✅ Yes ✅ Yes
UI language English & Chinese Chinese-first
Best for Cross-platform users who need mobile Power users who want rich presets

💡 TL;DR: Need it on your phone too? → Go with Chatbox. Want 300+ built-in AI assistants and a feature-packed desktop experience? → Go with Cherry Studio. Both are free — there’s no harm in trying both.


Why ClaudeAPI.com Instead of Anthropic Direct?

Anthropic Official ClaudeAPI.com
Accessibility ❌ Unreliable in many regions ✅ Direct access from anywhere, no VPN needed
Payment Credit card (international only) ✅ Stripes and more
API format Anthropic format only ✅ OpenAI-compatible + Anthropic format
Billing currency USD ✅ USD with transparent pricing
Minimum top-up ~$5 ✅ Free trial credits to get started

ClaudeAPI.com is the fastest way to get Claude API access without the usual friction — no geo-restrictions, no international card required.

Want to go deeper? Check out our guide:考 How to Use Claude API for SEO in 2026: 7 Workflows + Ready-to-Use Prompts


Pro Tips

Customizing the System Prompt

Set a custom system prompt in your conversation settings to give Claude a specific role. Here’s one we use in production:

You are a senior Python engineer with deep expertise in code review 
and system architecture. Keep your answers concise.
You are a senior Python engineer with deep expertise in code review 
and system architecture. Keep your answers concise.

Real-world result: Across 3 rounds of back-and-forth, responses were sharp, accurate, and to the point — total time: 22.41s.

Keeping Costs Under Control: Model Selection Guide

Based on real usage data, here’s what we recommend for each use case:

Use Case Recommended Model Est. Cost per Request Notes
Everyday chat claude-sonnet-4-6 ~$0.001-0.01 Short input & output
Code generation claude-sonnet-4-6 ~$0.01-0.05 Medium output volume
Long-form writing claude-sonnet-4-6 ~$0.05-0.1 Tested: 476 tokens out, ~35 tokens in
Deep analysis claude-opus-4-6 ~$0.1-0.5 Higher per-token cost, best quality
Simple translation / classification claude-haiku-4-5-20251001 ~$0.001 Cheapest and fastest

💡 Real benchmark: We used Sonnet 4.6 to generate a 478-word article (“Why Every Developer Should Learn AI”) — 35 tokens in, 476 tokens out, completed in 42.22s. Total cost at ClaudeAPI.com pricing: ~$0.003.

Document Chat

Both Chatbox and Cherry Studio support uploading documents and chatting with Claude directly:

  • 📄 Upload a PDF → Ask Claude to summarize key points or answer questions

  • 💻 Upload a code file → Let Claude do a code review

  • 🖼️ Upload a screenshot → Have Claude analyze UI issues or error messages

  • 💡 Document chat leverages Claude’s multimodal capabilities. For best results, use Sonnet 4.6 or Opus 4.6.


FAQ

Messages not going through after setup / getting errors?

Run through this checklist:

  • [ ] Is the Base URL set to https://claudeapi.com
  • [ ] Did you copy the full API Key (starts with sk-)?
  • [ ] Is the model name spelled correctly?
  • [ ] Does your ClaudeAPI.com account have sufficient balance?

Getting “model not found” or “invalid model”?

This is almost always a model name typo. Model IDs are case-sensitive. Here are the verified, working model IDs:

claude-opus-4-6
claude-sonnet-4-6
claude-haiku-4-5-20251001
claude-sonnet-4-5-20250929
claude-sonnet-4-20250514
claude-opus-4-6
claude-sonnet-4-6
claude-haiku-4-5-20251001
claude-sonnet-4-5-20250929
claude-sonnet-4-20250514

Claude API responses feel slow — is that normal?

Here are real latency benchmarks for a simple prompt with ~200 tokens of output:

Model Response Time
Haiku 4.5 4.03s(⚡fastest)
Sonnet 4.5 4.63s
Sonnet 4 6.85s
Opus 4.6 7.01s
Sonnet 4.6 8.35s

If you’re seeing significantly slower times, check your network connection. Long-form generation (500+ tokens) typically takes 30–50s — that’s expected behavior.

Can’t see any models in Cherry Studio?

After adding a provider, you need to add models manually. In the provider settings, click “Add Model” and type in the model ID (e.g., claude-sonnet-4-6).

Does it support streaming output?

Yes. Streaming has been tested and works correctly:

  • Time to first token: 5.38s (Sonnet 4.6)

  • Response streams in progressively, delivered across 13 chunks

Both Chatbox and Cherry Studio have streaming enabled by default.

Is it complicated to connect Chatbox / Cherry Studio to Claude API?

Not at all — you can be up and running in under 3 minutes. ClaudeAPI.com provides a globally accessible endpoint that works from anywhere without a VPN. Just plug in the base URL and your API key, and you’re good to go.


Quick Test: Verify Your API Key with Code

Want to confirm your key works before setting up a GUI client? Run this Python snippet:

from openai import OpenAI

client = OpenAI(
    api_key="sk-your-Key",
    base_url="https://claudeapi.com"
)

# Basic test
response = client.chat.completions.create(
    model="claude-sonnet-4-6",
    messages=[{"role": "user", "content": "Hello, introduce yourself in one sentence."}],
    max_tokens=100
)
print(response.choices[0].message.content)

# Streaming test
stream = client.chat.completions.create(
    model="claude-sonnet-4-6",
    messages=[{"role": "user", "content": "What are the strengths of Python?"}],
    max_tokens=200,
    stream=True
)
for chunk in stream:
    if chunk.choices and chunk.choices[0].delta.content:
        print(chunk.choices[0].delta.content, end="", flush=True)
from openai import OpenAI

client = OpenAI(
    api_key="sk-your-Key",
    base_url="https://claudeapi.com"
)

# Basic test
response = client.chat.completions.create(
    model="claude-sonnet-4-6",
    messages=[{"role": "user", "content": "Hello, introduce yourself in one sentence."}],
    max_tokens=100
)
print(response.choices[0].message.content)

# Streaming test
stream = client.chat.completions.create(
    model="claude-sonnet-4-6",
    messages=[{"role": "user", "content": "What are the strengths of Python?"}],
    max_tokens=200,
    stream=True
)
for chunk in stream:
    if chunk.choices and chunk.choices[0].delta.content:
        print(chunk.choices[0].delta.content, end="", flush=True)

Wrap-Up: Claude API in 3 Minutes

No coding required. Works from anywhere. Here’s all it takes to get Claude running in a desktop client:

Step What to Do
1. Get your API key Go to claudeapi.com → Sign up → Top up → Create a token → Copy your key
Download a client ChatBox or Cherry Studio
3. Configure the API Set provider to OpenAI-compatible or Anthropic, set Base URL to https://claudeapi.com, paste your key
4. Pick a model Enter claude-sonnet-4-6 (recommended) and start chatting

Quick-Reference Config

Base URL:  https://claudeapi.com
API Key:   sk-...
Models:   claude-sonnet-4-6(everyday use) / claude-haiku-4-5-20251001(fast + budget-friendly)
Base URL:  https://claudeapi.com
API Key:   sk-...
Models:   claude-sonnet-4-6(everyday use) / claude-haiku-4-5-20251001(fast + budget-friendly)

What Our Testing Confirmed

  • ✅ All 5 Claude models passed — working correctly
  • ✅ Streaming output works — time to first token: 5.38s
  • ✅ Multi-turn context maintained across 3 rounds — 22.41s total
  • ✅ Long-form generation works — 478-word article in 42.22s

👉 ** Get your API key at claudeapi.com and put Claude to work as your AI assistant today.**


Written by the ClaudeAPI.com engineering team. All benchmarks were collected in 2026. Have questions or running into issues? Visit ClaudeAPI.com — we’re happy to help.

Related Articles