LLM Chat
Conversational AI via community GPUs
- Mistral 7B, Llama, and more open models
- Token-by-token streaming responses
- Multi-turn conversation history
- Market-based pricing — varies by model and demand
A Cargo is a portable, signed unit of computation that runs on Islands across the Archipelag.io network.
Think of it as a shipping container for code — each one packages a model, runtime, and resource requirements into a single deployable artifact. Consumers submit jobs, the coordinator finds the best Island, and the Cargo executes in a secure sandbox.
The network currently supports these Cargo types. More are being added during beta.
Conversational AI via community GPUs
Text-to-image on community hardware
Run models too large for any single device, or process batches across dozens of Islands at once.
Shard a large model across multiple Islands
Split parallel work across many Islands at once
Define multi-step workflows as DAGs. The coordinator resolves dependencies, dispatches steps in parallel across Islands, maps data between steps, and merges results.
Translate text, classify the output, then summarize — all in one API call. Each step runs on the best-fit Island. If a step fails, the workflow handles it.
Cache, test, train, and protect — all on the same distributed network.
Similar prompts served from cache instantly
Compare model versions with real traffic
Train models without centralizing data
Encrypted end-to-end — Islands can't see your data
confidential: trueIf your code works with OpenAI, it works with Archipelag.io. Change the base URL and you're done.
# Python — streaming chat
from openai import OpenAI
client = OpenAI(
base_url="https://app.archipelag.io/api/v1",
api_key="your-key"
)
stream = client.chat.completions.create(
model="mistral-7b",
messages=[{"role": "user", "content": "Hello!"}],
stream=True
)
for chunk in stream:
print(chunk.choices[0].delta.content, end="")// JavaScript — streaming chat
import OpenAI from 'openai';
const client = new OpenAI({
baseURL: 'https://app.archipelag.io/api/v1',
apiKey: 'your-key'
});
const stream = await client.chat.completions.create({
model: 'mistral-7b',
messages: [{ role: 'user', content: 'Hello!' }],
stream: true
});
for await (const chunk of stream) {
process.stdout.write(chunk.choices[0]?.delta?.content ?? '');
}Standard OpenAI chat completions format. Supports streaming, temperature, max_tokens, and all common parameters.
Bearer token via API key (read/write scopes) or session cookie for the web UI. Keys are managed in your dashboard.
Optional policy parameter restricts job placement to Islands in specific regions (EU, Switzerland, custom). For teams with compliance requirements.
Prices are set by supply and demand on the compute exchange, not by us.
The compute exchange shows current clearing prices for each Cargo type. Check the exchange for live rates.
The beta uses virtual credits — no real money changes hands. You get 10,000 credits on signup and they auto-refill. After beta: buy credits, no subscriptions.
Every job shows its clearing price before execution. You see exactly what you'll pay and which Island will run it.
Use the full web platform or build your own integration.
Full-featured web chat with conversation history, model selection, and host preferences. Real-time streaming built in.
Web UI for text-to-image with configurable dimensions, inference steps, and a gallery of past generations.
Try the API without signing up. Interactive testing with live code examples in Python and JavaScript.
Submit batches of inputs via the API or web UI. Monitor progress in real time with a visual dashboard.
Browse 133+ Cargos in the registry. Reviews, ratings, and publisher tiers help you find trusted models.
Build and publish your own Cargos. Track usage, manage submissions, and earn revenue per execution.
Email or GitHub. You get free beta credits immediately—no credit card needed.
Use the web chat UI, point your OpenAI client at our base URL, or use the Python/JS SDK.
The coordinator finds the nearest Island. Responses stream back token-by-token in real time.
Create an account and run your first inference job. Free credits, no credit card.