The coordinator measures real RTT to every Island and places jobs on the lowest-latency match. Your request doesn't cross an ocean to reach a GPU.
Compute exchange
Hosts set asking prices. Users set bid preferences. The coordinator clears a market rate for each Cargo type. No fixed tiers, no surprise bills.
Data residency
Jobs default to regional hosts. For compliance, set explicit jurisdiction policies—EU-only, Swiss, or custom—and your data never leaves.
Streaming-first
LLM responses stream token-by-token over WebSocket or SSE. Image generation reports progress. No polling, no waiting for a full response.
Zero-trust hosts
Identity verification via Stripe. Cargos are cosign-signed containers running under seccomp profiles. Hosts never see raw user data.
Karma & reputation
Every Island builds a public karma score from success rate, latency, and uptime. Higher karma means more jobs and better rates.
Multi-Island compute
Models too large for one device run across multiple Islands in a pipeline. A 70B model on four laptops. Batch jobs fan out to fifty phones. Transparent to consumers.
Cargo Marketplace
Browse 133+ Cargos in the registry. Developers publish models and earn per execution. Reviews, ratings, and publisher tiers build trust.
Mobile compute
iPhones and iPads become Islands via the iOS agent. On-device LLM inference with CoreML — no GPU workstation needed.
Batch processing
Submit up to 100 inputs in a single API call. Each dispatched to a different Island. Progress tracked in real time, results merged automatically.
Workflow orchestration
Chain models into DAG pipelines. Steps run in parallel across Islands, data flows between them via templates, and results merge automatically.
Inference caching
Similar prompts served from cache instantly — free. Semantic matching catches near-duplicates. 30-60% cost reduction for repetitive workloads.
A/B testing
Deploy multiple model versions, split traffic, measure latency and accuracy per variant. Auto-promote the winner with statistical significance.
Federated fine-tuning
Train models across Islands without centralizing data. Each Island trains locally, shares only gradients. Your data never leaves your machines.
Confidential inference
Encrypted end-to-end. Islands can't see your data, even while processing it. TEE hardware attestation or homomorphic encryption — your choice.
05 — For Developers
Drop-in API, distributed backend
OpenAI-compatible chat endpoint. Switch your base URL and your existing code runs on community GPUs instead of a single cloud provider.
OpenAI-compatible /v1/chat/completions
Python and JavaScript/React SDKs
Web chat, image generation, and API playground
Batch processing and real-time streaming
Jurisdiction routing for data residency compliance
# Works with any OpenAI-compatible clientfrom openai import OpenAI
client = OpenAI(
base_url="https://app.archipelag.io/api/v1",
api_key="your-key"
)
stream = client.chat.completions.create(
model="mistral-7b",
messages=[{"role": "user", "content": "Hello!"}],
stream=True
)
for chunk in stream:
print(chunk.choices[0].delta.content, end="")
Your machineYour Island, online
Job arrives via NATSSigned container pulled & executed
Credits earnedAt your asking price, paid via Stripe
06 — For Islands
Your hardware, your price
Turn any machine with a GPU into an Island. Set your own asking price and availability schedule. The coordinator matches you with nearby consumers who need compute.
Desktop GPU or mobile device — iPhones work too
You set the price — market clearing, not fixed tiers
Sandboxed containers — seccomp profiles, no host access
Karma score — good uptime means more jobs
Or publish Cargos — earn per execution as a developer