Skip to main content
Blaxel is a cloud computing infrastructure built for AI agents. Our computing platform provides compute runtimes and hosting services that AI can use — from code sandboxes to background task infrastructure.
An AI agent is any application that leverages generative AI models to take autonomous actions in the real world. These agents often require computing power to execute their interactions.
  • Conversational agents that are able to take action in the world while keeping a human in the loop for activation or validation: for example code generation agents with real-time previews.
  • AI-powered data pipelines: for example a data transformation pipeline that retrieves unstructured video files and uses an AI model to extract structured data then act on it.
  • RAG agents: for example, a chatbot assistant that can better answer consumers’ queries by autonomously running scripts to access a relevant databases.
  • Autonomous system agents that handle machine-to-machine workflows: like a smart traffic monitoring system that analyzes video feeds, detects accidents in real-time, and automatically dispatches emergency services with AI-generated incident reports.
Blaxel’s infrastructure platform gives production-grade agents their own computing environments including code sandboxes, tool servers, and LLMs. It offers infrastructure to run these agents on a global network that makes them run fast and reliably.
This portal provides comprehensive documentation and API, SDK and CLI reference to help you operate Blaxel Platform.

Essential concepts

Blaxel is a cloud designed for agentic AI. It doesn’t force you into any kind of workflow or shaped box. While we encourage you to exploit architecture designs that we consider are more reliable, our toolkit gives you all the pieces you need to build reliable agentic systems exactly the way you want. Blaxel consists of modular services that are engineered to work seamlessly together, but you can also just use any one of them independently. Think of it as a purpose-built set of building blocks that you can use to power agents.

The building blocks

At the heart of Blaxel is our flagship Sandbox service. They are secure instant-launching compute environments that you can use for running AI code.
  • Automatic scale-to-zero after 5s inactivity, resume from standby under 25ms even after weeks.
  • Micro-VMs with full access to file system, processes and logs — and native support for Zero Data Retention (ZDR).
  • Preview URLs with your own custom domain.
The rest of Blaxel’s cloud services include:
  • Agents Hosting - Deploy your AI agents as serverless auto scalable endpoints. Completely framework agnostic: just bring your code, Blaxel builds it and runs it for you.
  • Batch Jobs - Scalable compute engine designed for agents to schedule and execute many AI processing tasks in parallel in the background
  • MCP Servers Hosting - Deploy custom tool servers on a fast-starting infrastructure to extend your agents’ capabilities.
  • Model Gateway - Intelligent routing layer to LLM providers with built-in telemetry, token cost control, and fallbacks capabilities
  • Full observability — out-of-the-box

A cloud built for agents

Agents will transform how we work in the coming years. Traditional cloud providers weren’t designed to handle them and their one-size-fits-all architecture holds them back. We built Blaxel to fix that. Blaxel is a cloud where AI agents themselves are the primary users. All products are accessible through MCP servers, allowing agents to create and manage resources via tool calls. Blaxel provides agents with all the compute they need to scale and perform optimally: products like Sandboxes give them their own dedicated personal computer(s) / computing environments, while Batch Jobs enable them to schedule background tasks at scale.
As the ultimate AI builder’s toolkit, Blaxel doesn’t require you to learn and adopt a framework or architecture. However, we do recommend best-practices from our experience working with top AI teams and aim to provide guardrails and framing when you build your agents.
  • Break down and distribute your agents whenever possible. A single monolithic agent handling all tool calls, LLM calls, and task workflows can be deployed to Blaxel—but it will be harder to maintain, monitor, and will use resources inefficiently. Blaxel SDK allows builders to split services and connect them from your code.
  • You can call LLM providers directly from your code, but we recommend you go through Blaxel’s Model Gateway for telemetry.
  • Similarly, while direct tool calls are possible, deploying separate MCP servers improves reusability, optimizes resources, and simplifies monitoring. Blaxel also optimizes placement globally when your serverless tool server needs to make multiple backend calls.
  • Break large agents into smaller, specialized sub-agents when possible—they’re easier to debug and observe.
  • Agentic systems naturally connect with many services both inside and outside your network, mixing North-South and East-West traffic in cloud terms. Strong observability is essential for reliability.
  • Reliability is the biggest challenge in agentic AI. Focus on fine-tuning your prompts, tool calls, data access, and orchestration logic—Blaxel will handle the execution.

Which component should I use?

When building your agentic system, you’ll need to make architecture design choices. Blaxel offers several high-perf compute options, summarized below in order of latency performance:
  • Sandboxes: Perfect for maximum workload flexibility. These micro VMs provide full access to filesystem, network, and processes, booting from hibernation in under 25ms.
  • Agents Hosting (sync mode): Ideal for running HTTP API services that process requests within a few seconds.
  • Agents Hosting (async mode): Best for running HTTP API services handling longer requests without maintaining an open connection.
  • Batch Jobs: Designed for asynchronous tasks that may run for extended periods where boot latency is less critical. These jobs are triggered by providing specific input parameters, unlike Agents that are a fully hosted API.
ProductTypical useTypical workload durationBoot timeInput type
SandboxesGiving an agent its own compute runtimeseconds to hours~25ms (from standby)Fully custom
Agents Hosting (sync mode)Agent API that answers fasta few seconds (maximum 100 s)~2-4sCustom API endpoints
Agents Hosting (async mode)Agent API that processes data for a whilea few minutes (maximum 10 mins)~5sCustom API endpoints
Batch JobsSub-tasks scheduled in an agentic workflowminutes to hours (maximum 24 h)~30sSpecific input parameters
MCP Servers HostingRunning an MCP server APIseconds to minutes (maximum 10 mins)~2-4sAPI following MCP

The Blaxel powerhouse

When you deploy workloads to Blaxel, they run on a technical backbone called the Global Agentics Network. Its natively serverless architecture automatically scales computing resources without any server management on your part. Global Agentics Network serves as the powerhouse for the entire Blaxel platform, from Agents Hosting to Sandboxes. It is natively distributed in order to optimize for low-latency or other strategies. It allows for multi-region deployment, enabling AI workloads (such as an AI agent processing inference requests) to run across multiple geographic areas or cloud providers. This is accomplished by decoupling this execution layer from a data layer made of a smart distributed network that federates all those execution locations. Finally, the platform implements advanced security measures, including fine-grained authentication and authorization through Blaxel IAM, ensuring that your AI infrastructure remains protected. It can be interacted with through various methods, including APIs, CLI, web console, and MCP servers.

Documentation structure

You might want to start with any of the following articles:
  • Get started: Deploy your first workload on Blaxel in just 3 minutes.
  • Product Documentation
    • Sandboxes: Equip your agents with fast & secure virtual machines to run AI code.
    • Agents Hosting: Host and run AI agents as serverless auto-scalable endpoints.
    • Batch jobs: Background tasks for your AI workflows that run in batches.
    • MCP Servers Hosting: Expose capabilities and execute tool calls using MCP.
    • Model APIs: Learn about supported model types on our global AI gateway.
    • Integrations: Discover how Blaxel works with other tools, frameworks, and platforms.
    • Observability: Monitor logs, traces and metrics for your agent runs.
    • Policies Governance: Manage your AI deployment strategies.
    • Security: Implement robust security measures for your AI infrastructure.
    • Regions: Discover where Blaxel is available in the world.
  • API reference: Comprehensive guide to Blaxel’s APIs.
  • CLI reference: Learn how to use Blaxel’s command-line interface.
⌘I