The fastest Trust Layer for AI Agents
-
Updated
Feb 3, 2026 - Python
The fastest Trust Layer for AI Agents
AI-native application framework and runtime, simply write a YAML file.
Secure autonomous AI agent fleet platform — Docker-isolated, multi-provider, with built-in cost controls. OpenClaw alternative for production use.
Bud AI Foundry - A comprehensive inference stack for compound AI deployment, optimization and scaling. Bud Stack provides intelligent infrastructure automation, performance optimization, and seamless model deployment across multi-cloud/multi-hardware environments.
A minimal kernel for agentic systems. Runtime-first architecture for programmatic tool execution. Inspired by Anthropic's Code Execution with MCP.
A self-evolving, AI-native language and platform for intelligent agents and autonomous software.
L0: The Missing Reliability Substrate for AI. Streaming-first. Reliable. Replayable. Deterministic. Multimodal. Retries. Continuation. Fallbacks (provider & model). Consensus. Parallelization. Guardrails. Atomic event logs. Byte-for-byte replays.
Official LAW-N Runtime Environment (LNRE) — a lightweight virtual machine and execution engine for LAW-N programs, time-aware instructions, and N-SQL operations. Handles bytecode, timing, routing, and runtime introspection for the LAW-N / Mind’s Eye cognition stack.
Add a description, image, and links to the ai-runtime topic page so that developers can more easily learn about it.
To associate your repository with the ai-runtime topic, visit your repo's landing page and select "manage topics."