Skip to content
#

llm-runtime

Here are 3 public repositories matching this topic...

Language: All
Filter by language

L0: The Missing Reliability Substrate for AI. Streaming-first. Reliable. Replayable. Deterministic. Multimodal. Retries. Continuation. Fallbacks (provider & model). Consensus. Parallelization. Guardrails. Atomic event logs. Byte-for-byte replays.

  • Updated Dec 10, 2025
  • TypeScript

L0: The Missing Reliability Substrate for AI. Streaming-first. Reliable. Replayable. Deterministic. Multimodal. Retries. Continuation. Fallbacks (provider & model). Consensus. Parallelization. Guardrails. Atomic event logs. Byte-for-byte replays.

  • Updated Dec 10, 2025
  • Python

mindscript-runtime is the minimal, reference implementation of the MindScript engine. It provides: - a CLI for running .ms / .ms.md files - a parser that converts MindScript into an internal AST - a stage-based runtime that executes each stage sequentially - adapters for different LLM backends (OpenAI, Gemini, local stub)

  • Updated Dec 10, 2025
  • Python

Improve this page

Add a description, image, and links to the llm-runtime topic page so that developers can more easily learn about it.

Curate this topic

Add this topic to your repo

To associate your repository with the llm-runtime topic, visit your repo's landing page and select "manage topics."

Learn more