EU AI Act Ready (Aug 2026)

Deterministic AI Orchestration

Stop guessing with black boxes. Start engineering with auditable graphs. The only Deterministic engine for the Agentic Age.

Lár

The "PyTorch" for Agents

Lár provides the computational graph primitives to build cognitive architectures, powered by LiteLLM for universal model access. No magic. No hidden prompts. Just pure, debuggable Python.

Deterministic Primitives

Built on Graph Theory. Define nodes and strict edges. No "loops until success" magic.

Granular Cost Tracking

Track token usage and cost per node, not just per run. Know exactly which agent is burning your budget.

Zero Abstraction Leaks

You see the raw prompt. You see the raw response. You own the `traceback`. Nothing is hidden.

Universal Provider Support

Powered by LiteLLM. Switch between OpenAI, Anthropic, Bedrock, and Ollama in 1 second. No refactoring required.

"Did the AI just do that?"

A customer was refunded $500. A sensitive file was deleted. A regulated decision was made.

In the Black Box Era, you're guessing.
With Lár, you have the Flight Recorder.

The Paradigm Shift

From "Magic" to Engineering

We built the same agent using LangChain and Lár.
When we hit a 429 Rate Limit error, this is what happened:

The Black Box

// support_app_lc.py execution
final_response = full_agent_chain.invoke({"task": task})
Traceback (most recent call last):
File "langchain_core/runnables/base.py", line 2877, in invoke
input = context.run(step.invoke, input, config, **kwargs)
File "langchain_core/runnables/passthrough.py", line 495, in invoke
return self._call_with_config(self._invoke, input, config, **kwargs)
File "langchain_google_genai/chat_models.py", line 198, in _chat_with_retry
return _chat_with_retry(**kwargs)
File "tenacity/__init__.py", line 418, in exc_check
raise retry_exc.reraise()
google.api_core.exceptions.ResourceExhausted: 429 You exceeded your current quota.
* Quota exceeded for metric: generativelanguage.googleapis.com/generate_content_free_tier_requests
* limit: 2, model: gemini-2.5-pro
> System Crash. Stack depth: 129 lines.
> State lost.
VS

The Lár Glass Box

Structured errors. Instant clarity.

{
  "run_id": "ab7c23bb-481e-4241",
  "step": 4,
  "node": "LLMNode",
  "outcome": "error",
  "error": {
    "code": 429,
    "message": "You exceeded your current quota...",
    "status": "RESOURCE_EXHAUSTED",
    "details": [
      {
        "@type": "type.googleapis.com/google.rpc.QuotaFailure",
        "violations": [
          {
            "quotaMetric": "generate_content_free_tier_requests",
            "quotaValue": "2"
          }
        ]
      }
    ]
  }
}

1. The Benchmark

Stress Test: 60-Node Swarm

Don't burn money on
Magic Loops.

We built the same "Corporate Swarm" using standard agent frameworks (Chat Loops) and Lár (Assembly Line).

The difference isn't just speed. It's viability.

Cost Per Run
$0.00
vs $3.60 (Standard)
Latency
0.08s
vs 64s (Standard)
Crash Rate
0%

Standard agents hit RecursionLimit at step 25.
Lár ran 10,000+ steps without a single error.

2. The Guardrails

August 2026 Ready

Engineered for the
EU AI Act

Stop building "Black Box" agents that will be illegal in 2026. Lár is designed specifically for High-Risk AI Systems in Healthcare, Fintech, and Critical Infrastructure.

12
Article 12: Record Keeping

Native "State-Diff Ledger" produces forensic, immutable JSON logs for every step.

13
Article 13: Transparency

"Glass Box" architecture means no magic loops. Every decision path is explicit code.

14
Article 14: Human Oversight

**HumanJuryNode**: A dedicated primitive that enforces a "Hardware Stop," strictly preventing execution until a human explicitly approves via CLI or API.

flight_recorder.json ● Live
{"timestamp": "2026-08-12T10:00:01Z", "node": "TriageNode", "input": "Patient reports chest pain...", "decision": "high_risk", "confidence": 0.98, "model": "gpt-4o-2024-08-06", "latency_ms": 124 }, {"timestamp": "2026-08-12T10:00:02Z", "node": "Supervisor", "action": "INTERRUPT_TRIGGERED", "reason": "Human Review Required" }
Enterprise Flagship

Lár Juried Layer

The "Grand Unification" architecture for High-Risk AI. It combines LLM Reasoning ("Proposer") with Deterministic Policy ("Jury") and Human Interrupts to stop hallucinations before they execute.

3. The Brain

Agentic Metacognition

Traditional agents have static brains. Lár agents have Dynamic Graphs. They can introspect, spot their own limitations, and spawn new subgraphs at runtime. See Examples →

Flagship Showcase

Lár DMN

A Bicameral Cognitive Architecture that solves catastrophic forgetting. It has a conscious mind (fast) and a subconscious mind (slow) that sleeps, dreams, and consolidates memories when you're away.

Just-in-Time Integrations

Zero Maintenance.
Infinite Possibilities.

Other frameworks ship 500+ "wrapper tools" (e.g., HubSpotTool) that break whenever an API changes.

Lár takes a different approach. You don't need a library of stale wrappers. You need a prompt that teaches your IDE how to wrap any Python SDK in 30 seconds. Read the Pattern →

Always Fresh

Use the official, latest SDK. No waiting for framework updates.

Glass Box Ownership

You generate the code. You read it. You own it.

# 1. User drags "IDE_INTEGRATION_PROMPT.md"
User: "Make a Stripe tool for refunds"

# 2. IDE Generates Production Code (30s)
import stripe
from lar import ToolNode

def refund_charge(state):
  stripe.api_key = state["stripe_key"]
  return stripe.Refund.create(...)

# 3. Ready to use
stripe_tool = ToolNode(
  tool_function=refund_charge,
  ...
)
The Agentic Workflow

Don't want to read docs?

Lár is Pure Python. There is no custom DSL or "Magic Chain" syntax to learn. This means Cursor, Windsurf, and Copilot are already Lár experts.

User: "Build a Lár agent that researches stocks."

Cursor: "Done. Since Lár is just Python graphs, I used the `RAG Researcher` pattern. Here is the strict type-checked code..."

Step 1
Clone the Repo
git clone snath-ai/lar
Step 2
Open in Cursor

Or Windsurf / VS Code

Step 3
Just Ask

"Build a Research Agent"

Power Your IDE

Make Cursor or Windsurf an expert Lár Architect.

Step 1: The Brain

Reference the Master Rules file to load the constraints.

IDE_MASTER_PROMPT.md
Step 2: The Hands

Generate any integration (Stripe, Linear, etc) in 30s.

IDE_INTEGRATION_PROMPT.md
Step 3: The Ask

Use the template to scaffold a new agent.

IDE_PROMPT_TEMPLATE.md
(Requires cloning the repo.) | Read the Guide

Ready for Production?

Lár is just a Python library. This means you deploy it like any other backend service. No proprietary "serving layers." No vendor lock-in.

# examples/basic/4_fastapi_server.py
from fastapi import FastAPI
from lar import GraphExecutor

@app.post("/run")
def run_agent(task):
  return executor.run(task)
FastAPI Ready

Deploy to AWS, Railway, or Heroku in minutes.

Ready to Build?

We have 30+ examples ranging from "Hello World" to "Self-Healing Swarms".

Browse the Full Library →

Built by @axdithyaxo