Quick Start¶
This guide gets you from zero to a running AgentLang pipeline or workflow in five minutes.
Prerequisites¶
- Python 3.14+
- This repository cloned locally
Verify your Python version:
1. Verify the install¶
AgentLang has no external dependencies in its core. Run a syntax check to confirm everything is wired up:
No output means success.
2. Run your first pipeline¶
The blog.agent example defines a simple two-step pipeline: research a topic, then draft an article.
Output:
Mock mode
By default, pipelines run in mock mode — all task handlers are deterministic local functions that return structured placeholders. No API key required.
3. Run your first workflow¶
The higher-level workflow surface is the recommended authoring model for multi-agent handoffs and review loops.
python main.py examples/multiagent_blog.agent publish_topic_blog \
--input '{"topic":"agent memory systems"}'
Lower it to explicit pipeline IR:
This shows the generated run, while, and break structure that the runtime actually executes.
4. Try parallel execution¶
The compare.agent pipeline runs two research tasks in parallel, then merges results:
Output:
{
"result": "[reviewer] Option A vs B\nA: [planner] key points for 'vector database option A'\nB: [planner] key points for 'vector database option B'"
}
The two research calls ran concurrently — you'll see both results merged before the compare step executes.
5. See retry and fallback¶
The reliability.agent pipeline uses retries and on_fail use to handle transient failures gracefully.
Run with a low failure count (succeeds before fallback kicks in):
python main.py examples/reliability.agent resilient_brief \
--input '{"topic":"api-status","fail_count":1}'
Force the fallback by exceeding the retry budget:
python main.py examples/reliability.agent resilient_brief \
--input '{"topic":"api-status","fail_count":5}'
What just happened?
fail_count: 5 exceeds the retries 2 budget in the pipeline, so the on_fail use clause provides a fallback value. The if/else block then routes execution based on whether the fallback was used.
6. Start the REPL¶
For interactive exploration:
AgentLang REPL (adapter=mock). Type 'exit' to quit.
> examples/blog.agent blog_post {"topic":"agent memory"}
{
"result": "[writer] Draft article:\n[planner] key points for 'agent memory'"
}
> exit
Each line at the > prompt takes the form <source_file> <pipeline_or_workflow_name> [json_input]. Errors are printed and the REPL continues — no restart needed.
7. Trace a live run¶
When debugging live agent behavior, enable tracing. This works with both --adapter live (OpenAI) and --adapter anthropic (Claude):
python main.py examples/incident_runbook.agent respond_to_incident \
--adapter live \
--trace-live \
--input '{"incident":"database failover drill"}'
Or with Anthropic/Claude:
python main.py examples/incident_runbook.agent respond_to_incident \
--adapter anthropic \
--trace-live \
--input '{"incident":"database failover drill"}'
Trace lines are printed to stderr and show:
- agent task start/end
- LLM request mode (OpenAI or Anthropic)
- tool calls and tool results
- final structured task outputs