Benchmarks, migration guides, and feature comparisons for JamJet vs other AI agent frameworks.
| Directory | What's inside |
|---|---|
benchmarks/ |
Latency benchmarks: JamJet vs LangGraph vs raw LLM call |
migrate/from-langgraph/ |
Side-by-side code + concept mapping |
migrate/from-crewai/ |
Side-by-side code + concept mapping |
migrate/from-openai-direct/ |
Side-by-side code + concept mapping |
feature-matrix.md |
Full feature comparison across 4 frameworks |
Works with any OpenAI-compatible endpoint. Tested with Ollama locally.
cd benchmarks
pip install -r requirements.txt
# Local (Ollama)
export OPENAI_API_KEY="ollama"
export OPENAI_BASE_URL="http://localhost:11434/v1"
export MODEL_NAME="qwen3:8b"
python bench_single_call.py --json results/$(date +%Y-%m-%d)-ollama-qwen3-8b.json
# OpenAI
export OPENAI_API_KEY="sk-..."
export MODEL_NAME="gpt-4.1-mini"
python bench_single_call.py --json results/$(date +%Y-%m-%d)-openai-gpt41mini.jsonThe benchmark measures framework orchestration overhead on top of a raw LLM call. All three runners (Raw, JamJet, LangGraph) make the identical LLM call through the same OpenAI-compatible client. The difference in latency is purely the framework tax.
Metrics per framework: mean, median, p95, p99, min, max, stdev (all in ms), overhead vs raw baseline.
See benchmarks/results/ for recorded runs.
Each guide has:
- The original framework code (unchanged, idiomatic)
- The JamJet equivalent
- A concept-mapping table
- Key differences explained
Results from different hardware, models, and endpoints are welcome. Open a PR with your results/YYYY-MM-DD-<model>-<notes>.json.
Corrections to the feature matrix are welcome too — if we marked something wrong, open an issue or PR.