Quick reference for mini-a. Bookmark this page.
Installation
Alias (Optional)
# Zsh
echo 'alias mini-a="opack exec mini-a"' >> ~/.zshrc && . ~/.zshrc
If alias setup is not available, run commands as opack exec mini-a [...].
Model Setup
| Provider |
Config |
| OpenAI |
export OAF_MODEL="(type: openai, model: gpt-5-mini, key: '...')" |
| Google |
export OAF_MODEL="(type: gemini, model: gemini-2.0-flash, key: '...')" |
| Anthropic |
export OAF_MODEL="(type: anthropic, model: claude-sonnet-4-20250514, key: '...')" |
| Ollama |
export OAF_MODEL="(type: ollama, model: 'llama3', url: 'http://localhost:11434')" |
| Bedrock |
export OAF_MODEL="(type: bedrock, options: (region: eu-west-1, model: 'anthropic.claude-sonnet-4-20250514-v1:0'))" |
| GitHub |
export OAF_MODEL="(type: openai, url: 'https://models.github.ai/inference', model: openai/gpt-5, key: $(gh auth token), apiVersion: '')" |
Running mini-a
| Command |
Description |
mini-a |
Interactive console |
mini-a goal='...' |
Direct goal |
mini-a onport=8080 |
Web UI |
mini-a chatbotmode=true |
Chatbot mode |
mini-a modelman=true |
Model manager |
mini-a --skills |
Starter self-contained SKILL.yaml template |
Console Commands
| Command |
Action |
/help |
Show commands |
/model |
Current model info |
/models |
Show all configured model tiers (main, LC, validation) with provider and source |
/show [prefix] |
Show active parameters |
/skills [prefix] |
List discovered skills |
/compact [n] |
Compact older history, keep up to latest n exchanges (default 6) |
/summarize [n] |
Summarize older history, keep up to latest n exchanges (default 6) |
/context |
Show token/context breakdown |
/reset |
Reset conversation |
/last [md] |
Reprint last final answer (md for raw markdown) |
/save <path> |
Save last final answer to a file |
/stats [mode] [out=file.json] |
Usage statistics (summary, detailed, tools, memory) |
/history [n] |
Show recent user goals from conversation history |
/exit |
Exit mini-a |
/clear |
Reset conversation history and accumulated metrics |
/cls |
Clear screen |
Key Parameters
| Parameter |
Default |
Description |
agent |
- |
Agent profile (markdown file path or inline markdown) |
agentfile |
- |
Backward-compatible alias for agent |
goal |
- |
The task to accomplish |
useshell |
false |
Enable shell commands |
readwrite |
false |
Allow file writes |
chatbotmode |
false |
Chat-only mode |
maxsteps |
15 |
Max agent steps |
maxcontext |
- |
Max context tokens |
maxcontent |
- |
Alias for maxcontext |
maxtokens |
- |
Max output tokens |
deepresearch |
false |
Enable iterative research/validation cycles |
validationgoal |
- |
Quality criteria for deep research |
valgoal |
- |
Alias for validationgoal |
vmodel |
- |
Optional dedicated validation model used in deep-research scoring |
modelval |
- |
Runtime override for the validation model (OAF_VAL_MODEL format) |
maxcycles |
3 |
Max deep research cycles |
validationthreshold |
PASS |
Validation verdict/score required to stop |
persistlearnings |
true |
Carry learnings forward between cycles |
useplanning |
false |
Enable planning |
useutils |
false |
Built-in utilities |
mini-a-docs |
false |
Docs-aware Mini Utils root (markdownFiles) when utilsroot is unset |
miniadocs |
false |
Alias for mini-a-docs |
useskills |
false |
Expose skill operations in Mini Utils Tool (requires useutils=true) |
usetools |
false |
Enable tool use |
usejsontool |
false |
Compatibility json tool for some tool-calling models |
toollog |
- |
JSSLON channel for MCP tool call logs (input/output) |
debugfile |
- |
Write debug output to a NDJSON file (implies debug=true) |
debugvalch |
- |
Separate debug channel for the validation model |
outfileall |
- |
Deep research only: save full cycle outputs (not only final answer) |
shelltimeout |
- |
Max shell command runtime (ms) before timeout |
shellmaxbytes |
8000 |
Cap shell output size and truncate with head/tail excerpt |
shellallowpipes |
false |
Allow pipes, redirection, and shell control operators |
usestream |
false |
Stream responses |
showthinking |
false |
Surface XML-tagged <thinking>...</thinking> blocks as thought logs |
usemath |
false |
Enable LaTeX math guidance for KaTeX rendering in web UI |
usesvg |
false |
Enable SVG generation for custom visuals and infographics |
usediagrams |
false |
Enable diagram generation |
usemermaid |
false |
Alias for usediagrams |
usecharts |
false |
Enable chart generation |
usevectors |
false |
Enable vector bundle (usesvg=true + usediagrams=true) |
mcp |
- |
MCP servers to load |
mcpproxy |
false |
MCP proxy mode |
mcpproxytoon |
false |
TOON serialization for spilled proxy results |
mcpprogcall |
false |
Enable localhost bridge for programmatic MCP tool calls |
mcpprogcallport |
0 |
Programmatic MCP bridge port (0 auto-selects) |
mcpprogcallmaxbytes |
4096 |
Max inline bridge response size before spill |
mcpprogcallresultttl |
600 |
TTL (seconds) for spilled bridge results |
mcpprogcallbatchmax |
10 |
Max tool calls accepted per bridge batch request |
maxpromptchars |
120000 |
Max accepted prompt size for incoming web prompts |
lccontextlimit |
0 |
Escalate to main model when low-cost model context gets too large |
deescalate |
3 |
Successful steps before returning from main model to low-cost model |
modellock |
auto |
Force model tier: main, lc, or auto |
modelstrategy |
default |
Model orchestration profile: default or advisor (LC executes, main model consulted selectively) |
advisormaxuses |
2 |
Max advisor consultations per run when modelstrategy=advisor |
advisorcooldownsteps |
2 |
Min steps between advisor consultations |
memoryuser |
false |
File-backed session + global working memory preset with auto-promotion and 30-day stale sweep |
memorypromote |
"" |
Auto-promote selected sections from session to global memory at session end |
memorystaledays |
0 |
Mark unconfirmed global memory entries as stale after N days |
memoryinject |
summary |
Inject memory as section counts (summary) or full compact entries (full) |
llmcomplexity |
false |
Validate medium-complexity routing with an LC model check |
metricsch |
- |
SLON/JSON channel for recording periodic metrics snapshots |
usedelegation |
false |
Agent delegation |
workers |
- |
Remote worker URLs for delegation |
onport |
- |
Web UI port |
Agent Files
Load a complete agent profile from a markdown file with YAML frontmatter:
mini-a agent=my-agent.agent.md goal="..."
mini-a --agent # Print a starter template
Supported frontmatter keys: name, description, model, capabilities, tools, constraints, rules, knowledge, youare, mini-a (parameter overrides).
agentfile= is a backward-compatible alias for agent=.
File Inclusion
mini-a useutils=true goal='@file.txt Summarize this file'
mini-a useutils=true goal='@data.csv Analyze it'
Custom Commands, Skills, Hooks
| Item |
Path/Command |
| Custom slash template |
~/.openaf-mini-a/commands/<name>.md |
| Skill (folder) |
~/.openaf-mini-a/skills/<name>/SKILL.md |
| Skill (file) |
~/.openaf-mini-a/skills/<name>.md |
| Hooks |
~/.openaf-mini-a/hooks/*.yaml|*.yml|*.json |
| Extra command dirs |
extracommands=/path/a,/path/b |
| Extra skill dirs |
extraskills=/path/a,/path/b |
| Extra hook dirs |
extrahooks=/path/a,/path/b |
| Run one template |
mini-a exec="/<name> arg1 arg2" |
| List skills |
/skills |
Mode Presets
| Mode |
Enables |
shell |
Read-only shell access (useshell=true) |
shellrw |
Shell + write access + non-interactive shell approvals |
utils |
Mini Utils Tool preset (useutils=true mini-a-docs=true usetools=true) |
chatbot |
Chat-only mode |
internet |
Internet-focused MCP/tool mode with docs-aware utils |
news |
Internet + RSS news-focused MCP mode |
poweruser |
High-capability preset with shell, utils, proxy tuning, and docs-aware defaults |
web |
Browser UI preset with MCP tools enabled |
webfull |
Full web UI preset with history, attachments, proxying, and richer rendering modes |
Custom modes: create ~/.openaf-mini-a_modes.yaml with a modes: map. Custom definitions are merged with built-ins and override duplicates.
# ~/.openaf-mini-a_modes.yaml
modes:
mypreset:
useshell: true
readwrite: true
maxsteps: 30
mini-a mode=mypreset goal='your goal here'
Dual-Model Setup
export OAF_MODEL="(type: openai, model: gpt-5-mini, key: '...')"
export OAF_LC_MODEL="(type: openai, model: gpt-5-nano, key: '...')"
export OAF_VAL_MODEL="(type: openai, model: gpt-5-mini, key: '...')"
# Saves 50-70% on token costs
# Optional prompt-mode controls (Gemini main auto-enables when unset)
export OAF_MINI_A_NOJSONPROMPT=true
export OAF_MINI_A_LCNOJSONPROMPT=true
Docker
# Interactive
docker run -it -e OAF_MODEL="(type: openai, model: gpt-5-mini, key: '...')" openaf/mini-a
# Web UI
docker run -p 8080:8080 -e OAF_MODEL="(type: openai, model: gpt-5-mini, key: '...')" openaf/mini-a onport=8080
Common MCP Servers
| Server |
Tools |
Example |
mcp-time |
Current time, timezone conversion |
mcp="(cmd: 'ojob mcps/mcp-time.yaml')" |
mcp-db |
SQL queries, schema inspection |
mcp="(cmd: 'ojob mcps/mcp-db.yaml')" |
mcp-web |
Web search, fetch pages |
mcp="(cmd: 'ojob mcps/mcp-web.yaml')" |
mcp-file |
File operations |
mcp="(cmd: 'ojob mcps/mcp-file.yaml')" |
mcp-shell |
Shell command execution |
mcp="(cmd: 'ojob mcps/mcp-shell.yaml')" |