Quick reference for mini-a. Bookmark this page.

Installation

opack install mini-a

Alias (Optional)

# Zsh
echo 'alias mini-a="opack exec mini-a"' >> ~/.zshrc && . ~/.zshrc

If alias setup is not available, run commands as opack exec mini-a [...].

Model Setup

Provider Config
OpenAI export OAF_MODEL="(type: openai, model: gpt-5-mini, key: '...')"
Google export OAF_MODEL="(type: gemini, model: gemini-2.0-flash, key: '...')"
Anthropic export OAF_MODEL="(type: anthropic, model: claude-sonnet-4-20250514, key: '...')"
Ollama export OAF_MODEL="(type: ollama, model: 'llama3', url: 'http://localhost:11434')"
Bedrock export OAF_MODEL="(type: bedrock, options: (region: eu-west-1, model: 'anthropic.claude-sonnet-4-20250514-v1:0'))"
GitHub export OAF_MODEL="(type: openai, url: 'https://models.github.ai/inference', model: openai/gpt-5, key: $(gh auth token), apiVersion: '')"

Running mini-a

Command Description
mini-a Interactive console
mini-a goal='...' Direct goal
mini-a onport=8080 Web UI
mini-a chatbotmode=true Chatbot mode
mini-a modelman=true Model manager
mini-a --skills Starter self-contained SKILL.yaml template

Console Commands

Command Action
/help Show commands
/model Current model info
/models Show all configured model tiers (main, LC, validation) with provider and source
/show [prefix] Show active parameters
/skills [prefix] List discovered skills
/compact [n] Compact older history, keep up to latest n exchanges (default 6)
/summarize [n] Summarize older history, keep up to latest n exchanges (default 6)
/context Show token/context breakdown
/reset Reset conversation
/last [md] Reprint last final answer (md for raw markdown)
/save <path> Save last final answer to a file
/stats [mode] [out=file.json] Usage statistics (summary, detailed, tools, memory)
/history [n] Show recent user goals from conversation history
/exit Exit mini-a
/clear Reset conversation history and accumulated metrics
/cls Clear screen

Key Parameters

Parameter Default Description
agent - Agent profile (markdown file path or inline markdown)
agentfile - Backward-compatible alias for agent
goal - The task to accomplish
useshell false Enable shell commands
readwrite false Allow file writes
chatbotmode false Chat-only mode
maxsteps 15 Max agent steps
maxcontext - Max context tokens
maxcontent - Alias for maxcontext
maxtokens - Max output tokens
deepresearch false Enable iterative research/validation cycles
validationgoal - Quality criteria for deep research
valgoal - Alias for validationgoal
vmodel - Optional dedicated validation model used in deep-research scoring
modelval - Runtime override for the validation model (OAF_VAL_MODEL format)
maxcycles 3 Max deep research cycles
validationthreshold PASS Validation verdict/score required to stop
persistlearnings true Carry learnings forward between cycles
useplanning false Enable planning
useutils false Built-in utilities
mini-a-docs false Docs-aware Mini Utils root (markdownFiles) when utilsroot is unset
miniadocs false Alias for mini-a-docs
useskills false Expose skill operations in Mini Utils Tool (requires useutils=true)
usetools false Enable tool use
usejsontool false Compatibility json tool for some tool-calling models
toollog - JSSLON channel for MCP tool call logs (input/output)
debugfile - Write debug output to a NDJSON file (implies debug=true)
debugvalch - Separate debug channel for the validation model
outfileall - Deep research only: save full cycle outputs (not only final answer)
shelltimeout - Max shell command runtime (ms) before timeout
shellmaxbytes 8000 Cap shell output size and truncate with head/tail excerpt
shellallowpipes false Allow pipes, redirection, and shell control operators
usestream false Stream responses
showthinking false Surface XML-tagged <thinking>...</thinking> blocks as thought logs
usemath false Enable LaTeX math guidance for KaTeX rendering in web UI
usesvg false Enable SVG generation for custom visuals and infographics
usediagrams false Enable diagram generation
usemermaid false Alias for usediagrams
usecharts false Enable chart generation
usevectors false Enable vector bundle (usesvg=true + usediagrams=true)
mcp - MCP servers to load
mcpproxy false MCP proxy mode
mcpproxytoon false TOON serialization for spilled proxy results
mcpprogcall false Enable localhost bridge for programmatic MCP tool calls
mcpprogcallport 0 Programmatic MCP bridge port (0 auto-selects)
mcpprogcallmaxbytes 4096 Max inline bridge response size before spill
mcpprogcallresultttl 600 TTL (seconds) for spilled bridge results
mcpprogcallbatchmax 10 Max tool calls accepted per bridge batch request
maxpromptchars 120000 Max accepted prompt size for incoming web prompts
lccontextlimit 0 Escalate to main model when low-cost model context gets too large
deescalate 3 Successful steps before returning from main model to low-cost model
modellock auto Force model tier: main, lc, or auto
modelstrategy default Model orchestration profile: default or advisor (LC executes, main model consulted selectively)
advisormaxuses 2 Max advisor consultations per run when modelstrategy=advisor
advisorcooldownsteps 2 Min steps between advisor consultations
memoryuser false File-backed session + global working memory preset with auto-promotion and 30-day stale sweep
memorypromote "" Auto-promote selected sections from session to global memory at session end
memorystaledays 0 Mark unconfirmed global memory entries as stale after N days
memoryinject summary Inject memory as section counts (summary) or full compact entries (full)
llmcomplexity false Validate medium-complexity routing with an LC model check
metricsch - SLON/JSON channel for recording periodic metrics snapshots
usedelegation false Agent delegation
workers - Remote worker URLs for delegation
onport - Web UI port

Agent Files

Load a complete agent profile from a markdown file with YAML frontmatter:

mini-a agent=my-agent.agent.md goal="..."
mini-a --agent   # Print a starter template

Supported frontmatter keys: name, description, model, capabilities, tools, constraints, rules, knowledge, youare, mini-a (parameter overrides).

agentfile= is a backward-compatible alias for agent=.

File Inclusion

mini-a useutils=true goal='@file.txt Summarize this file'
mini-a useutils=true goal='@data.csv Analyze it'

Custom Commands, Skills, Hooks

Item Path/Command
Custom slash template ~/.openaf-mini-a/commands/<name>.md
Skill (folder) ~/.openaf-mini-a/skills/<name>/SKILL.md
Skill (file) ~/.openaf-mini-a/skills/<name>.md
Hooks ~/.openaf-mini-a/hooks/*.yaml|*.yml|*.json
Extra command dirs extracommands=/path/a,/path/b
Extra skill dirs extraskills=/path/a,/path/b
Extra hook dirs extrahooks=/path/a,/path/b
Run one template mini-a exec="/<name> arg1 arg2"
List skills /skills

Mode Presets

Mode Enables
shell Read-only shell access (useshell=true)
shellrw Shell + write access + non-interactive shell approvals
utils Mini Utils Tool preset (useutils=true mini-a-docs=true usetools=true)
chatbot Chat-only mode
internet Internet-focused MCP/tool mode with docs-aware utils
news Internet + RSS news-focused MCP mode
poweruser High-capability preset with shell, utils, proxy tuning, and docs-aware defaults
web Browser UI preset with MCP tools enabled
webfull Full web UI preset with history, attachments, proxying, and richer rendering modes

Custom modes: create ~/.openaf-mini-a_modes.yaml with a modes: map. Custom definitions are merged with built-ins and override duplicates.

# ~/.openaf-mini-a_modes.yaml
modes:
  mypreset:
    useshell: true
    readwrite: true
    maxsteps: 30
mini-a mode=mypreset goal='your goal here'

Dual-Model Setup

export OAF_MODEL="(type: openai, model: gpt-5-mini, key: '...')"
export OAF_LC_MODEL="(type: openai, model: gpt-5-nano, key: '...')"
export OAF_VAL_MODEL="(type: openai, model: gpt-5-mini, key: '...')"
# Saves 50-70% on token costs
# Optional prompt-mode controls (Gemini main auto-enables when unset)
export OAF_MINI_A_NOJSONPROMPT=true
export OAF_MINI_A_LCNOJSONPROMPT=true

Docker

# Interactive
docker run -it -e OAF_MODEL="(type: openai, model: gpt-5-mini, key: '...')" openaf/mini-a
# Web UI
docker run -p 8080:8080 -e OAF_MODEL="(type: openai, model: gpt-5-mini, key: '...')" openaf/mini-a onport=8080

Common MCP Servers

Server Tools Example
mcp-time Current time, timezone conversion mcp="(cmd: 'ojob mcps/mcp-time.yaml')"
mcp-db SQL queries, schema inspection mcp="(cmd: 'ojob mcps/mcp-db.yaml')"
mcp-web Web search, fetch pages mcp="(cmd: 'ojob mcps/mcp-web.yaml')"
mcp-file File operations mcp="(cmd: 'ojob mcps/mcp-file.yaml')"
mcp-shell Shell command execution mcp="(cmd: 'ojob mcps/mcp-shell.yaml')"