Skip to content

Latest commit

 

History

History
301 lines (256 loc) · 14.1 KB

File metadata and controls

301 lines (256 loc) · 14.1 KB
layout page
title Cheatsheet
permalink /cheatsheet/

Quick reference for mini-a. Bookmark this page. {: .cheatsheet}

Installation

opack install mini-a

Alias (Optional)

# Zsh
echo 'alias mini-a="opack exec mini-a"' >> ~/.zshrc && . ~/.zshrc

If alias setup is not available, run commands as opack exec mini-a [...].

Model Setup

Provider Config
OpenAI export OAF_MODEL="(type: openai, model: gpt-5-mini, key: '...')"
Google export OAF_MODEL="(type: gemini, model: gemini-2.0-flash, key: '...')"
Anthropic export OAF_MODEL="(type: anthropic, model: claude-sonnet-4-20250514, key: '...')"
Ollama export OAF_MODEL="(type: ollama, model: 'llama3', url: 'http://localhost:11434')"
Bedrock export OAF_MODEL="(type: bedrock, options: (region: eu-west-1, model: 'anthropic.claude-sonnet-4-20250514-v1:0'))"
GitHub export OAF_MODEL="(type: openai, url: 'https://models.github.ai/inference', model: openai/gpt-5, key: $(gh auth token), apiVersion: '')"

Running mini-a

Command Description
mini-a Interactive console
mini-a goal='...' Direct goal
mini-a onport=8080 Web UI
mini-a chatbotmode=true Chatbot mode
mini-a modelman=true Model manager
mini-a --skills Starter self-contained SKILL.yaml template

Console Commands

Command Action
/help Show commands
/model Current model info
/models Show all configured model tiers (main, LC, validation) with provider and source
/show [prefix] Show active parameters
/skills [prefix] List discovered skills
/compact [n] Compact older history, keep up to latest n exchanges (default 6)
/summarize [n] Summarize older history, keep up to latest n exchanges (default 6)
/context Show token/context breakdown
/reset Reset conversation
/last [md] Reprint last final answer (md for raw markdown)
/save <path> Save last final answer to a file
/stats [mode] [out=file.json] Usage statistics (summary, detailed, tools, memory, wiki)
/wiki list [prefix] List wiki pages (optionally filtered by prefix)
/wiki read <page.md> Print a wiki page
/wiki search <query> Full-text search across wiki pages
/wiki lint Run wiki health checks (broken links, orphans, stale pages)
/history [n] Show recent user goals from conversation history
/exit Exit mini-a
/clear Reset conversation history and accumulated metrics
/cls Clear screen

Key Parameters

Parameter Default Description
agent - Agent profile (markdown file path or inline markdown)
agentfile - Backward-compatible alias for agent
goal - The task to accomplish
useshell false Enable shell commands
readwrite false Allow file writes
chatbotmode false Chat-only mode
maxsteps 15 Max agent steps
maxcontext - Max context tokens
maxcontent - Alias for maxcontext
maxtokens - Max output tokens
deepresearch false Enable iterative research/validation cycles
validationgoal - Quality criteria for deep research
valgoal - Alias for validationgoal
vmodel - Optional dedicated validation model used in deep-research scoring
modelval - Runtime override for the validation model (OAF_VAL_MODEL format)
maxcycles 3 Max deep research cycles
validationthreshold PASS Validation verdict/score required to stop
persistlearnings true Carry learnings forward between cycles
useplanning false Enable planning
useutils false Built-in utilities
mini-a-docs false Docs-aware Mini Utils root (markdownFiles) when utilsroot is unset
miniadocs false Alias for mini-a-docs
useskills false Expose skill operations in Mini Utils Tool (requires useutils=true)
usetools false Enable tool use
usetoolslc false Register MCP tools only on the low-cost model; main model stays in prompt/action mode
usejsontool false Compatibility json tool for some tool-calling models
toollog - JSSLON channel for MCP tool call logs (input/output)
auditch - Channel for agent activity audit log (tool calls, shell commands, goal events)
debugch - Channel for main-model LLM request/response payloads
debuglcch - Channel for low-cost model LLM request/response payloads
debugfile - Write debug output to a NDJSON file (implies debug=true)
debugvalch - Separate debug channel for the validation model
outfileall - Deep research only: save full cycle outputs (not only final answer)
shelltimeout - Max shell command runtime (ms) before timeout
shellmaxbytes 8000 Cap shell output size and truncate with head/tail excerpt
shellallowpipes false Allow pipes, redirection, and shell control operators
usestream false Stream responses
showexecs false Show shell/exec events as separate lines in the interaction stream
showseparator true Show a separator line between interaction events (disable for compact view)
showthinking false Surface XML-tagged <thinking>...</thinking> blocks as thought logs
promptprofile balanced System prompt verbosity: minimal, balanced, or verbose (auto when debug=true)
systempromptbudget Max estimated tokens for the system prompt; drops lower-priority sections when exceeded
goalprefix Prefix automatically prepended to every goal before the agent sees it
usemath false Enable LaTeX math guidance for KaTeX rendering in web UI
usesvg false Enable SVG generation for custom visuals and infographics
usediagrams false Enable diagram generation
usemermaid false Alias for usediagrams
usecharts false Enable chart generation
usevectors false Enable vector bundle (usesvg=true + usediagrams=true)
mcp - MCP servers to load
mcpproxy false MCP proxy mode
mcpproxythreshold 0 Proxy auto-spill byte threshold (0 disables); large results written to temp file
mcpproxytoon false TOON serialization for spilled proxy results
mcpprogcall false Enable localhost bridge for programmatic MCP tool calls
mcpprogcallport 0 Programmatic MCP bridge port (0 auto-selects)
mcpprogcallmaxbytes 4096 Max inline bridge response size before spill
mcpprogcallresultttl 600 TTL (seconds) for spilled bridge results
mcpprogcallbatchmax 10 Max tool calls accepted per bridge batch request
maxpromptchars 120000 Max accepted prompt size for incoming web prompts
lccontextlimit 0 Escalate to main model when low-cost model context gets too large
deescalate 3 Successful steps before returning from main model to low-cost model
modellock auto Force model tier: main, lc, or auto
modelstrategy default Model orchestration profile: default or advisor (LC executes, main model consulted selectively)
advisormaxuses 2 Max advisor consultations per run when modelstrategy=advisor
advisorcooldownsteps 2 Min steps between advisor consultations
usewiki false Enable wiki knowledge base
wikiaccess ro Wiki access: ro (read-only) or rw (read-write)
wikibackend fs Wiki backend: fs (filesystem) or s3
wikiroot . Root directory for the fs backend
wikibucket - S3 bucket name (s3 backend)
memoryuser false File-backed session + global working memory preset with auto-promotion and 30-day stale sweep
memorypromote "" Auto-promote selected sections from session to global memory at session end
memorystaledays 0 Mark unconfirmed global memory entries as stale after N days
memoryinject summary Inject memory as section counts (summary) or full compact entries (full)
memorysessionheader HTTP request header name used to derive the memory session ID in web mode (e.g. X-User-Id)
llmcomplexity false Validate medium-complexity routing with an LC model check
metricsch - SLON/JSON channel for recording periodic metrics snapshots
usedelegation false Agent delegation
workers - Remote worker URLs for delegation
workertags Comma-separated tags appended to the default worker skill in the AgentCard
usea2a false Use A2A HTTP+JSON endpoints for remote worker delegation
apitoken Bearer token for authenticating requests to the worker API server
autodelegation false Auto-delegate noisy/oversized tool results to a summarization sub-agent (requires usedelegation=true)
autodelegationthreshold 8192 Byte size threshold that triggers auto-delegation
autodelegationmaxperstep 2 Max auto-delegations per agent step
noisytools "" Comma-separated tool names always auto-delegated regardless of result size
subtasks "" Pipe-separated startup scout goals submitted before the main loop
subtasksfile "" Path to YAML/JSON file of {goal, fork, args, timeout} startup task objects
subtaskssequential false Run startup scouts (and all subtasks) one at a time instead of in parallel
forkstatemaxbytes 65536 Max bytes of serialized fork state transmitted to a remote worker
onport - Web UI port
homedir Override the home directory used to locate the .openaf-mini-a folder

Agent Files

Load a complete agent profile from a markdown file with YAML frontmatter:

mini-a agent=my-agent.agent.md goal="..."
mini-a --agent   # Print a starter template

Supported frontmatter keys: name, description, model, capabilities, tools, constraints, rules, knowledge, youare, mini-a (parameter overrides).

agentfile= is a backward-compatible alias for agent=.

File Inclusion

mini-a useutils=true goal='@file.txt Summarize this file'
mini-a useutils=true goal='@data.csv Analyze it'

Custom Commands, Skills, Hooks

Item Path/Command
Custom slash template ~/.openaf-mini-a/commands/<name>.md
Skill (folder) ~/.openaf-mini-a/skills/<name>/SKILL.md
Skill (file) ~/.openaf-mini-a/skills/<name>.md
Hooks `~/.openaf-mini-a/hooks/*.yaml
Extra command dirs extracommands=/path/a,/path/b
Extra skill dirs extraskills=/path/a,/path/b
Extra hook dirs extrahooks=/path/a,/path/b
Alternate config base homedir=/opt/shared/mini-a-config
Run one template mini-a exec="/<name> arg1 arg2"
List skills /skills

Mode Presets

Mode Enables
shell Read-only shell access (useshell=true)
shellrw Shell + write access + non-interactive shell approvals
utils Mini Utils Tool preset (useutils=true mini-a-docs=true usetools=true)
chatbot Chat-only mode
internet Internet-focused MCP/tool mode with docs-aware utils
news Internet + RSS news-focused MCP mode
poweruser High-capability preset with shell, utils, proxy tuning, and docs-aware defaults
web Browser UI preset with MCP tools enabled
webfull Full web UI preset with history, attachments, proxying, and richer rendering modes

Custom modes: create ~/.openaf-mini-a_modes.yaml with a modes: map. Custom definitions are merged with built-ins and override duplicates.

# ~/.openaf-mini-a_modes.yaml
modes:
  mypreset:
    useshell: true
    readwrite: true
    maxsteps: 30
mini-a mode=mypreset goal='your goal here'

Dual-Model Setup

export OAF_MODEL="(type: openai, model: gpt-5-mini, key: '...')"
export OAF_LC_MODEL="(type: openai, model: gpt-5-nano, key: '...')"
export OAF_VAL_MODEL="(type: openai, model: gpt-5-mini, key: '...')"
# Saves 50-70% on token costs
# Optional prompt-mode controls (Gemini main auto-enables when unset)
export OAF_MINI_A_NOJSONPROMPT=true
export OAF_MINI_A_LCNOJSONPROMPT=true

Full Debug

Capture everything — agent activity, main-model payloads, and low-cost model payloads — to separate JSON files:

mini-a goal="your goal" \
  auditch="(type: file, options: (file: audit.json))" \
  debugch="(type: file, options: (file: debug.json))" \
  debuglcch="(type: file, options: (file: debuglc.json))"

Add debugvalch when using a validation model (deepresearch=true):

mini-a goal="research task" deepresearch=true \
  auditch="(type: file, options: (file: audit.json))" \
  debugch="(type: file, options: (file: debug.json))" \
  debuglcch="(type: file, options: (file: debuglc.json))" \
  debugvalch="(type: file, options: (file: debugval.json))"

Config Home (homedir)

By default Mini-A reads ~/.openaf-mini-a. Override with homedir=:

# Shared team config
mini-a homedir=/opt/shared/mini-a-config goal="..."

# Per-project isolated config
mini-a homedir=./my-project-config goal="..."

# Shared base + extra project skills
mini-a homedir=/opt/shared/mini-a-config extraskills=./project-skills goal="..."

Docker

# Interactive
docker run -it -e OAF_MODEL="(type: openai, model: gpt-5-mini, key: '...')" openaf/mini-a
# Web UI
docker run -p 8080:8080 -e OAF_MODEL="(type: openai, model: gpt-5-mini, key: '...')" openaf/mini-a onport=8080

Common MCP Servers

Server Tools Example
mcp-time Current time, timezone conversion mcp="(cmd: 'ojob mcps/mcp-time.yaml')"
mcp-db SQL queries, schema inspection mcp="(cmd: 'ojob mcps/mcp-db.yaml')"
mcp-web Web search, fetch pages mcp="(cmd: 'ojob mcps/mcp-web.yaml')"
mcp-file File operations mcp="(cmd: 'ojob mcps/mcp-file.yaml')"
mcp-shell Shell command execution mcp="(cmd: 'ojob mcps/mcp-shell.yaml')"