- Document modern AI development workflow using specialized agents - Add guidelines for when to use git worktree/subagents vs main conversation - Define agent roles: ticket mgmt, design, code, test, docs, review - Include implementation examples with Task tool - Update test count (35 → 54) with new market_schedule tests Closes #9 Co-Authored-By: Claude Sonnet 4.5 <noreply@anthropic.com>
6.3 KiB
CLAUDE.md
This file provides guidance to Claude Code (claude.ai/code) when working with code in this repository.
Git Workflow Policy
CRITICAL: All code changes MUST follow this workflow. Direct pushes to main are ABSOLUTELY PROHIBITED.
- Create Gitea Issue First — All features, bug fixes, and policy changes require a Gitea issue before any code is written
- Create Feature Branch — Branch from
mainusing formatfeature/issue-{N}-{short-description} - Implement Changes — Write code, tests, and documentation on the feature branch
- Create Pull Request — Submit PR to
mainbranch referencing the issue number - Review & Merge — After approval, merge via PR (squash or merge commit)
Never commit directly to main. This policy applies to all changes, no exceptions.
Agent Workflow
Modern AI development leverages specialized agents for concurrent, efficient task execution.
Parallel Execution Strategy
Use git worktree or subagents (via the Task tool) to handle multiple requirements simultaneously:
- Each task runs in independent context
- Parallel branches for concurrent features
- Isolated test environments prevent interference
- Faster iteration with distributed workload
Specialized Agent Roles
Deploy task-specific agents as needed instead of handling everything in the main conversation:
- Conversational Agent (main) — Interface with user, coordinate other agents
- Ticket Management Agent — Create/update Gitea issues, track task status
- Design Agent — Architectural planning, RFC documents, API design
- Code Writing Agent — Implementation following specs
- Testing Agent — Write tests, verify coverage, run test suites
- Documentation Agent — Update docs, docstrings, CLAUDE.md, README
- Review Agent — Code review, lint checks, security audits
- Custom Agents — Created dynamically for specialized tasks (performance analysis, migration scripts, etc.)
When to Use Agents
Prefer spawning specialized agents for:
- Complex multi-file changes requiring exploration
- Tasks with clear, isolated scope (e.g., "write tests for module X")
- Parallel work streams (feature A + bugfix B simultaneously)
- Long-running analysis (codebase search, dependency audit)
- Tasks requiring different contexts (multiple git worktrees)
Use the main conversation for:
- User interaction and clarification
- Quick single-file edits
- Coordinating agent work
- High-level decision making
Implementation
# Example: Spawn parallel test and documentation agents
task_tool(
subagent_type="general-purpose",
prompt="Write comprehensive tests for src/markets/schedule.py",
description="Write schedule tests"
)
task_tool(
subagent_type="general-purpose",
prompt="Update README.md with global market feature documentation",
description="Update README"
)
Use run_in_background=True for independent tasks that don't block subsequent work.
Build & Test Commands
# Install all dependencies (production + dev)
pip install ".[dev]"
# Run full test suite with coverage
pytest -v --cov=src --cov-report=term-missing
# Run a single test file
pytest tests/test_risk.py -v
# Run a single test by name
pytest tests/test_brain.py -k "test_parse_valid_json" -v
# Lint
ruff check src/ tests/
# Type check (strict mode, non-blocking in CI)
mypy src/ --strict
# Run the trading agent
python -m src.main --mode=paper
# Docker
docker compose up -d ouroboros # Run agent
docker compose --profile test up test # Run tests in container
Architecture
Self-evolving AI trading agent for Korean stock markets (KIS API). The main loop in src/main.py orchestrates four components in a 60-second cycle per stock:
-
Broker (
src/broker/kis_api.py) — Async KIS API client with automatic OAuth token refresh, leaky-bucket rate limiter (10 RPS), and POST body hash-key signing. Uses a custom SSL context with disabled hostname verification for the VTS (virtual trading) endpoint due to a known certificate mismatch. -
Brain (
src/brain/gemini_client.py) — Sends structured prompts to Google Gemini, parses JSON responses intoTradeDecisionobjects. Forces HOLD when confidence < threshold (default 80). Falls back to safe HOLD on any parse/API error. -
Risk Manager (
src/core/risk_manager.py) — READ-ONLY by policy (seedocs/agents.md). Circuit breaker halts all trading viaSystemExitwhen daily P&L drops below -3.0%. Fat-finger check rejects orders exceeding 30% of available cash. -
Evolution (
src/evolution/optimizer.py) — Analyzes high-confidence losing trades from SQLite, asks Gemini to generate newBaseStrategysubclasses, validates them by running the full pytest suite, and simulates PR creation.
Data flow per cycle: Fetch orderbook + balance → calculate P&L → get Gemini decision → validate with risk manager → execute order → log to SQLite (src/db.py).
Key Constraints (from docs/agents.md)
core/risk_manager.pyis READ-ONLY. Changes require human approval.- Circuit breaker threshold (-3.0%) may only be made stricter, never relaxed.
- Fat-finger protection (30% max order size) must always be enforced.
- Confidence < 80 must force HOLD — this rule cannot be weakened.
- All code changes require corresponding tests. Coverage must stay >= 80%.
- Generated strategies must pass the full test suite before activation.
Configuration
Pydantic Settings loaded from .env (see .env.example). Required vars: KIS_APP_KEY, KIS_APP_SECRET, KIS_ACCOUNT_NO (format XXXXXXXX-XX), GEMINI_API_KEY. Tests use in-memory SQLite (DB_PATH=":memory:") and dummy credentials via tests/conftest.py.
Test Structure
54 tests across four files. asyncio_mode = "auto" in pyproject.toml — async tests need no special decorator. The settings fixture in conftest.py provides safe defaults with test credentials and in-memory DB.
test_risk.py(11) — Circuit breaker boundaries, fat-finger edge casestest_broker.py(6) — Token lifecycle, rate limiting, hash keys, network errorstest_brain.py(18) — JSON parsing, confidence threshold, malformed responses, prompt constructiontest_market_schedule.py(19) — Market open/close logic, timezone handling, DST, lunch breaks