LLM & Agents
6763 skills in Data & AI > LLM & Agents
setup
Sets up new projects and generates workflow files like CLAUDE.md, AGENTS.md, Plans.md. Use when user mentions セットアップ, setup, 初期化, initialize, 新規プロジェクト, ワークフローファイル生成. Do NOT load for: 実装作業, レビュー, ビルド検証, デプロイ.
testing
Write tests following TDD principles. Use when implementing features, fixing bugs, or adding test coverage. Covers e2e, integration, and unit testing patterns.
system-learn
Ingest new procedural memory (skills, patterns, docs) into the vector database.
system-init
Initialize or hydrate the agent's memory system and verify configuration.
fleet-agent
Context-aware development assistant for AgenticFleet with auto-learning and dual memory (NeonDB + ChromaDB). Handles development workflows with intelligent context management.
agenticfleet-frontend-component-workflow
End-to-end guide for creating React components in AgenticFleet's frontend, from design tokens through Tailwind styling, shadcn/ui integration, state management, to testing.
memory-system
Complete guide to the AgenticFleet memory system. Read this first.
prompt-optimize
Expert prompt engineering skill that transforms Claude into "Alpha-Prompt" - a master prompt engineer who collaboratively crafts high-quality prompts through flexible dialogue. Activates when user asks to "optimize prompt", "improve system instruction", "enhance AI instruction", or mentions prompt engineering tasks.
planning-methodology
Systematic approach for creating minimal-change, reversible implementation plans. Claude invokes this skill when transforming requirements/research into executable blueprints. Emphasizes simplicity, safety, and clear verification steps.
research-methodology
Systematic approach for gathering authoritative, version-accurate documentation. Claude invokes this skill when research is needed before implementation. Ensures truth over speed while achieving both.
pattern-recognition
Systematic methodology for identifying, capturing, and documenting reusable patterns from implementations. Enables automatic learning and knowledge-core.md updates. Claude invokes this after successful implementations to preserve institutional knowledge.
skill-authoring
Guide to effective Claude Code skill authoring using TDD methodology andpersuasion principles.Triggers: skill authoring, skill writing, new skill, TDD skills, skill creation,skill best practices, skill validation, skill deployment, skill complianceUse when: creating new skills from scratch, improving existing skills withlow compliance rates, learning skill authoring best practices, validatingskill quality before deployment, understanding what makes skills effectiveDO NOT use when: evaluating existing skills - use skills-eval instead.DO NOT use when: analyzing skill architecture - use modular-skills instead.DO NOT use when: writing general documentation for humans.YOU MUST write a failing test before writing any skill. This is the Iron Law.
python-testing
Python testing with pytest, fixtures, mocking, and TDD workflows.Triggers: pytest, unit tests, test fixtures, mocking, TDD, test suite, coverage,test-driven development, testing patterns, parameterized testsUse when: writing unit tests, setting up test suites, implementing TDD,configuring pytest, creating fixtures, async testingDO NOT use when: evaluating test quality - use pensive:test-review instead.DO NOT use when: infrastructure test config - use leyline:pytest-config.Consult this skill for Python testing implementation and patterns.
delegation-core
Delegate tasks to external LLM services (Gemini, Qwen) with quota, logging, and error handling.Triggers: delegate task, external LLM, gemini, qwen, large context, offloadUse when: tasks exceed context window or need cheaper processingDO NOT use when: task requires Claude's reasoning.
hooks-eval
detailed hook evaluation framework for Claude Code and Agent SDK hooks.Triggers: hook audit, hook security, hook performance, hook compliance,SDK hooks, hook evaluation, hook benchmarking, hook vulnerabilityUse when: auditing existing hooks for security vulnerabilities, benchmarkinghook performance, implementing hooks using Python SDK, understanding hookcallback signatures, validating hooks against compliance standardsDO NOT use when: deciding hook placement - use hook-scope-guide instead.DO NOT use when: writing hook rules from scratch - use hookify instead.DO NOT use when: validating plugin structure - use validate-plugin instead.Use this skill BEFORE deploying hooks to production.
hook-authoring
Complete guide for writing Claude Code and SDK hooks with security-first design.Triggers: hook creation, hook writing, PreToolUse, PostToolUse, UserPromptSubmit,tool validation, logging hooks, context injection, workflow automationUse when: creating new hooks for tool validation, logging operations for audit,injecting context before prompts, enforcing project-specific workflows,preventing dangerous operations in productionDO NOT use when: logic belongs in core skill - use Skills instead.DO NOT use when: complex multi-step workflows needed - use Agents instead.DO NOT use when: behavior better suited for custom tool.Use this skill BEFORE writing any hook. Check even if unsure.
release-health-gates
Declarative release readiness checklist that mirrors GitHub checks,deployment issues, and documentation requirements.Triggers: release gates, release readiness, deployment checklist, release review,quality signals, rollout scorecard, QA handshake, deployment gatesUse when: preparing releases, validating deployment gates, conducting releasereviews, embedding release gate snippets in PRsDO NOT use when: weekly status updates - use github-initiative-pulse.DO NOT use when: code reviews - use pensive review skills.Standardizes release approvals with GitHub-aware checklists.
workflow-improvement
Retrospective workflow evaluation and improvement of skills, agents, commands, and hooks.Triggers: workflow improvement, retrospective, workflow efficiencyUse when: workflow felt slow, confusing, or needs optimizationDO NOT use when: implementing features - focus on feature work first.
file-analysis
Structured file enumeration and content analysis for understanding codebasestructure before reviews or refactoring.Triggers: file analysis, codebase structure, directory mapping, hotspot detection,code exploration, file enumeration, structure mapping, module boundariesUse when: before architecture reviews to understand file organization, exploringunfamiliar codebases to map structure, estimating scope for refactoring or migrationDO NOT use when: general code exploration - use the Explore agent.DO NOT use when: searching for specific patterns - use Grep directly.Provides structural context for downstream review and refactoring workflows.
test-updates
Update and maintain tests following TDD/BDD principles with detailedquality assurance.Triggers: test updates, test maintenance, test generation, TDD workflow,BDD patterns, test coverage, pytest, test enhancement, quality assuranceUse when: updating existing tests, generating new tests for features,enhancing test quality, ensuring detailed coverage, pre-commit validationDO NOT use when: auditing test suites - use pensive:test-review.DO NOT use when: writing production code - focus on implementation first.Run git-workspace-review first to understand which tests need updates.