polaris-local-forge
Snowflake-Labs/polaris-local-forge**[REQUIRED]** Use for **ALL** requests involving local Apache Polaris: setup, API queries, catalog operations, cleanup, teardown. **AUTO-ACTIVATE:** If `.snow-utils/snow-utils-manifest.md` contains `polaris-local-forge:` this skill MUST handle ALL operations including cleanup. **DO NOT** use `polaris` CLI (does not exist), curl to Polaris endpoints (needs OAuth), or docker ps checks - invoke this skill first. Triggers: polaris local, local iceberg catalog, local polaris setup, rustfs setup, create polaris cluster, try polaris locally, get started with polaris, apache polaris quickstart, polaris dev environment, local data lakehouse, replay from manifest, reset polaris catalog, teardown polaris, clean up, cleanup, delete cluster, remove resources, polaris status, list catalogs, show namespaces, list tables, show catalog, describe table, list principals, show principal roles, list views, polaris namespaces, polaris catalogs, query data, query table, query iceberg, query catalog data, show my data, show table data, show records, how many rows, count rows, count records, run sql, run query, duckdb query, select from, group by, aggregate.
SKILL.md
name: polaris-local-forge
description: "[REQUIRED] Use for ALL requests involving local Apache Polaris: setup, API queries, catalog operations, cleanup, teardown. AUTO-ACTIVATE: If .snow-utils/snow-utils-manifest.md contains polaris-local-forge: this skill MUST handle ALL operations including cleanup. DO NOT use polaris CLI (does not exist), curl to Polaris endpoints (needs OAuth), or docker ps checks - invoke this skill first. Triggers: polaris local, local iceberg catalog, local polaris setup, rustfs setup, create polaris cluster, try polaris locally, get started with polaris, apache polaris quickstart, polaris dev environment, local data lakehouse, replay from manifest, reset polaris catalog, teardown polaris, clean up, cleanup, delete cluster, remove resources, polaris status, list catalogs, show namespaces, list tables, show catalog, describe table, list principals, show principal roles, list views, polaris namespaces, polaris catalogs, query data, query table, query iceberg, query catalog data, show my data, show table data, show records, how many rows, count rows, count records, run sql, run query, duckdb query, select from, group by, aggregate."
location: user
Apache Polaris Local Forge
Set up a complete local Apache Polaris development environment with RustFS S3-compatible storage, PostgreSQL metastore, and k3d/k3s Kubernetes — all on your machine.
Query Apache Iceberg tables locally with DuckDB, PyIceberg, or any Iceberg REST-compatible engine.
MANIFEST FILE: .snow-utils/snow-utils-manifest.md (exact path, always .md)
PREREQUISITE: NO PREREQUISITE -- this skill is self-contained infrastructure.
AUTH MODEL:
- Bootstrap credentials: Auto-generated admin credentials for Apache Polaris realm setup (in
.bootstrap-credentials.env) - Principal credentials: API-generated client_id/client_secret for catalog access (in
work/principal.txt) - RustFS credentials: Static
admin/passwordfor S3-compatible storage (no IAM)
SENSITIVE DATA: principal.txt contains realm,client_id,client_secret. When displaying to user, show ONLY the realm. Mask credentials: client_id: ****xxxx (last 4 chars). NEVER show client_secret at all.
Prerequisites
This skill requires the following tools installed on your machine:
| Tool | Purpose | Install |
|---|---|---|
| Podman (default) | Container runtime (OSS) | Already installed - Podman is a dependency of Cortex Code |
| Docker (alternative) | Container runtime | Docker Desktop (>= 4.27) |
| k3d | k3s-in-Docker/Podman | brew install k3d or k3d.io |
| Python | >= 3.12 | python.org |
| uv | Python package manager | curl -LsSf https://astral.sh/uv/install.sh | sh |
Note: Podman is a dependency of Cortex Code and should already be installed on your system. If not present, install via brew install podman or podman.io.
Container Runtime (Auto-Detected): The CLI automatically detects which runtime to use during init:
flowchart TD
Start[init command] --> CheckDockerRunning{Docker Desktop running?}
CheckDockerRunning -->|Yes| UseDocker[Use Docker]
CheckDockerRunning -->|No| CheckPodmanRunning{Podman machine running?}
CheckPodmanRunning -->|Yes| UsePodman[Use Podman]
CheckPodmanRunning -->|No| CheckInstalled{What's installed?}
CheckInstalled -->|Both| PromptUser[Prompt user to choose]
CheckInstalled -->|Podman only| UsePodman2[Use Podman]
CheckInstalled -->|Docker only| UseDocker2[Use Docker]
CheckInstalled -->|Neither| Fail[Fail with error]
Detection priority: Running runtime preferred over installed. Docker preferred when both running. Override by setting PLF_CONTAINER_RUNTIME=docker or podman in .env.
First-time Podman users: Run doctor --fix after init to automatically create and start the Podman machine. See docs/podman-setup.md for manual setup.
Optional:
| Tool | Purpose | Install |
|---|---|---|
| DuckDB CLI | SQL verification | brew install duckdb |
| AWS CLI | S3 bucket operations on RustFS | brew install awscli |
| direnv | Auto-load env vars | brew install direnv |
Workflow
CRITICAL: SETUP TAKES 2-5 MINUTES. Tell user upfront so they don't wait anxiously.
PERMISSIONS: Cortex Code prompts for approval on each new command.
- Before
init: Select "Allow using 'uv' for this session" - After
init: Select "Allow using 'plf' for this session" (or option 5)
This avoids repeated prompts. See Cortex Code Permissions for details.
SETUP APPROACH FOR CORTEX CODE (REQUIRED):
-
First, show the plan:
## Apache Polaris Local Forge Setup (Total: ~2-5 minutes) **Permissions:** Cortex Code prompts for approval on each new command. Select "Allow using 'plf' for this session" (option 5) to run all steps smoothly. **Container runtime:** Auto-detected based on what's running. - Docker Desktop running → uses Docker - Podman machine running → uses Podman - Neither running → will prompt you to choose Steps to complete: - [ ] Step 1: Initialize & detect runtime (~5s) - [ ] Step 2: Doctor check & fix (~5-30s) - [ ] Step 3: Generate configuration files (~10s) - [ ] Step 4: Create k3d cluster (~30s) - [ ] Step 5: Wait for RustFS and PostgreSQL (~60-90s) - [ ] Step 6: Deploy Apache Polaris (~60s) - [ ] Step 7: Setup catalog (~30s) - [ ] Step 8: Verify with DuckDB (~5s) Ready to proceed with cluster name: "<directory-name>" Say "proceed" to continue, or specify a different cluster name.STOP: Wait for user response. If user says "proceed" or similar, use the default cluster name. If user provides a name, use that instead. Then proceed to Step 1.
-
Then run each step individually with a header message.
FORMATTING: Before each step header, output a blank line. This ensures proper visual separation in Cortex Code.
Step 1: Initialize & detect runtime
NOTE: Before
init, there is no./bin/plfwrapper. Useuv run --project <SKILL_DIR>for these commands.First, detect runtime availability:
uv run --project <SKILL_DIR> polaris-local-forge runtime detect --jsonThis outputs JSON with detection result:
{"status": "detected", "runtime": "docker|podman", ...}- Runtime found, proceed{"status": "choice", "options": ["docker", "podman"], ...}- User choice required{"status": "error", ...}- Neither installed, show error
Based on detection result:
-
If
"status": "detected": Run init (runtime will be auto-set):uv run --project <SKILL_DIR> polaris-local-forge --work-dir <WORK_DIR> init -
If
"status": "choice": Present choice using AskQuestion tool with radio options:Option Label Description dockerUse Docker User will start Docker Desktop manually podmanUse Podman (recommended) Machine will be created/started by doctor --fixThen run init with explicit runtime:
uv run --project <SKILL_DIR> polaris-local-forge --work-dir <WORK_DIR> init --runtime <user_choice>(The
--runtimeflag bypasses the interactive prompt that would "Abort" in non-interactive shells) -
If
"status": "error": Display the error message and stop
After
initcompletes, the./bin/plfwrapper is created. Use it for ALL subsequent commands.
Step 2: Doctor check & fix
Run:
./bin/plf doctor --fixThis automatically:
- Creates Podman machine (
k3d) if using Podman and machine doesn't exist - Starts Podman machine if stopped
- Kills gvproxy if blocking port 19000 (required by RustFS)
- Sets up SSH config for Podman VM access
Step 3: Generate configuration files
Run:
./bin/plf prepare
Step 4: Create k3d cluster
Run:
./bin/plf cluster create
Step 5: Wait for RustFS and PostgreSQL
Run:
./bin/plf cluster wait --tags bootstrap
Step 6: Deploy Apache Polaris
Run:
./bin/plf polaris deployWait:./bin/plf cluster wait --tags polarisRun:./bin/plf polaris bootstrap
Step 7: Setup catalog
Run:
./bin/plf catalog setup
Step 8: Verify with DuckDB
Run:
./bin/plf catalog verify-sqlFor interactive exploration:
./bin/plf catalog explore-sql
All steps complete!
After Step 8 succeeds, output the completion summary (see "Step 4: Summary" section below).
-
After each step completes, briefly confirm success before moving to next step.
-
FORMATTING RULES (critical for Cortex Code display):
- Always output a blank line before headers and code blocks
- Always output a blank line after code blocks before continuing text
- Never concatenate output directly after closing triple backticks
- Use
---horizontal rules between major sections
FORBIDDEN ACTIONS -- NEVER DO THESE:
- NEVER use
sed/awk/bashto edit.envor configuration files -- use the Edit/StrReplace tool - NEVER hardcode credentials in scripts -- always read from
.envorwork/principal.txt - NEVER assume the cluster is running -- always check status with
k3d cluster listbefore operations - NEVER run destructive commands (
cluster delete,polaris purge) without explicit user confirmation - NEVER skip teardown confirmation because user did setup in the same session -- ALWAYS confirm destruction regardless of session context
- NEVER delete
.snow-utils/directory -- this contains the manifest needed for replay/audit. Teardown and cleanup commands always preserve.snow-utils/ - NEVER expose
principal.txtcontents in output -- show only the realm. Mask client_id: show****+ last 4 chars. NEVER show client_secret at all. Example:realm: POLARIS, client_id: ****a1b2, client_secret: ******** - NEVER modify files in the skill directory (
<SKILL_DIR>) --k8s/,polaris-forge-setup/,src/,example-manifests/are read-only source. Only the user's--work-diris writable - NEVER create
.snow-utils/in the skill directory -- ALWAYS ask for a work directory FIRST, then create.snow-utils/there - NEVER start setup (with or without manifest) without a work directory -- the skill repo is READ-ONLY
- NEVER guess or invent CLI commands or options -- ONLY use commands from the CLI Reference tables below (e.g.,
plf cluster statusNOTplf status). If a command fails, run./bin/plf --helpor./bin/plf <group> --helpand use ONLY the commands/options shown there - NEVER run raw
duckdbcommands -- use./bin/plf catalog query --sql "..."for data queries,./bin/plf catalog verify-sqlfor initial setup verification, or./bin/plf catalog explore-sqlfor interactive sessions - NEVER construct DuckDB SQL manually EXCEPT for
catalog query --sqlwhich requires the caller to provide the SQL - NEVER extract credentials from
principal.txtto pass to other commands -- credentials are automatically loaded by CLI commands - NEVER use
uv run --project ... polaris-local-forgeafter init -- ALWAYS use the./bin/plfwrapper script which handles paths, env vars, and suppresses warnings - NEVER use
plfcommands for REST API queries (list catalogs, show namespaces, list tables, etc.) -- use Ansible playbooks instead (see Apache Polaris API Queries) - NEVER look for or try to run a
polarisCLI -- there is NOpolariscommand in this skill. The only CLI isplf(for infrastructure) and Ansible (for API queries) - NEVER run
which polarisorpolaris --help-- thepolarisbinary does not exist
COMMAND ROUTING:
- plf commands: Only for CLI operations listed in CLI Reference (init, prepare, cluster, catalog setup/cleanup, teardown, doctor, verify-sql, etc.)
- REST API queries: For query intents (list/show/describe catalogs, namespaces, tables, principals, roles, views, grants) -- use
./bin/plf api query <endpoint>(see Apache Polaris API Queries) - General chat: Explanations, questions, troubleshooting advice -- respond directly without running commands
PROJECT CONTEXT DETECTION (AUTO-ACTIVATION):
When user is in a directory containing ANY of these files, treat it as an active polaris-local-forge project and auto-activate this skill for ALL operations (setup, queries, cleanup, teardown):
| Context Signal | Check | Meaning |
|---|---|---|
.snow-utils/snow-utils-manifest.md with polaris-local-forge: |
grep -q "polaris-local-forge:" .snow-utils/snow-utils-manifest.md |
This skill is installed (primary) |
.env with PLF_POLARIS_CATALOG_NAME |
grep -q "PLF_POLARIS_CATALOG_NAME" .env |
Polaris config exists (secondary) |
bin/plf wrapper script |
[ -x bin/plf ] |
CLI initialized (tertiary) |
CRITICAL - When context detected:
- For cleanup/teardown requests → ALWAYS present confirmation dialog with options (see Teardown Flow)
- For API queries → Use
./bin/plf api query(see Apache Polaris API Queries) - For setup/replay → Use appropriate flow from this skill
- NEVER offer generic cleanup dialogs - always use
plfcommands with the structured teardown dialog
When context detected AND user asks API-related queries (list/show/describe/what catalogs/namespaces/tables/principals/roles/views/grants):
⚠️ FIRST — CHECK FOR WRITE OPERATIONS (FAIL FAST):
Before doing ANYTHING else, check if user is asking to create, update, delete, add, remove, modify, drop, or grant/revoke. If YES, respond IMMEDIATELY:
⚠️ Write operations are not supported yet.
This skill currently supports **read-only** API queries (GET only):
- list/show catalogs, namespaces, tables, views
- describe/get principals, roles, grants
Create/update/delete operations will be added in a future phase.
DO NOT proceed to construct endpoints or run Ansible for write operations. Stop here.
DO NOT (FORBIDDEN):
- ❌
polarisCLI — DOES NOT EXIST (nopolaris namespaces list, nopolaris --help, nopolaris namespace list) - ❌
curlto any Polaris endpoint — needs OAuth token you don't have (nocurl http://localhost:8181/..., nocurl http://localhost:18181/...) - ❌
jqpiped from curl — same reason, curl won't work without OAuth - ❌
kubectl get ns— "namespaces" means Polaris catalog namespaces, NOT Kubernetes - ❌
docker ps/docker ps --filter "name=polaris"— assume infra is running, don't check - ❌ Port
8181— wrong port, Polaris runs on18181in this skill - ❌ Snowflake disambiguation ("Is this Snowflake or Polaris?") — it's Polaris
- ❌ "No Polaris service detected" fallback — don't give up, use Ansible
- ❌
ansible-playbookdirectly — causes asdf/tool-versions conflict, MUST useuv run - ❌
cd <dir> && ansible-playbook— wrong pattern, useuv run --project - ❌ POST/PUT/DELETE operations — write operations NOT supported, fail fast (see above)
- ❌
-e "method=POST"or any method other than GET — read-only phase - ❌
-e '{"body": ...}'or any JSON body parameter — not supported - ❌ Constructing JSON request bodies — not supported
- ❌ Asking user for JSON input — not supported
DO (REQUIRED) — USE plf api query COMMAND:
./bin/plf api query /api/catalog/v1/polardb/namespaces
./bin/plf api query /api/management/v1/catalogs
./bin/plf api query /api/management/v1/principals
Key points:
- Use
./bin/plf api query <endpoint>— clean, simple, handles all the complexity - Endpoint must start with
/api/...— construct from user intent (see table below) - Only GET is supported (read-only phase)
- Output is JSON by default, use
-o textfor plain text - Use
-vfor verbose Ansible output if debugging
Quick context check (run silently before API queries):
grep -q "polaris-local-forge:" .snow-utils/snow-utils-manifest.md 2>/dev/null || grep -q "PLF_POLARIS_CATALOG_NAME" .env 2>/dev/null || [ -x bin/plf ]
API QUERY ROUTING (CRITICAL): When user asks about Apache Polaris data (catalogs, namespaces, tables, views, principals, roles, grants):
- Do NOT search for CLI tools (
polaris,plf query, etc.) -- they don't exist - Do NOT try
curlto any endpoint -- it won't work (needs OAuth token) - Do NOT run
kubectlcommands to check services or namespaces -- this is NOT Kubernetes - Do NOT check if Polaris is running -- assume the setup is complete
- Go DIRECTLY to
./bin/plf api query <endpoint>with the constructed endpoint path
DISAMBIGUATION -- "namespaces" = Polaris catalog namespaces (NOT Kubernetes):
When user says "namespaces", "list namespaces", "show namespaces" → this means Apache Polaris catalog namespaces, accessed via /api/catalog/v1/{catalog}/namespaces. Do NOT interpret as Kubernetes namespaces.
INTERACTIVE PRINCIPLE: This skill is designed to be interactive. At every decision point, ASK the user and WAIT for their response before proceeding. When presenting choices (e.g., runtime selection, execution mode), use the AskQuestion tool with radio options for better UX in Cortex Code.
EXCEPTION: Once user selects "Run all (recommended)" at the config review step, the skill switches to autonomous mode for that setup flow. No further confirmation prompts until setup completes or an error occurs. Track this as EXECUTION_MODE=run_all.
DISPLAY PRINCIPLE: When showing configuration or status, substitute actual values from .env and the manifest. The user should see real values, not raw ${...} placeholders.
OUTPUT PRINCIPLE: Terminal output gets collapsed/truncated in Cortex Code UI. When running --dry-run or any diagnostic command, you MUST copy the relevant output into your response. The user CANNOT see collapsed terminal output. Format with proper code blocks: ```text for summaries, ```bash for commands.
RESILIENCE PRINCIPLE: Always update the manifest IMMEDIATELY after each resource creation step, not in batches. This ensures Resume Flow can recover from any interruption.
Pattern:
1. Set overall Status: IN_PROGRESS at START of resource creation
2. Update each resource row to DONE immediately after creation
3. Set Status: COMPLETE only at the END when ALL resources verified
If user aborts mid-flow, the manifest preserves progress:
- Overall Status stays IN_PROGRESS
- Completed resources show DONE
- Pending resources show PENDING/REMOVED
- Resume Flow picks up from first non-DONE resource
IDEMPOTENCY PRINCIPLE: Before editing any file, CHECK if the change is already applied.
Pattern for manifest updates:
grep -q "Status.*COMPLETE" .snow-utils/snow-utils-manifest.md && echo "Already complete" || echo "Needs update"
Pattern for file edits:
1. Read current file state
2. Check if desired content already exists
3. Only edit if change is needed
4. Skip with message: "Already applied: [description]"
ENVIRONMENT REQUIREMENT: The skill uses local RustFS for S3-compatible storage. AWS CLI commands target http://localhost:19000 with static credentials. No real AWS account is needed.
Note: This skill configures Apache Polaris with local RustFS (S3-compatible) storage only. For real AWS S3 support, see Phase 2 in SKILL_README.md.
CRITICAL RULE - ALL OPERATIONS VIA CLI:
NEVER use direct
podman,docker,k3d,kubectl, orhelmcommands. ALL operations MUST go throughpolaris-local-forgeCLI commands. Direct commands cause port binding issues (gvproxy) and stale state. See "FORBIDDEN COMMANDS" section in CLI Reference for details.
Pre-Check Rules (Fail Fast):
| Command | Pre-Check | If Fails |
|---|---|---|
| Any command | NOT using direct podman/k3d/kubectl | Stop: "Use CLI commands only. See FORBIDDEN COMMANDS." |
init |
Docker or Podman installed | Auto-detects runtime; prompts user if both installed but neither running; fails if neither installed |
doctor |
Tools installed | Report missing tools with install instructions |
doctor |
No ghost clusters | Report stale Docker resources if found |
doctor --fix |
(none - fixes issues) | Creates Podman machine if missing; starts if stopped; kills gvproxy if blocking port 19000; sets up SSH config; cleans up ghost clusters |
cluster create |
Cluster doesn't exist | Stop: "Cluster already exists. Use cluster delete first or setup to resume." |
cluster create |
No ghost cluster | Stop: "Ghost cluster detected. Use --force to clean up stale references." |
cluster create |
Ports 19000, 19001, 18181 available | Stop: "Port in use. Run doctor --fix first." |
catalog setup |
Cluster running, Apache Polaris ready | Stop: "Cluster not ready. Run setup first." |
polaris deploy |
Cluster running | Stop: "Cluster not running. Run cluster create first." |
polaris purge |
Apache Polaris deployed | Stop: "Apache Polaris not deployed. Nothing to purge." |
teardown |
Any resources exist | Proceed gracefully (idempotent with --yes) |
catalog cleanup |
Catalog exists | Proceed gracefully (idempotent with --yes) |
Step 0: Initialize Project Directory
FIRST: Check if in skill directory (NEVER initialize here):
if [ -f "SKILL.md" ] && [ -d "src/polaris_local_forge" ]; then
IN_SKILL_DIR="true"
else
IN_SKILL_DIR="false"
fi
If IN_SKILL_DIR is true, show:
⚠️ You're in the polaris-local-forge skill directory.
This directory contains the source code and is READ-ONLY.
All workspace files must be created in a separate directory.
Where would you like to create your Apache Polaris workspace?
1. ~/polaris-dev (recommended)
2. Specify custom path: ___________
STOP: Wait for user input. NEVER offer "current directory" as an option when in skill directory.
After user provides work directory:
- Create the directory if needed:
mkdir -p <path> - Change to that directory using Shell
working_directoryparameter for subsequent commands - PRESERVE any prior user choices (e.g., if user already selected "Run all", continue in autonomous mode)
If NOT in skill directory, detect if user already has a workspace set up:
if [ -f .env ] && [ -f pyproject.toml ]; then
echo "Existing workspace detected: $(pwd)"
[ -d .snow-utils ] && echo " Found: .snow-utils/"
fi
If existing workspace detected -> go to Step 0a (Prerequisites Check).
If NOT in an existing workspace (and NOT in skill directory), ask user:
Where would you like to create your Apache Polaris workspace?
Options:
1. Use current directory: $(pwd)
2. Create a new directory (e.g., polaris-dev)
STOP: Wait for user input.
Initialize the workspace with the CLI:
uv run --project <SKILL_DIR> polaris-local-forge --work-dir . init
This command:
- Copies template files (
.env,.envrc,.gitignore) - Creates required directories (
.kube/,work/,bin/,k8s/) - Sets
PROJECT_HOME,K3D_CLUSTER_NAME, andSKILL_DIRin.env - Creates
bin/plfwrapper script for simplified CLI invocation
To use a custom cluster name:
uv run --project <SKILL_DIR> polaris-local-forge --work-dir . init --cluster-name my-cluster
IMPORTANT: All subsequent CLI commands use
--work-dirto point generated files here. The skill directory (<SKILL_DIR>) stays read-only. For a second cluster, create another directory and re-run the skill.
Use the wrapper script (created by init):
CRITICAL: After running init, ALWAYS use ./bin/plf for ALL commands. NEVER use uv run --project ... directly - the wrapper handles everything:
./bin/plf doctor
./bin/plf cluster create
./bin/plf catalog verify-sql
The wrapper sources .env automatically and handles all paths.
IMPORTANT - Command Execution: Use Shell tool's working_directory parameter, NOT cd &&:
Shell(command="./bin/plf <COMMAND>", working_directory="<PROJECT_HOME>")
Step 0a: Configuration Review and Confirmation
Display current configuration for user review:
Configuration Review
────────────────────
Config file: .env
Work directory: $(pwd)
PLF_CONTAINER_RUNTIME: ${PLF_CONTAINER_RUNTIME} # ADAPT: podman (default) or docker
PLF_PODMAN_MACHINE: ${PLF_PODMAN_MACHINE} # macOS only (default: k3d)
K3D_CLUSTER_NAME: ${K3D_CLUSTER_NAME} # ADAPT: defaults to project directory name
K3S_VERSION: ${K3S_VERSION}
KUBECONFIG: .kube/config
AWS_ENDPOINT_URL: http://localhost:19000
AWS_REGION: us-east-1
AWS_ACCESS_KEY_ID: admin
POLARIS_URL: http://localhost:18181
POLARIS_REALM: ${POLARIS_REALM} # ADAPT: customizable
PLF_POLARIS_S3_BUCKET: ${PLF_POLARIS_S3_BUCKET} # ADAPT: customizable
PLF_POLARIS_CATALOG_NAME: ${PLF_POLARIS_CATALOG_NAME} # ADAPT: customizable
PLF_POLARIS_PRINCIPAL_NAME: ${PLF_POLARIS_PRINCIPAL_NAME} # ADAPT: customizable
Review the configuration above. Would you like to change any values?
1. Accept all (recommended for first-time setup)
2. Edit a specific value
3. Cancel
STOP: Wait for user input.
If user selects "2. Edit a specific value":
-
Ask: "Which value to edit? (e.g., K3D_CLUSTER_NAME)"
-
User provides the variable name
-
Ask: "Enter new value for {variable}:"
-
Update
.envusing StrReplace tool:StrReplace(.env, "{VARIABLE}=old-value", "{VARIABLE}=new-value") -
Re-display configuration with updated value
-
Return to confirmation prompt (user can edit more or accept)
For PLF_PODMAN_MACHINE edits (macOS only):
If user wants to use a different Podman machine, list available machines:
podman machine ls --format "{{.Name}}"
Display options:
Podman Machine Selection:
0. Create new 'k3d' machine (recommended for isolation)
1. k3d (if exists)
2. podman-machine-default
3. [other existing machines...]
Enter machine name or number:
If user selects "0. Create new":
podman machine init k3d --cpus 4 --memory 16384 --now
podman system connection default k3d
Then run ./bin/plf doctor --fix to configure SSH access.
After confirmation, proceed to prerequisites check.
Step 0b: Prerequisites Check
Check manifest for cached tool verification:
grep "^tools_verified:" .snow-utils/snow-utils-manifest.md 2>/dev/null
If tools_verified: exists with a date: Skip tool checks, continue to Step 0c.
Otherwise, run prerequisite check:
./bin/plf doctor
The doctor command checks:
- Required tools: podman/docker, k3d, kubectl, uv
- Podman machine status (macOS only)
- SSH config for Podman VM access (macOS only)
- Port availability: 19000 (RustFS), 19001 (RustFS Console), 18181 (Apache Polaris), 6443 (k3d API)
With --fix flag (macOS with Podman):
- Creates Podman machine (
k3d) if it doesn't exist - Starts Podman machine if stopped
- Kills gvproxy if blocking port 19000 (required by RustFS)
- Sets up SSH config for Podman VM access
Note: doctor --fix is now fully automatic - it creates and configures the Podman machine without manual intervention. See docs/podman-setup.md for advanced configuration options.
If any tool is missing, stop and provide installation instructions from the Prerequisites table above.
STOP: Do not proceed until all prerequisites pass.
After all tools verified, update manifest:
grep -q "^tools_verified:" .snow-utils/snow-utils-manifest.md 2>/dev/null || \
echo "tools_verified: $(date +%Y-%m-%d)" >> .snow-utils/snow-utils-manifest.md 2>/dev/null || true
Step 0c: Detect or Initialize Manifest
CRITICAL: Work Directory Check
BEFORE any manifest operations, verify we are NOT in the skill directory:
if [ -f "SKILL.md" ] && [ -d "src/polaris_local_forge" ]; then
echo "ERROR: Currently in skill directory (read-only source)"
echo " Cannot create .snow-utils/ here"
exit 1
fi
If in skill directory, STOP and ask:
⚠️ You're in the skill source directory (read-only).
Where would you like to create your Apache Polaris workspace?
1. Create new directory: ~/polaris-dev
2. Specify custom path: ___________
Enter choice [1]:
STOP: Wait for user input. NEVER proceed with manifest operations in skill directory.
After user provides work directory:
- Create the directory if needed:
mkdir -p <path> - Change to that directory: Use Shell
working_directoryparameter - THEN proceed with manifest detection below
Remote Manifest URL Detection
If the user provides a URL (in their prompt or pasted), detect and normalize it before local manifest detection:
Supported URL patterns and translation rules:
- GitHub blob:
https://github.com/{owner}/{repo}/blob/{branch}/{path}-> replace host withraw.githubusercontent.comand remove/blob/segment - GitHub raw:
https://raw.githubusercontent.com/...-> use as-is - GitHub gist:
https://gist.github.com/{user}/{id}-> append/rawif not already present - Any other HTTPS URL ending in
.md-> use as-is
After translating, show user and confirm:
Found manifest URL. Download URL:
<translated_raw_url>
Download to current directory as <filename>? [yes/no]
STOP: Wait for user confirmation.
If yes:
curl -fSL -o <filename> "<translated_raw_url>"
After successful download, continue with local manifest detection below.
Check for existing manifest:
WORKING_MANIFEST=""
SHARED_MANIFEST=""
if [ -f .snow-utils/snow-utils-manifest.md ]; then
WORKING_MANIFEST="EXISTS"
WORKING_STATUS=$(grep "^\*\*Status:\*\*" .snow-utils/snow-utils-manifest.md | head -1)
echo "Working manifest: .snow-utils/snow-utils-manifest.md (${WORKING_STATUS})"
fi
for f in *-manifest.md; do
[ -f "$f" ] && grep -q "## shared_info\|CORTEX_CODE_INSTRUCTION" "$f" 2>/dev/null && {
SHARED_MANIFEST="$f"
echo "Shared manifest: $f"
}
done
Decision matrix:
| Working Manifest | Shared Manifest | Action |
|---|---|---|
| None | None | Fresh start -> Step 0e |
| None | Exists | Copy shared to .snow-utils/ -> Step 0d |
| Exists (REMOVED) | None | Replay Flow (reuse existing config) |
| Exists (COMPLETE) | None | Ask user: re-run, reset, or skip |
| Exists (IN_PROGRESS) | None | Resume Flow -- run pending steps from CLI Reference |
| Exists | Exists | Conflict -- ask user which to use |
If BOTH manifests exist, show:
Found two manifests:
1. Working manifest: .snow-utils/snow-utils-manifest.md
Status: <WORKING_STATUS>
2. Shared manifest: <SHARED_MANIFEST>
(contains resource definitions from another setup)
Which manifest should we use?
A. Resume working manifest (continue where you left off)
B. Start fresh from shared manifest (backup working, adapt values)
C. Cancel
STOP: Wait for user choice.
If using example manifest (user says "get started with apache polaris using example manifest"):
NOTE: This assumes the Work Directory Check above has passed. The example manifest is copied FROM the skill's
example-manifests/directory TO the user's work directory.
# In user's work directory (NOT skill directory!)
mkdir -p .snow-utils && chmod 700 .snow-utils
cp <SKILL_DIR>/example-manifests/polaris-local-forge-manifest.md .snow-utils/snow-utils-manifest.md
chmod 600 .snow-utils/snow-utils-manifest.md
Then proceed to Step 0d to check for adaptive markers.
Step 0d: Shared Manifest Adapt-Check
ALWAYS run this step when using a shared or example manifest. Prompt user ONLY if # ADAPT: markers are found.
ADAPT_COUNT=$(grep -c "# ADAPT:" .snow-utils/snow-utils-manifest.md 2>/dev/null)
echo "ADAPT markers found: ${ADAPT_COUNT}"
If ADAPT_COUNT > 0 (markers found):
Extract all values with # ADAPT: markers and present to user:
Manifest Value Review
─────────────────────
The following values can be customized for your environment:
Setting Default Value Marker
──────────────────────────── ───────────────────────── ──────────────────────
PLF_CONTAINER_RUNTIME: podman # ADAPT: podman or docker
K3D_CLUSTER_NAME: (project directory name) # ADAPT: customizable
POLARIS_REALM: POLARIS # ADAPT: customizable
PLF_POLARIS_S3_BUCKET: polaris # ADAPT: customizable
PLF_POLARIS_CATALOG_NAME: polardb # ADAPT: customizable
PLF_POLARIS_PRINCIPAL_NAME: super_user # ADAPT: customizable
KUBECONFIG: (derived from cluster name)
KUBECTL_PATH: (derived from cluster name)
Options:
1. Accept all defaults (recommended for first-time setup)
2. Edit a specific value
3. Cancel
STOP: Wait for user choice.
| Choice | Action |
|---|---|
| 1 -- Accept all | Proceed with defaults |
| 2 -- Edit specific | Ask which value, update manifest and .env in-place, re-display |
| 3 -- Cancel | Stop |
If user changes PLF_CONTAINER_RUNTIME: Update .env with the new value. If switching to Docker, clear PLF_PODMAN_MACHINE from .env.
If user changes K3D_CLUSTER_NAME: Automatically update derived values (KUBECONFIG, KUBECTL_PATH, resource table row 1) in the manifest. Also update .env with the new cluster name.
If user changes POLARIS_REALM or any PLF_POLARIS_* value: Update both the manifest and .env.
If ADAPT_COUNT = 0 (no markers): Proceed silently with values as-is.
Step 0e: Initialize Manifest
Use the --with-manifest flag to initialize the manifest along with the workspace:
./bin/plf init --with-manifest
This creates .snow-utils/snow-utils-manifest.md with proper permissions (chmod 700 for directory, chmod 600 for file).
If already initialized, you can add the manifest separately:
./bin/plf init --with-manifest --force
Step 1: Environment Setup
SHOW -- what we're about to do:
Set up the lightweight Python environment for querying and exploration. This creates a virtual environment and installs query dependencies (
duckdb,pyiceberg,boto3,pandas) from the workspacepyproject.toml. The infrastructure CLI runs from the skill directory separately.
STOP: Wait for user confirmation before proceeding.
DO:
uv python pin 3.12
uv venv
uv sync --all-extras
SUMMARIZE:
Environment ready. Python venv created with query/notebook dependencies. Configuration loaded from
.env.
Step 2: Full Setup
SHOW -- what we're about to do (tell user this):
I'll set up the complete Apache Polaris environment. This takes 2-5 minutes.
6 steps:
- Pre-check & start Podman (~5s)
- Generate configs (~10s)
- Create k3d cluster (~30s)
- Wait for RustFS & PostgreSQL (~60-90s)
- Deploy Apache Polaris (~60s)
- Setup catalog & verify (~35s)
Services will be available at:
- Apache Polaris API: localhost:18181
- RustFS S3: localhost:19000
- RustFS Console: localhost:19001
Ready to proceed?
STOP: Ask user for execution preference:
How would you like to proceed?
Present using AskQuestion tool with radio options:
| Option | Label | Description |
|---|---|---|
run_all |
Run all (recommended) | Execute all setup steps automatically |
step_by_step |
Step by step | Pause after each step for confirmation |
CRITICAL: Track EXECUTION_MODE for the rest of the session:
EXECUTION_MODE = user's choice ("run_all" or "step_by_step")
Based on user choice:
- If "run_all": Execute ALL steps below without pausing. NO confirmation prompts. Only stop on actual errors.
- If "step_by_step": Run each step, show output, then ask "Continue?" before the next step.
IMPORTANT: Once user selects "run_all", ALL subsequent stopping points in this session are SKIPPED. The skill proceeds automatically through Steps 1-8 without asking for confirmation. This includes:
- Prerequisite checks (auto-proceed if passing)
- Each CLI command execution (run immediately)
- Manifest updates (update immediately)
Only stop for:
- Fatal errors that prevent continuation
- Explicit destructive operations (teardown, purge) that were NOT part of the original setup flow
NOTE: Cortex Code will still show its command approval dialog (platform security feature). Tell user: "Select option 5 ('Allow using plf for this session') on the first
plfcommand to avoid approval prompts for remaining steps."
DO -- Run each step as a SEPARATE command (REQUIRED for Cortex Code visibility):
./bin/plf doctor --fix # Step 1: Pre-check, auto-starts Podman
./bin/plf init --with-manifest # Step 2: Copy templates, create dirs, init manifest
The init --with-manifest command creates the .snow-utils/snow-utils-manifest.md file with the resources tracking table.
Continue with remaining setup steps:
./bin/plf prepare # Step 3: Generate k8s configs
./bin/plf cluster create # Step 4: Create k3d cluster
./bin/plf cluster wait --tags bootstrap # Wait for RustFS + PostgreSQL
./bin/plf polaris deploy # Step 5: Deploy Apache Polaris to cluster
./bin/plf cluster wait --tags polaris # Wait for Apache Polaris + bootstrap job
./bin/plf polaris bootstrap # Step 6: Bootstrap Apache Polaris (create principal)
./bin/plf catalog setup # Step 7: Setup demo catalog
./bin/plf catalog verify-sql # Step 8: Verify with DuckDB
CRITICAL RULES:
- HONOR EXECUTION_MODE: If user selected "run_all", DO NOT ask for confirmation between steps. Proceed automatically.
- Run all mode: Execute full sequence, NO prompts between steps. Just run each command and report results.
- Step by step mode: Pause after each step, ask "Continue? [yes/no]"
- Run EACH command above as a SEPARATE bash invocation (for Cortex Code visibility)
- NEVER combine multiple commands (hides output from Cortex Code)
- NEVER use
podman machine startdirectly (usedoctor --fix) - DESTRUCTIVE COMMANDS (teardown, delete, cleanup, purge): ALWAYS STOP and ask user for explicit confirmation BEFORE running. After user confirms, pass
--yesto skip CLI's interactive prompt (CLI prompts don't work in non-interactive shell) - TEARDOWN AFTER SETUP: Even if user completes setup in this session and then asks to teardown, you MUST still present the teardown confirmation dialog with options. Session context does NOT bypass destruction confirmation. See Teardown Flow.
- DO NOT re-ask work directory, runtime, cluster name, etc. after user has already answered them
After setup completes, set the scoped cluster environment for this session:
export KUBECONFIG="$(pwd)/.kube/config"
export PATH="$(pwd)/bin:$PATH"
set -a && source .env && set +a
Update manifest after each successful step (RESILIENCE PATTERN):
After each step completes successfully, update the manifest resource row status from PENDING to DONE:
# Example: After cluster create succeeds, update row 1
# Use Edit/StrReplace tool to change: | 1 | k3d cluster | infrastructure | PENDING |
# To: | 1 | k3d cluster | infrastructure | DONE |
If interrupted mid-flow, the manifest preserves progress:
- Overall Status stays IN_PROGRESS
- Completed resources show DONE
- Pending resources show PENDING
- Resume Flow picks up from first PENDING resource
After ALL steps complete, update manifest status to COMPLETE:
# Use Edit/StrReplace tool to change: **Status:** IN_PROGRESS
# To: **Status:** COMPLETE
Verify manifest was written:
cat .snow-utils/snow-utils-manifest.md
Expected: all 7 resource rows show DONE, Status shows COMPLETE.
SUMMARIZE:
Setup complete. Cluster
${K3D_CLUSTER_NAME}running with Apache Polaris, RustFS, and PostgreSQL. Catalog${PLF_POLARIS_CATALOG_NAME}created with principal${PLF_POLARIS_PRINCIPAL_NAME}. Credentials saved towork/principal.txt. Manifest:.snow-utils/snow-utils-manifest.md
Step 3: Verification
SHOW -- what we're about to do:
Verify the entire setup by running a DuckDB SQL script that:
- Connects to Apache Polaris REST catalog at
http://localhost:18181- Authenticates using the principal credentials from
work/principal.txt- Creates a test Iceberg table on RustFS
- Queries the table to confirm read/write works
This confirms: Apache Polaris API, RustFS S3, catalog, principal, and RBAC grants are all working.
DO:
./bin/plf catalog verify-sql
Check the result:
- If SUCCEEDED: Continue to Step 3a (generate notebook)
- If FAILED: Check Troubleshooting section. Common issues: Apache Polaris not ready, RustFS not accessible, principal credentials invalid.
SUMMARIZE (on success):
Verification passed! DuckDB successfully queried Iceberg tables via Apache Polaris REST catalog on RustFS.
Step 3a: Generate Notebook
SHOW -- what we're about to do:
Generate a Jupyter notebook for interactive Apache Polaris and Iceberg exploration. The notebook is pre-configured with your catalog connection details and principal credentials, ready to run.
DO:
./bin/plf catalog generate-notebook
SUMMARIZE:
Notebook generated at
notebooks/verify_polaris.ipynb. Open it withjupyter notebook notebooks/verify_polaris.ipynbfor interactive exploration.
Step 4: Summary
SUMMARIZE -- Setup Complete:
Output the following summary (substitute actual values from .env). Remember: blank line before AND after the code block.
Apache Polaris Local Forge -- Setup Complete!
Service URLs:
Apache Polaris API: http://localhost:18181
RustFS S3: http://localhost:19000
RustFS Console: http://localhost:19001
Catalog:
Name: ${PLF_POLARIS_CATALOG_NAME}
S3 Bucket: ${PLF_POLARIS_S3_BUCKET}
Principal: ${PLF_POLARIS_PRINCIPAL_NAME}
Credentials:
RustFS: admin / password
Apache Polaris: See work/principal.txt
Bootstrap: See k8s/polaris/.bootstrap-credentials.env
Next steps:
jupyter notebook notebooks/verify_polaris.ipynb # Interactive exploration
./bin/plf catalog verify-sql # Re-run verification
./bin/plf catalog explore-sql # Interactive DuckDB SQL
Shell setup:
export KUBECONFIG=$(pwd)/.kube/config
export PATH=$(pwd)/bin:$PATH
set -a && source .env && set +a # AWS config for RustFS
kubectl get pods -n polaris
aws s3 ls
Manifest: .snow-utils/snow-utils-manifest.md
Update manifest status to COMPLETE.
Catalog-Only Flows
These flows operate on the catalog without rebuilding the cluster.
Catalog Setup (cluster must be running)
Trigger: "setup catalog only", "create catalog"
./bin/plf catalog setup
Catalog Cleanup
Trigger: "cleanup catalog", "remove catalog"
STOP: Confirm with user before executing.
./bin/plf catalog cleanup --yes
Updates manifest status to REMOVED. Generated files in work/ are preserved.
Catalog Reset
Trigger: "reset polaris catalog", "recreate catalog"
STOP: Confirm with user before executing.
./bin/plf catalog cleanup --yes
./bin/plf catalog setup
Runs cleanup + setup. Generates new principal.txt (new credentials).
Full Catalog Reset
Trigger: "full catalog reset", "purge polaris database"
STOP: This is destructive. Confirm with user.
./bin/plf polaris reset --yes
./bin/plf catalog setup
Purges the entire Apache Polaris database and recreates from scratch.
Teardown Flow
Trigger: "teardown polaris", "delete everything", "clean up", "cleanup", "remove cluster"
CRITICAL - ALWAYS CONFIRM, NO EXCEPTIONS:
Even if user just completed setup:all in this same session, you MUST STOP and confirm before any destruction.
STOP - PRESENT OPTIONS TO USER:
Before running ANY teardown command, present this dialog:
Teardown Options:
1. Cluster only (default)
- Deletes k3d cluster, Polaris, RustFS, PostgreSQL
- Keeps all local files for quick replay
- Use `setup:replay` to restore
2. Cluster + generated files
- Deletes cluster AND cleans generated files
- Use `setup:all` for fresh start
CLEANUP_PATHS: .kube work k8s scripts bin notebooks .env .aws .envrc .gitignore .venv
NOTE: .snow-utils/ is ALWAYS preserved. Replay is possible after BOTH options
unless you manually delete .snow-utils/snow-utils-manifest.md
Which option? [1/2] (or 'cancel' to abort)
Wait for user response - do NOT proceed without explicit selection.
After user selects option:
Option 1 (cluster only):
./bin/plf teardown --yes
Option 2 (cluster + files):
# CLEANUP_PATHS from Taskfile.yml - keep in sync!
./bin/plf teardown --yes
rm -rf .kube work k8s scripts bin notebooks .env .aws .envrc .gitignore .venv
The --yes flag is passed because CLI prompts don't work in non-interactive shell, NOT to skip user confirmation from the agent.
After completion, output EXACTLY this message (DO NOT list individual commands or resources):
Teardown complete.
.snow-utils/ preserved for audit and replay.
Run **replay** anytime to restore.
DO NOT list individual CLI commands. The agent can run replay when user asks.
Updates manifest status to REMOVED.
Podman Machine (macOS)
On macOS with Podman, teardown stops the Podman machine by default when using --yes (fully releases ports 19000, 19001, 18181). To keep Podman running:
./bin/plf teardown --yes --no-stop-podman
Replay Flow
Trigger: "replay polaris local forge", "recreate environment"
When manifest has Status: REMOVED:
- Read config values from manifest
- Show replay plan to user
- On confirmation, execute Steps 1-10
- Reuse existing
.envand work directory layout - Re-activate scoped environment (
KUBECONFIG,PATHfor kubectl) - Regenerate catalog-level files (new principal credentials)
- Update manifest to COMPLETE
Apache Polaris API Queries
EXPERIMENTAL FEATURE: REST API query support via natural language is experimental. Currently supports read-only operations. Write operations (create, delete, update) will be added in a future phase.
AUTO-ACTIVATION: If user's current directory has
.snow-utils/snow-utils-manifest.mdwithpolaris-local-forge:, or.envwithPLF_POLARIS_CATALOG_NAME, this is an active polaris-local-forge project. Skip all disambiguation and useplf api query(see PROJECT CONTEXT DETECTION above).
When to use: User asks to list, show, describe, or query catalogs, namespaces, tables, views, principals, roles, or grants.
IMPORTANT — GO DIRECTLY TO ANSIBLE:
- There is NO
polarisCLI — do NOT runwhich polarisorpolaris --help - There is NO
plf queryorplf polarissubcommand - Do NOT search for CLI tools — go DIRECTLY to the Ansible playbook
- Do NOT try
curlorjq— Ansible handles OAuth and credentials automatically - Construct the endpoint path dynamically based on user intent
How: Use ./bin/plf api query <endpoint> — the CLI handles Ansible, OAuth, and output formatting.
SEMANTIC UNDERSTANDING: Recognize intent variations — these all mean the same thing:
- "list catalogs" = "show me catalogs" = "what catalogs exist" = "get all catalogs"
- "show namespaces" = "list namespaces" = "namespaces in catalog" = "what namespaces"
- "describe table X" = "show table X" = "get table X details" = "table X schema"
CRITICAL DISAMBIGUATION — "namespaces" means POLARIS NAMESPACES:
- When user says "namespaces", "list namespaces", "show namespaces" — this means Apache Polaris catalog namespaces, NOT Kubernetes namespaces
- Do NOT run
kubectl get nsor check Kubernetes namespaces- Do NOT run
curlto any endpoint — Ansible handles this- Do NOT check if Polaris is running or look for services — assume setup is complete
- Go DIRECTLY to Ansible with endpoint
/api/catalog/v1/{catalog}/namespaces
Why Ansible (not curl/jq)
- No extra dependencies — users don't need
jqinstalled - Credentials hidden — OAuth token retrieval uses
no_log: true, never exposed - Consistent pattern — same approach as existing
catalog_setup.yml - JSON output — Ansible returns structured JSON for easy formatting
- Reads .env — Automatically uses user's configured catalog name
Endpoint Construction (Generic Pattern)
The playbook accepts an endpoint parameter. Construct the endpoint path based on user intent:
Management API Endpoints (/api/management/v1/...):
| User Intent | Endpoint Path |
|---|---|
| "list catalogs", "show catalogs" | /api/management/v1/catalogs |
| "show catalog {name}", "describe catalog {name}" | /api/management/v1/catalogs/{name} |
| "list principals", "show principals" | /api/management/v1/principals |
| "show principal {name}", "describe principal {name}" | /api/management/v1/principals/{name} |
| "list principal roles", "show my roles" | /api/management/v1/principal-roles |
| "show principal role {name}" | /api/management/v1/principal-roles/{name} |
| "list roles for principal {name}" | /api/management/v1/principals/{name}/principal-roles |
| "list catalog roles for {catalog}" | /api/management/v1/catalogs/{catalog}/catalog-roles |
| "show catalog role {role} in {catalog}" | /api/management/v1/catalogs/{catalog}/catalog-roles/{role} |
| "list grants for role {role} in {catalog}" | /api/management/v1/catalogs/{catalog}/catalog-roles/{role}/grants |
Catalog API Endpoints (/api/catalog/v1/...):
| User Intent | Endpoint Path |
|---|---|
| "list namespaces", "show namespaces" | /api/catalog/v1/{catalog}/namespaces |
| "show namespace {ns}", "describe namespace {ns}" | /api/catalog/v1/{catalog}/namespaces/{ns} |
| "list tables in {ns}", "show tables" | /api/catalog/v1/{catalog}/namespaces/{ns}/tables |
| "show table {table} in {ns}", "describe table {table}" | /api/catalog/v1/{catalog}/namespaces/{ns}/tables/{table} |
| "list views in {ns}", "show views" | /api/catalog/v1/{catalog}/namespaces/{ns}/views |
| "show view {view} in {ns}", "describe view {view}" | /api/catalog/v1/{catalog}/namespaces/{ns}/views/{view} |
Endpoint Construction Rules
- Parse user intent to identify the operation type (list/show/describe)
- Extract entity names from the request (catalog, namespace, table, view, principal, role)
- Use defaults when not specified:
{catalog}→ use value from.envPLF_POLARIS_CATALOG_NAME(typicallypolardb){ns}→ usedefaultif not specified
- Substitute placeholders in the endpoint path
- Execute with the constructed endpoint
Execution Pattern
Use ./bin/plf api query <endpoint> — the CLI handles Ansible, OAuth, environment, and output formatting:
# List all catalogs
./bin/plf api query /api/management/v1/catalogs
# List namespaces in polardb catalog
./bin/plf api query /api/catalog/v1/polardb/namespaces
# Show table details
./bin/plf api query /api/catalog/v1/polardb/namespaces/default/tables/penguins
# List grants for a catalog role
./bin/plf api query /api/management/v1/catalogs/polardb/catalog-roles/admin/grants
# Text output instead of JSON
./bin/plf api query /api/management/v1/catalogs -o text
# Verbose mode for debugging
./bin/plf api query /api/management/v1/catalogs -v
Use Shell tool with working_directory set to <WORK_DIR>:
Shell(command="./bin/plf api query /api/catalog/v1/polardb/namespaces", working_directory="<WORK_DIR>")
Messaging Template
When running API queries, show clear progress:
**Querying Apache Polaris:** {intent description}
Running: `plf api query {endpoint}`
Example:
**Querying Apache Polaris:** list namespaces in polardb catalog
Running: `plf api query /api/catalog/v1/polardb/namespaces`
Output Formatting
Ansible returns JSON in result.json. Format as markdown table for the user:
**Namespaces in catalog `polardb`:**
| Namespace |
|-----------|
| default |
| analytics |
Export for Sharing Flow
Trigger phrases: "export manifest for sharing", "share polaris manifest"
Purpose: Create a portable copy of the manifest that another developer can use to replay the entire setup on their machine.
Precondition
Cortex Code MUST verify manifest has Status: COMPLETE or Status: REMOVED:
grep "^\*\*Status:\*\*" .snow-utils/snow-utils-manifest.md
If status is IN_PROGRESS, refuse with: "Cannot export incomplete setup. Complete the setup first or teardown to set status to REMOVED."
Export Steps
-
Read active manifest from
.snow-utils/snow-utils-manifest.md -
Read
project_namefrom## project_recipesection -
Determine
shared_byfrom machine username (fall back to asking user):SHARED_BY=$(whoami) echo "Exporting as: ${SHARED_BY}" -
Ask user for export location:
Export manifest for sharing: Filename: {project_name}-manifest.md Default location: ./ (project root) Save to [./]:STOP: Wait for user input.
-
Ask about container runtime:
Container Runtime Export Option ─────────────────────────────── Your setup used: {container_runtime from manifest} How should the exported manifest handle container runtime? 1. Keep current value ({container_runtime}) - recipient reproduces exact environment 2. Clear for auto-detection - recipient's CLI detects their runtime Choice [2]:STOP: Wait for user input. Default is 2 (auto-detect) for maximum flexibility.
-
If file already exists at target: Ask overwrite / rename with timestamp / cancel
-
Create export file with these transformations:
- Inject
<!-- CORTEX_CODE_INSTRUCTION -->HTML comment at top (seeexample-manifests/polaris-local-forge-manifest.mdfor format) - Add
## shared_infosection withshared_by: {SHARED_BY},shared_date,original_project_dir,notes - Preserve
## installed_skillswith skill URL (enables Cortex self-installation) - Change
**Status:** COMPLETEto**Status:** REMOVED - Add
# ADAPT: customizablemarkers on user-customizable values - If user chose option 2: Set
container_runtime: # Auto-detected by CLI - If user chose option 1: Keep
container_runtime: {value} - Remove
podman_machineline (always auto-detected by CLI)
- Inject
-
Show confirmation:
Exported: {project_name}-manifest.md Location: ./{project_name}-manifest.md Shared by: {SHARED_BY} Container runtime: {kept|cleared for auto-detection} Status set to: REMOVED Share this file with your colleague. They can open it in Cursor and ask Cortex Code: "setup from shared manifest"
Note: The exported file is in the project root, NOT in
.snow-utils/. Skills only read from.snow-utils/snow-utils-manifest.mdso the export is invisible to all skill flows.
Consuming Projects: Minimal Setup
A separate project that wants to query the Apache Polaris catalog needs only:
In the project directory:
.envwith:
POLARIS_URL=http://localhost:18181
AWS_ENDPOINT_URL=http://localhost:19000
AWS_ACCESS_KEY_ID=admin
AWS_SECRET_ACCESS_KEY=password
AWS_REGION=us-east-1
# From work/principal.txt in the polaris workspace:
POLARIS_REALM=POLARIS
CLIENT_ID=<from principal.txt>
CLIENT_SECRET=<from principal.txt>
- A notebook (
.ipynb) or SQL scripts for querying pyproject.tomlwith query deps (copyuser-project/pyproject.tomlfrom the skill repo)
NOT needed: k8s manifests, ansible, polaris-local-forge CLI source.
Stopping Points
- Step 0: Ask for workspace directory (if not detected)
- Step 0a: Configuration review -- wait for user confirmation
- Step 0b: If prerequisites missing (show install instructions)
- Step 0c: Manifest detection (ask which to use if conflict)
- Step 0d: Adapt-check (if shared manifest has
# ADAPT:markers) - Step 1: Before environment setup (uv venv)
- Step 2: Before full setup (show 6-step plan, wait for OK)
- Step 3: Before verification (DuckDB test)
- Step 3a: Before generating notebook
- Catalog-only flows: Before cleanup/reset
- Teardown: Before destructive operations
CLI Reference
All commands use the ./bin/plf wrapper script (created by init):
./bin/plf <command>
The wrapper sources .env automatically and handles all paths. Aliased as plf in tables below.
Command Execution
IMPORTANT: Run commands from the project directory using the Shell tool's working_directory parameter:
Shell(command="./bin/plf <COMMAND>", working_directory="<PROJECT_HOME>")
DO NOT use cd && ./bin/plf pattern - use working_directory instead for cleaner output.
The wrapper sources .env (created by init) which contains:
PROJECT_HOME=/path/to/your/project
K3D_CLUSTER_NAME=your-project-name
SKILL_DIR=/path/to/polaris-local-forge
Commands this skill executes:
| Logical Command | Purpose |
|---|---|
plf doctor |
Check prerequisites |
plf doctor --fix |
Fix Podman/SSH issues |
plf init |
Initialize work directory |
plf prepare |
Generate configuration files |
plf cluster create |
Create k3d cluster |
plf cluster wait |
Wait for deployments |
plf polaris deploy |
Deploy Apache Polaris |
plf polaris bootstrap |
Create admin principal |
plf catalog setup |
Setup Iceberg catalog |
plf catalog verify-sql |
Verify with DuckDB |
plf cluster delete --yes |
Delete cluster (destructive) |
plf teardown --yes |
Full cleanup (destructive) |
Recommended permission setting: When Cortex Code prompts for command approval:
Before init (uses uv):
- Option 5: "Allow using 'uv' for this session" - Recommended
After init (uses ./bin/plf wrapper):
- Option 5: "Allow using 'plf' for this session" - Recommended for "Run all" mode
- Option 4: "Always allow using 'plf'" - If you use this skill frequently
TIP: When running "Run all (recommended)" mode, select option 5 on the first
plfcommand to avoid approval prompts for each of the 8+ setup steps.
This is safe because:
- You control the skill source code
- Commands are well-defined CLI operations
uvonly runs Python from your trusted project
Global options (before any subcommand):
| Option | Description |
|---|---|
--work-dir PATH |
Working directory for generated files (default: skill directory) |
--env-file PATH |
Path to .env file (default: <work-dir>/.env) |
OPTION NAMES (NEVER guess or invent options):
ONLY use options listed in the tables below. If a command fails with "No such option", run
./bin/plf <command> --helpto see actual available options and use ONLY those. NEVER invent, abbreviate, or rename options.
--yes flag: Pass --yes ONLY after getting explicit user confirmation from the agent conversation. The flag skips the CLI's interactive prompt (which doesn't work in Cortex Code's non-interactive shell), but the agent MUST still ask the user for confirmation before running the command. All destructive commands support --dry-run to preview.
COMMAND NAMES (exact -- do NOT substitute):
catalog setup-- NOT "catalog create", "catalog init"catalog cleanup-- NOT "catalog delete", "catalog remove"catalog verify-sql-- NOT "catalog verify", "verify-sql"catalog explore-sql-- NOT "catalog explore", "explore-sql"cluster create-- NOT "cluster setup", "cluster init"cluster delete-- NOT "cluster remove", "cluster destroy"
FORBIDDEN FILE EDIT COMMANDS (NEVER use for config files):
CRITICAL: NEVER use
sed,awk, or bash string manipulation to edit.env, manifest files, or any configuration files. Use the Edit/StrReplace tool instead.
| NEVER Use | Use Instead |
|---|---|
sed -i 's/old/new/' .env |
Edit/StrReplace tool |
awk '{...}' file > newfile |
Edit/StrReplace tool |
echo "VAR=value" >> .env |
Edit/StrReplace tool |
cat > .env << EOF |
Write tool (only for new files) |
If ports are blocked (gvproxy error on port 19000):
Port 19000 is required by RustFS. On macOS with Podman, the gvproxy network process may hold this port. The doctor --fix command handles this automatically:
# Automatic fix (recommended)
./bin/plf doctor --fix
This will:
- Kill gvproxy processes blocking port 19000
- Stop and restart the Podman machine
- Set up SSH config for VM access
Manual fix (if automatic fails):
# Only use if doctor --fix doesn't work
pkill -f gvproxy
podman machine stop k3d
sleep 5
podman machine start k3d
./bin/plf doctor --fix # Re-run to verify
After fixing ports, continue setup:
./bin/plf prepare
./bin/plf cluster create
./bin/plf cluster wait --tags bootstrap
./bin/plf polaris deploy
./bin/plf cluster wait --tags polaris
./bin/plf polaris bootstrap
./bin/plf catalog setup
./bin/plf catalog verify-sql
Doctor (Prerequisites Check)
| Command | Description |
|---|---|
plf doctor |
Check prerequisites, tools, Podman machine, ports |
plf doctor --fix |
Auto-fix: create Podman machine if missing, start if stopped, kill gvproxy on port 19000, setup SSH |
plf doctor --output json |
Output status as JSON |
Runtime
| Command | Description |
|---|---|
plf runtime detect |
Detect container runtime (docker/podman/choice) |
plf runtime detect --json |
Output detection result as JSON (for agents) |
plf runtime docker-host |
Output DOCKER_HOST value for current runtime |
Init & Prepare
| Command | Description |
|---|---|
plf init |
Initialize project directory (.env, .envrc, .gitignore, directories) |
plf init --force |
Re-initialize, overwriting existing files |
plf init --runtime docker|podman |
Initialize with explicit runtime (skips interactive prompt) |
plf prepare |
Generate config files from templates (runs ansible prepare.yml) |
plf prepare --tags <tags> |
Run specific ansible tags |
plf prepare --dry-run |
Preview without executing |
Cluster
| Command | Description |
|---|---|
plf cluster create |
Create k3d cluster with RustFS and PostgreSQL (waits for API server) |
plf cluster create --force |
Clean up ghost/stale cluster references before creating |
plf cluster create --wait-timeout N |
Custom API server wait timeout in seconds (default: 120) |
plf cluster create --skip-wait |
Skip API server readiness check (for debugging) |
plf cluster create --dry-run |
Preview cluster creation |
plf cluster delete --yes |
Delete k3d cluster |
plf cluster delete --dry-run |
Preview cluster deletion |
plf cluster wait --tags bootstrap |
Wait for RustFS + PostgreSQL to be ready |
plf cluster wait --tags polaris |
Wait for Apache Polaris deployment + bootstrap job |
plf cluster list |
List k3d clusters |
plf cluster list --output json |
List clusters in JSON format |
plf cluster status |
Show cluster and services status |
plf cluster status --output json |
Show status in JSON format |
Teardown
| Command | Description |
|---|---|
plf teardown |
Complete teardown: catalog cleanup, cluster delete, prompt to stop Podman |
plf teardown --yes |
Teardown with confirmation skipped (stops Podman by default on macOS) |
plf teardown --yes --no-stop-podman |
Teardown but keep Podman running |
plf teardown --dry-run |
Preview teardown operations |
Apache Polaris
| Command | Description |
|---|---|
plf polaris deploy |
Deploy Apache Polaris to cluster |
plf polaris deploy --dry-run |
Preview Apache Polaris deployment |
plf polaris purge |
Delete Apache Polaris deployment |
plf polaris purge --dry-run |
Preview Apache Polaris purge |
plf polaris bootstrap |
Run bootstrap job (create principal and catalog) |
plf polaris bootstrap --dry-run |
Preview bootstrap job |
Catalog
| Command | Description |
|---|---|
plf catalog setup |
Configure Apache Polaris catalog via Ansible |
plf catalog setup --tags <tags> |
Run specific ansible tags |
plf catalog setup --dry-run |
Preview catalog setup |
plf catalog cleanup --yes |
Clean up catalog via Ansible |
plf catalog cleanup --dry-run |
Preview catalog cleanup |
plf catalog verify-sql |
Run DuckDB verification (loads + inserts data) |
plf catalog explore-sql |
Open interactive DuckDB session with catalog pre-loaded |
plf catalog query --sql "..." |
Execute read-only SQL query (no inserts) |
Data Query Triggers (Read-Only)
CRITICAL: When user asks about existing data (counts, stats, queries), use catalog query NOT verify-sql.
verify-sql inserts data every time - running it twice doubles the penguin count!
Trigger phrases: "how many penguins", "count penguins", "penguin count", "query data", "show data"
| User Intent | SQL to Pass |
|---|---|
| "how many penguins we load?" | SELECT COUNT(*) as total_penguins FROM polaris_catalog.wildlife.penguins |
| "show penguins by species" | SELECT species, COUNT(*) as count FROM polaris_catalog.wildlife.penguins GROUP BY species ORDER BY species |
| "show tables" / "list tables" | SHOW ALL TABLES |
| "show island distribution" | SELECT island, COUNT(*) as count FROM polaris_catalog.wildlife.penguins GROUP BY island |
Example command:
./bin/plf catalog query --sql "SELECT COUNT(*) as total_penguins FROM polaris_catalog.wildlife.penguins"
When to use each:
| Command | Use When |
|---|---|
catalog verify-sql |
Initial setup verification (runs once after catalog setup) |
catalog query --sql |
Any subsequent data queries (read-only, safe to run multiple times) |
Status & Verification Commands
Use CLI commands for all status checks and verification:
| Task | CLI Command |
|---|---|
| Check prerequisites | plf doctor |
| Fix prerequisites | plf doctor --fix |
| List clusters | plf cluster list |
| Check cluster & services status | plf cluster status |
| Wait for bootstrap (RustFS + PostgreSQL) | plf cluster wait --tags bootstrap |
| Wait for Apache Polaris | plf cluster wait --tags polaris |
| Verify catalog with DuckDB | plf catalog verify-sql |
| Query existing data (read-only) | plf catalog query --sql "SELECT ..." |
| Interactive SQL exploration | plf catalog explore-sql |
| View config | Read .env file directly |
Logs & Troubleshooting (kubectl allowed for debugging only)
When CLI commands don't provide enough detail for debugging, use kubectl directly for logs and diagnostics:
| Command | Description |
|---|---|
kubectl logs -f -n polaris deployment/polaris |
Stream Apache Polaris logs |
kubectl logs -f -n polaris statefulset/postgresql |
Stream PostgreSQL logs |
kubectl logs -f -n rustfs deployment/rustfs |
Stream RustFS logs |
kubectl get events -n polaris --sort-by='.lastTimestamp' |
Recent Apache Polaris events |
kubectl describe pod -n polaris -l app=polaris |
Diagnose Apache Polaris pod |
Note: For all other operations (status checks, waiting for resources, verification), use the CLI commands above. kubectl is only for log streaming and deep debugging.
Known Limitations
This skill is designed for local development and learning only:
| Limitation | Details |
|---|---|
| Not for production | Single-node cluster, no HA, ephemeral storage |
| macOS/Linux only | No native Windows support (WSL2 may work but untested) |
| Single cluster | One k3d cluster named by K3D_CLUSTER_NAME at a time |
| Local storage | RustFS data stored in Podman/Docker volumes (lost on teardown) |
| No real AWS | Uses RustFS with static credentials, not real AWS S3 |
| Port requirements | Needs ports 19000, 19001, 18181 free |
| Resource limits | Requires ~4GB RAM and ~10GB disk for Podman machine |
What this skill CANNOT do:
- Connect to real AWS S3 (use RustFS for local S3-compatible storage)
- Run multiple Apache Polaris instances
- Provide persistent data across teardown (data is ephemeral)
- Run on ARM Windows or native Windows (Podman/Docker limitation)
Troubleshooting
Always start with CLI status check:
./bin/plf cluster status # Check overall status
./bin/plf cluster wait --tags bootstrap # Wait for RustFS + PostgreSQL
./bin/plf cluster wait --tags polaris # Wait for Apache Polaris
Ghost cluster (cluster create fails with "already exists" but cluster list is empty):
This happens when Docker/Podman has stale resources from a previous cluster that wasn't fully cleaned up (e.g., after switching container runtimes).
# Detect ghost clusters
./bin/plf doctor
# If ghost-cluster check fails, clean up with:
./bin/plf doctor --fix
# Or force cleanup during create:
./bin/plf cluster create --force
API server not ready after cluster creation:
The cluster create command waits for the API server by default (120s timeout). If you see "server could not find the requested resource" errors:
# Wait for API server manually
./bin/plf cluster wait
# Or recreate with longer timeout
./bin/plf cluster delete --yes
./bin/plf cluster create --wait-timeout 180
Apache Polaris pod stuck in ContainerCreating:
# First check status with CLI
./bin/plf cluster status
# If still stuck, use kubectl for deeper debugging
kubectl get events -n polaris --sort-by='.lastTimestamp'
kubectl describe pod -n polaris -l app=polaris
# Then redeploy
./bin/plf polaris deploy
RustFS not accessible:
# Check status first
./bin/plf cluster status
# Then verify S3 access
aws s3 ls --endpoint-url http://localhost:19000
Bootstrap job fails:
# Check logs for debugging
kubectl logs -f -n polaris jobs/polaris-bootstrap
# Purge and re-bootstrap using CLI
./bin/plf polaris purge
./bin/plf polaris deploy
./bin/plf cluster wait --tags polaris
./bin/plf polaris bootstrap
Catalog setup fails (S3 bucket error):
# Verify S3 access
aws s3 ls --endpoint-url http://localhost:19000
# Reset catalog using CLI
./bin/plf catalog cleanup --yes
./bin/plf catalog setup
DuckDB verification fails:
# Use CLI to verify catalog
./bin/plf catalog verify-sql
# If that fails, check credentials
cat work/principal.txt
Official documentation:
- Apache Polaris Documentation
- Apache Polaris Management API Spec
- Apache Polaris Catalog API Spec (Swagger)
- RustFS Documentation
- Apache Iceberg
Security Notes
Credential Storage
- Bootstrap credentials: Generated RSA keys and admin credentials are stored in
k8s/polaris/with restricted permissions (chmod 600) - Principal credentials:
work/principal.txtcontains sensitiveclient_idandclient_secret - RustFS credentials: Static
admin/passwordfor local development only -- not suitable for production use - KUBECONFIG: Scoped to project directory (
.kube/config) to isolate from system kubeconfig - kubectl binary: Downloaded to project
bin/directory to ensure version compatibility with the cluster - .env file: Contains configuration but no secrets by default -- add to
.gitignoreif you add sensitive values - Manifest directory:
.snow-utils/directory uses chmod 700; manifest files use chmod 600 - Network isolation: All services run on localhost ports (18181, 19000, 19001) -- not exposed externally by default
Output Masking (Cortex Code Rules)
When displaying sensitive values to users, Cortex Code MUST mask credentials:
Display Rules for Cortex Code:
| Credential | SHOW | NEVER Show |
|---|---|---|
realm |
Full value | - |
client_id |
Last 4 chars: ****a1b2 |
Full value |
client_secret |
******** only |
ANY part of the value |
| RustFS credentials | admin/password (known static) |
- |
IMPORTANT for automation:
- NEVER echo or print
client_secretvalues directly - When reading
principal.txt, display:realm: POLARIS, client_id: ****a1b2, client_secret: ******** - When reading
.env, mask any password or secret values before displaying to user
Directory Structure
User Workspace (--work-dir)
After skill-based setup, the user's project directory contains:
my-polaris-project/ # User's --work-dir
├── .env # Environment configuration (from .env.example)
├── pyproject.toml # Lightweight query deps (from user-project/)
├── .venv/ # Python virtual environment (uv sync)
├── .snow-utils/
│ └── snow-utils-manifest.md # Resource tracking manifest
├── .kube/
│ └── config # Cluster kubeconfig (chmod 600)
├── bin/
│ └── kubectl # Version-matched kubectl binary
├── k8s/ # Generated + copied k8s manifests
│ ├── features/
│ │ ├── rustfs.yaml # RustFS deployment (copied from skill)
│ │ ├── polaris.yaml # Generated Apache Polaris Helm values
│ │ └── postgresql.yaml # Generated PostgreSQL Helm values
│ └── polaris/
│ ├── kustomization.yaml # Copied from skill
│ ├── polaris-secrets.yaml # Generated secrets
│ ├── .bootstrap-credentials.env
│ ├── .polaris.env
│ ├── rsa_key / rsa_key.pub # RSA key pair
│ └── jobs/ # Bootstrap/purge jobs (copied from skill)
├── work/
│ └── principal.txt # Catalog credentials (chmod 600)
├── notebooks/
│ └── verify_polaris.ipynb # Generated verification notebook
└── scripts/
└── explore_catalog.sql # Generated SQL verification script
Skill Repository (read-only source)
polaris-local-forge/ # SKILL_DIR -- read-only
├── .env.example # Template copied to user workspace
├── config/cluster-config.yaml # k3d cluster configuration
├── k8s/ # Static k8s manifests (source of truth)
├── polaris-forge-setup/ # Ansible playbooks + templates
├── scripts/explore_catalog.py # Python verification script
├── src/polaris_local_forge/ # CLI source
├── user-project/
│ └── pyproject.toml # Lightweight deps template
├── pyproject.toml # Full CLI + infrastructure deps
├── SKILL.md # This file
├── SKILL_README.md # Skills documentation
└── example-manifests/
└── polaris-local-forge-manifest.md