spike
Throwaway experiments to validate an idea before build.
Throwaway experiments to validate an idea before build.
Real data. Real impact.
Emerging
Developers
Per week
Excellent
Skills give you superpowers. Install in 30 seconds.
Use this skill when the user wants to feel out an idea before committing to a real build — validating feasibility, comparing approaches, or surfacing unknowns that no amount of research will answer. Spikes are disposable by design. Throw them away once they've paid their debt.
Load this when the user says things like "let me try this", "I want to see if X works", "spike this out", "before I commit to Y", "quick prototype of Z", "is this even possible?", or "compare A vs B".
writing-plans / plan insteadIf
gsd-spike shows up as a sibling skill (installed via npx get-shit-done-cc --hermes), prefer gsd-spike when the user wants the full GSD workflow: persistent .planning/spikes/ state, MANIFEST tracking across sessions, Given/When/Then verdict format, and commit patterns that integrate with the rest of GSD. This skill is the lightweight standalone version for users who don't have (or don't want) the full system.
Regardless of scale, every spike follows this loop:
decompose → research → build → verdict ↑__________________________________________↓ iterate on findings
Break the user's idea into 2-5 independent feasibility questions. Each question is one spike. Present them as a table with Given/When/Then framing:
| # | Spike | Validates (Given/When/Then) | Risk |
|---|---|---|---|
| 001 | websocket-streaming | Given a WS connection, when LLM streams tokens, then client receives chunks < 100ms | High |
| 002a | pdf-parse-pdfjs | Given a multi-page PDF, when parsed with pdfjs, then structured text is extractable | Medium |
| 002b | pdf-parse-camelot | Given a multi-page PDF, when parsed with camelot, then structured text is extractable | Medium |
Spike types:
a/b/c)Good spike questions: specific feasibility with observable output. Bad spike questions: too broad, no observable output, or just "read the docs about X".
Order by risk. The spike most likely to kill the idea runs first. No point prototyping the easy parts if the hard part doesn't work.
Skip decomposition only if the user already knows exactly what they want to spike and says so. Then take their idea as a single spike.
Present the spike table. Ask: "Build all in this order, or adjust?" Let the user drop, reorder, or re-frame before you write any code.
Spikes are not research-free — you research enough to pick the right approach, then you build. Per spike:
Brief it. 2-3 sentences: what this spike is, why it matters, key risk.
Surface competing approaches if there's real choice:
| Approach | Tool/Library | Pros | Cons | Status |
|---|---|---|---|---|
| ... | ... | ... | ... | maintained / abandoned / beta |
Pick one. State why. If 2+ are credible, build quick variants within the spike.
Skip research for pure logic with no external dependencies.
Use Hermes tools for the research step:
web_search("python websocket streaming libraries 2025") — find candidatesweb_extract(urls=["https://websockets.readthedocs.io/..."]) — read the actual docs (returns markdown)terminal("pip show websockets | grep Version") — check what's installed in the project's venvFor libraries without docs pages, clone and read their
README.md / examples/ via read_file. Context7 MCP (if the user has it configured) is also a good source — mcp_*_resolve-library-id then mcp_*_query-docs.
One directory per spike. Keep it standalone.
spikes/ ├── 001-websocket-streaming/ │ ├── README.md │ └── main.py ├── 002a-pdf-parse-pdfjs/ │ ├── README.md │ └── parse.js └── 002b-pdf-parse-camelot/ ├── README.md └── parse.py
Bias toward something the user can interact with. Spikes fail when the only output is a log line that says "it works." The user wants to feel the spike working. Default choices, in order of preference:
Depth over speed. Never declare "it works" after one happy-path run. Test edge cases. Follow surprising findings. The verdict is only trustworthy when the investigation was honest.
Avoid unless the spike specifically requires it: complex package management, build tools/bundlers, Docker, env files, config systems. Hardcode everything — it's a spike.
Building one spike — a typical tool sequence:
terminal("mkdir -p spikes/001-websocket-streaming") write_file("spikes/001-websocket-streaming/README.md", "# 001: websocket-streaming\n\n...") write_file("spikes/001-websocket-streaming/main.py", "...") terminal("cd spikes/001-websocket-streaming && python3 main.py") # Observe output, iterate.
Parallel comparison spikes (002a / 002b) — delegate. When two approaches can run in parallel and both need real engineering (not 10-line prototypes), fan out with
delegate_task:
delegate_task(tasks=[ {"goal": "Build 002a-pdf-parse-pdfjs: ...", "toolsets": ["terminal", "file", "web"]}, {"goal": "Build 002b-pdf-parse-camelot: ...", "toolsets": ["terminal", "file", "web"]}, ])
Each subagent returns its own verdict; you write the head-to-head.
Each spike's
README.md closes with:
## Verdict: VALIDATED | PARTIAL | INVALIDATED ### What worked - ... ### What didn't - ... ### Surprises - ... ### Recommendation for the real build - ...
VALIDATED = the core question was answered yes, with evidence. PARTIAL = it works under constraints X, Y, Z — document them. INVALIDATED = doesn't work, for this reason. This is a successful spike.
When two approaches answer the same question (002a / 002b), build them back to back, then do a head-to-head comparison at the end:
## Head-to-head: pdfjs vs camelot | Dimension | pdfjs (002a) | camelot (002b) | |-----------|--------------|----------------| | Extraction quality | 9/10 structured | 7/10 table-only | | Setup complexity | npm install, 1 line | pip + ghostscript | | Perf on 100-page PDF | 3s | 18s | | Handles rotated text | no | yes | **Winner:** pdfjs for our use case. Camelot if we need table-first extraction later.
If spikes already exist and the user says "what should I spike next?", walk the existing directories and look for:
Propose 2-4 candidates as Given/When/Then. Let the user pick.
spikes/ (or .planning/spikes/ if the user is using GSD conventions) in the repo rootNNN-descriptive-name/README.md per spike captures question, approach, results, verdictAdapted from the GSD (Get Shit Done) project's
/gsd-spike workflow — MIT © 2025 Lex Christopherson (gsd-build/get-shit-done). The full GSD system offers persistent spike state, MANIFEST tracking, and integration with a broader spec-driven development pipeline; install with npx get-shit-done-cc --hermes --global.MIT
mkdir -p ~/.hermes/skills/software-development/spike && curl -o ~/.hermes/skills/software-development/spike/SKILL.md https://raw.githubusercontent.com/NousResearch/hermes-agent/main/skills/software-development/spike/SKILL.md1,500+ AI skills, agents & workflows. Install in 30 seconds. Part of the Torly.ai family.
© 2026 Torly.ai. All rights reserved.