Breaking News
Thursday, March 5, 2026
Show HN: Cognitive architecture for Claude Code – triggers, memory, docs https://ift.tt/w3azeZ7
Show HN: Cognitive architecture for Claude Code – triggers, memory, docs This started as a psychology research project (building a psychoemotional safety scoring model) and turned into something more general: a reusable cognitive architecture for long-running AI agent work. The core problem: Claude Code sessions lose context. Memory files live outside the repo and can silently disappear. Design decisions made in Session 3 get forgotten by Session 8. Documentation drifts from reality. Our approach — 12 mechanical triggers that fire at specific moments (before responding, before writing to disk, at phase boundaries, on user pushback). Principles without firing conditions remain aspirations. Principles with triggers become infrastructure. What's interesting: - Cognitive trigger system — T1 through T12 govern agent behavior: anti-sycophancy checks, recommend-against scans, process vs. substance classification, 8-order knock-on analysis before decisions. Not prompting tricks — structural firing conditions. - Self-healing memory — Auto-memory lives outside the git repo. A bootstrap script detects missing/corrupt state, restores from committed snapshots with provenance headers, and reports what happened. The agent's T1 (session start) runs the health check before doing anything else. - Documentation propagation chain — 13-step post-session cycle that pushes changes through 10 overlapping documents at different abstraction levels. Content guards prevent overwriting good state with empty files. Versioned archives at every cycle. - Git reconstruction from chat logs — The project existed before its repo. We rebuilt git history by replaying Write/Edit operations from JSONL transcripts, with a weighted drift score measuring documentation completeness. The divergence report became a documentation coverage report. - Structured decision resolution — 8-order knock-on analysis (certain → likely → possible → speculative → structural → horizon) with severity-tiered depth and consensus-or-parsimony binding. All built on Claude Code with Opus. The cognitive architecture (triggers, skills, memory pattern) transfers to any long-running agent project — the psychology domain is the first application, not a constraint. Design phase — architecture resolved, implementation of the actual psychology agent hasn't started. The infrastructure for building it is the interesting part. Code: https://ift.tt/9tXGuTF Highlights if you want to skip around: - Trigger system: docs/cognitive-triggers-snapshot.md - Bootstrap script: bootstrap-check.sh - Git reconstruction: reconstruction/reconstruct.py - Documentation chain: .claude/skills/cycle/SKILL.md - Decision resolution: .claude/skills/adjudicate/SKILL.md - Research journal: journal.md (the full narrative, 12 sections) Happy to discuss the trigger design, the memory recovery pattern, or why we think documentation propagation matters more than people expect for AI-assisted work. https://ift.tt/9tXGuTF March 5, 2026 at 10:05PM
Wednesday, March 4, 2026
Show HN: Gobble – Yet Another OSS Alternative to Google Analytics/PostHog, etc. https://ift.tt/hmcXVxf
Show HN: Gobble – Yet Another OSS Alternative to Google Analytics/PostHog, etc. https://ift.tt/O8rMZ9l March 5, 2026 at 01:12AM
Show HN: Qlog – grep for logs, but 100x faster https://ift.tt/I9cqWmY
Show HN: Qlog – grep for logs, but 100x faster I built qlog because I got tired of waiting for grep to search through gigabytes of logs. qlog uses an inverted index (like search engines) to search millions of log lines in milliseconds. It's 10-100x faster than grep and way simpler than setting up Elasticsearch. Features: - Lightning fast indexing (1M+ lines/sec using mmap) - Sub-millisecond searches on indexed data - Beautiful terminal output with context lines - Auto-detects JSON, syslog, nginx, apache formats - Zero configuration - Works offline - Pure Python Example: qlog index './logs/*/*.log' qlog search "error" --context 3 I've tested it on 10GB of logs and it's consistently 3750x faster than grep. The index is stored locally so repeated searches are instant. Demo: Run `bash examples/demo.sh` to see it in action. GitHub: https://ift.tt/0u6ZSFj Perfect for developers/DevOps folks who search logs daily. Happy to answer questions! https://ift.tt/0u6ZSFj March 5, 2026 at 12:17AM
Show HN: WooTTY - browser terminal in a single Go binary https://ift.tt/gBLso0e
Show HN: WooTTY - browser terminal in a single Go binary I needed a web terminal I could drop into K8s sidecars and internal tools without pulling in heavy dependencies or running a separate service. Existing options were either too opinionated about the shell or had fragile session handling around reconnects. WooTTY wraps any binary -- bash, ssh, or custom tools -- and serves a browser terminal over HTTP. Sessions survive reconnects via output replay. There's a Resume/Watch distinction so multiple people can attach to the same session without stepping on each other. https://ift.tt/nTG8Fk0 March 4, 2026 at 11:32PM
Tuesday, March 3, 2026
Show HN: Agent Action Protocol (AAP) – MCP got us started, but is insufficient https://ift.tt/tlja3ov
Show HN: Agent Action Protocol (AAP) – MCP got us started, but is insufficient Background: I've been working on agentic guardrails because agents act in expensive/terrible ways and something needs to be able to say "Maybe don't do that" to the agents, but guardrails are almost impossible to enforce with the current way things are built. Context: We keep running into so many problems/limitations today with MCP. It was created so that agents have context on how to act in the world, it wasn't designed to become THE standard rails for agentic behavior. We keep tacking things on to it trying to improve it, but it needs to die a SOAP death so REST can rise in it's place. We need a standard protocol for whenever an agent is taking action. Anywhere. I'm almost certainly the wrong person to design this, but I'm seeing more and more people tack things on to MCP rather than fix the underlying issues. The fastest way to get a good answer is to submit a bad one on the internet. So here I am. I think we need a new protocol. Whether it's AAP or something else, I submit my best effort. Please rip it apart, lets make something better. https://ift.tt/GtCop5h March 3, 2026 at 09:22PM
Show HN: A tool to give every local process a stable URL https://ift.tt/yav96BQ
Show HN: A tool to give every local process a stable URL In working with parallel agents in different worktrees, I found that I had a lot of port conflicts, went back and forth checking what incremented port my dev server was running on, and cookie bleed. This isnt a big issue if running a few servers with full a stack framework like Next, Nuxt, or Sveltekit, but if you run a Rust backend and a Vite frontend In multiple worktrees, it gets way more complicated, and the mental model starts to break. That's not even adding in databases, or caches. So I built Roxy, which is a single Go binary that wraps your dev servers (or any process actually) and gives you a stable .test domain based on the branch name and cwd. It runs a proxy and dns server that handles all the domain routing, tls, port mapping, and forwarding for you. It currently supports: - HTTP for your web apps and APIs - Most TCP connections for your db, cache and message/queue layers - TLS support so you can run HTTPS - Run multiple processes at once each with a unique URL, like Docker compose - Git and worktree awareness - Detached mode - Zero config startup My co-workers and I have been using it a lot with our workflow and i think it's ready for public use. We support MacOS and Linux I'll be working on some more useful features like Docker compose/Procfile compatibility and tunneling so you can access your dev environment remotely with a human readable URL Give it a try, and open an issue if something doesnt quite work right, or to request a feature! https://ift.tt/uSOrK4Y https://ift.tt/uSOrK4Y March 4, 2026 at 12:18AM
Subscribe to:
Comments (Atom)