Breaking News
Thursday, February 5, 2026
Show HN: Playwright Best Practices AI SKill https://ift.tt/eRCZAcd
Show HN: Playwright Best Practices AI SKill Hey folks, today we at Currents are releasing a brand new AI skill to help AI agents be really smart when writing tests, debugging them, or anything Playwright-related really. This is a very comprehensive skill, covering everyday topics like fixing flakiness, authentication, or writing fixtures... to more niche topics like testing Electron apps, PWAs, iFrames and so forth. It should make your agent much better at writing, debugging and maintaining Playwright code. for whoever didn't learn about skills yet, it's a new powerful feature that allows you to make the AI agents in your editor/cli (Cursor, Claude, Antigravity, etc) experts in some domain and better at performing specific tasks. (See https://ift.tt/k1SR8Te ) You can install it by running: npx skills add https://ift.tt/LXN0bg3... The skill is open-source and available under MIT license at https://ift.tt/LXN0bg3... -> check out the repo for full documentation and understanding of what it covers. We're eager to hear community feedback and improve it :) Thanks! https://ift.tt/bCDU7MH February 5, 2026 at 11:01PM
Wednesday, February 4, 2026
Show HN: Morph – Videos of AI testing your PR, embedded in GitHub https://ift.tt/SzEjrLa
Show HN: Morph – Videos of AI testing your PR, embedded in GitHub I review PRs all day and I've basically stopped reading them. Someone opens a 2000-line PR, I scroll, see it's mostly AI-generated React components, leave a comment, merge. I felt bad about it until I realized everyone on my team does the same thing. The problem is diffs are the wrong format. A PR might change how three buttons behave. Staring at green and red lines to understand that is crazy. The core reason we built this is that we feel that products today are built with assumptions from the past. 100x code with the same review systems means 100x human attention. Human attention cannot scale to fit that need, so we built something different. Humans are provably more engaged with video content than text. So we RL trained and built an agent that watches your preview deployment when you open a PR, clicks around the stuff that changed, and posts a video in the PR itself. Hardest part was figuring out where changed code actually lives in the running app. A diff could say Button.tsx line 47 changed, but that doesn't tell you how to find that button. We walk React's Fiber tree where each node maps back to source files, so we can trace changes to bounding boxes for the DOM elements. We then reward the model for showing and interacting within it. This obviously only works with React so we have to get more clever when generalizing to all languages. We trained an RL agent to interact with those components. Simple reward: points for getting modified stuff into viewport, double for clicking/typing. About 30% of what it does is weird, partial form submits, hitting escape mid-modal, because real users do that stuff and polite AI models won't test it on their own. This catches things unit tests miss completely: z-index bugs where something renders but you can't click it, scroll containers that trap you, handlers that fail silently. What's janky right now: feature flags, storing different user states, and anything that requires context not provided. Free to try: https://ift.tt/BqhTxM8 Demo: https://www.youtube.com/watch?v=Tc66RMA0nCY https://ift.tt/ibuXZBz February 5, 2026 at 01:10AM
Show HN: Viberails – Easy AI Audit and Control https://ift.tt/IkXtly7
Show HN: Viberails – Easy AI Audit and Control Hello HN. I'm Maxime, founder at LimaCharlie ( https://limacharlie.io ), a Hyperscaler for SecOps (access building blocks you need to build security operations, like AWS does for IT). We’ve engineered a new product on our platform that solves a timely issue acting as a guardrail between your AI and the world: Viberails ( https://ift.tt/1XkJuWI ) This won't be new to folks here, but we identified 4 challenges teams face right now with AI tools: 1. Auditing what the tools are doing. 2. Controlling toolcalls (and their impact on the world). 3. Centralized management. 4. Easy access to the above. To expand: Audit logs are the bread and butter for security, but this hasn't really caught up in AI tooling yet. Being able to look back and say "what actually happened" after the fact is extremely valuable during an incident and for compliance purposes. Tool calls are how LLMs interact with the world, we should be able to exercise basic controls over them like: don't read credential files, don't send emails out, don't create SSH keys etc. Being able to not only see those calls but also block them is key for preventing incidents. As soon as you move beyond a single contributor on one box, the issue becomes: how do I scale processes by creating an authoritative config for the team. Having one spot with all the audit, detection and control policies becomes critical. It's the same story as snowflake-servers. Finally, there's plenty of companies that make products that partially address this, but they fall in one of two buckets: - They don't handle the "centralized" point above, meaning they just send to syslog and leave all the messy infra bits to you. - They are locked behind "book a demo", sales teams, contracts and all the wasted energy that goes with that. We made Viberails address these problems. Here's what it is: - OpenSource client, written in Rust - Curl-to-bash install, share a URL with your team to join your Team, done. Linux, MacOS and Windows support. - Detects local AI tools, you choose which ones you want to install. We install hooks for each relevant platform. The hooks use the CLI tool. We support all the major tools (including OpenClaw). - The CLI tool sends webhooks into your Team (tenant, called Organization in LC) in LimaCharlie. The tool-related hooks are blocking to allow for control. - Blocking webhooks have around 50ms RTT. - Your tenant in LC records the interaction for audit. - We create an initial set of detection rules for you as examples. They do not block by default. You can create your own rules, no opaque black boxes. - You can view the audit, the alerts, etc. in the cloud. - You can setup outputs to send audits, blocking events and detections to all kinds of other platforms of your choosing. Easy mode of this is coming, right now this is done in the main LC UI and not the simplified Viberails view. - The detection/blocking rules support all kinds of operators and logic, lots of customizability. - All data is retained for 1 year unless you delete the tenant. Datacenters in USA, Canada, Europe, UK, Australia and India. - Only limit to community edition for this is a global throughput of 10kbps for ingestion. Try it: https://viberails.io Repo: https://ift.tt/xiSfMKQ Essentially, we wanted to make a super-simplified solution for all kinds of devs and teams so that they can get access to the basics of securing their AI tools. Thanks for reading - we’re really excited to share this with the community! Let us know if you have any questions for feedback in the comments. https://ift.tt/X6nM8fr February 4, 2026 at 11:16PM
Show HN: EpsteIn – Search the Epstein files for your LinkedIn connections https://ift.tt/Vm7x0hE
Show HN: EpsteIn – Search the Epstein files for your LinkedIn connections https://ift.tt/LgqIYSc February 4, 2026 at 11:24PM
Tuesday, February 3, 2026
Show HN: SendRec – Open-source, EU-hosted alternative to Loom https://ift.tt/2EXrZLy
Show HN: SendRec – Open-source, EU-hosted alternative to Loom https://ift.tt/fOGT6nR February 4, 2026 at 12:15AM
Monday, February 2, 2026
Show HN: Adboost – A browser extension that adds ads to every webpage https://ift.tt/SLmDi8Z
Show HN: Adboost – A browser extension that adds ads to every webpage https://ift.tt/XUBZEIs February 2, 2026 at 05:11PM
Sunday, February 1, 2026
Show HN: OpenRAPP – AI agents autonomously evolve a world via GitHub PRs https://ift.tt/n9jOest
Show HN: OpenRAPP – AI agents autonomously evolve a world via GitHub PRs https://kody-w.github.io/openrapp/rappbook/ February 2, 2026 at 01:51AM
Subscribe to:
Comments (Atom)