covid20212022
ads
ads
Saturday, April 18, 2026
Show HN: AI Subroutines – Run automation scripts inside your browser tab https://ift.tt/jJLSZRb
Show HN: AI Subroutines – Run automation scripts inside your browser tab We built AI Subroutines in rtrvr.ai. Record a browser task once, save it as a callable tool, replay it at: zero token cost, zero LLM inference delay, and zero mistakes. The subroutine itself is a deterministic script composed of discovered network calls hitting the site's backend as well as page interactions like click/type/find. The key architectural decision: the script executes inside the webpage itself, not through a proxy, not in a headless worker, not out of process. The script dispatches requests from the tab's execution context, so auth, CSRF, TLS session, and signed headers get added to all requests and propagate for free. No certificate installation, no TLS fingerprint modification, no separate auth stack to maintain. During recording, the extension intercepts network requests (MAIN-world fetch/XHR patch + webRequest fallback). We score and trim ~300 requests down to ~5 based on method, timing relative to DOM events, and origin. Volatile GraphQL operation IDs are detected and force a DOM-only fallback before they break silently on the next run. The generated code combines network calls with DOM actions (click, type, find) in the same function via an rtrvr.* helper namespace. Point the agent at a spreadsheet of 500 rows and with just one LLM call parameters are assigned and 500 Subroutines kicked off. Key use cases: - record sending IG DM, then have reusable and callable routine to send DMs at zero token cost - create routine getting latest products in site catalog, call it to get thousands of products via direct graphql queries - setup routine to file EHR form based on parameters to the tool, AI infers parameters from current page context and calls tool - reuse routine daily to sync outbound messages on LinkedIn/Slack/Gmail to a CRM using a MCP server We see the fundamental reason that browser agents haven't taken off is that for repetitive tasks going through the inference loop is unnecessary. Better to just record once, and get the LLM to generate a script leveraging all the possible ways to interact with a site and the wider web like directly calling backed API's, interacting with the DOM, and calling 3P tools/APIs/MCP servers. https://ift.tt/J5mrUDp April 18, 2026 at 04:03AM
Show HN: Praxis – Lab data to publication-ready figures in one Python package https://ift.tt/itwyvOA
Show HN: Praxis – Lab data to publication-ready figures in one Python package https://ift.tt/u5Nj9xO April 19, 2026 at 01:15AM
Show HN: I built Panda to get up to 99% token savings https://ift.tt/NL73vPK
Show HN: I built Panda to get up to 99% token savings https://ift.tt/dVw9mNM April 18, 2026 at 05:00PM
Friday, April 17, 2026
Show HN: Waputer – The WebAssembly Computer https://ift.tt/nlCwDAr
Show HN: Waputer – The WebAssembly Computer Waputer is an operating system that runs entirely in the browser. When you visit the website at https://waputer.app , a kernel written in JavaScript sets up a filesystem and launches a WebAssembly program, which in turn talks to the kernel to handle the display and input. A purely terminal-based version is at https://waputer.dev . My original intention was to create programs that run in the browser that have a lot more in common with the desktop. The traditional "hello world" program is not really suited for the web. Waputer changes that. The GitHub repo at https://ift.tt/g5z06Up gives a very brief overview of compiling a C program and running it on Waputer. There is a blog available from the main site that has a long-form explanation of Waputer and my motivations if you want some additional reading. https://waputer.app April 18, 2026 at 12:46AM
Show HN: Smol machines – subsecond coldstart, portable virtual machines https://ift.tt/ZBLptF2
Show HN: Smol machines – subsecond coldstart, portable virtual machines https://ift.tt/Ur5cJgS April 18, 2026 at 12:18AM
Show HN: Bird, a CLI for Tired Brains https://ift.tt/3XBzHEO
Show HN: Bird, a CLI for Tired Brains https://ift.tt/hSZ4xpo April 18, 2026 at 12:13AM
Show HN: PanicLock – Close your MacBook lid disable TouchID –> password unlock https://ift.tt/QFPhEV5
Show HN: PanicLock – Close your MacBook lid disable TouchID –> password unlock https://ift.tt/ivusXmS April 17, 2026 at 11:38PM
Thursday, April 16, 2026
Show HN: EDDI – Multi-agent AI engine where agent logic lives in JSON, not code https://ift.tt/FLnKJU5
Show HN: EDDI – Multi-agent AI engine where agent logic lives in JSON, not code I started EDDI in 2006 as a rule-based dialog engine. Back then it was pattern matching and state machines. When LLMs showed up, the interesting question wasn't "how do I call GPT" but "how do I keep control over what the AI does in production?" My answer was: agent logic belongs in JSON configs, not code. You describe what an agent should do, which LLM to use, what tools it can call, how it should behave. The engine reads that config and runs it. No dynamic code execution, ever. The LLM cannot run arbitrary code by design. The engine is strict so the AI can be creative. v6 is the version where this actually became practical. You can have groups of agents debating a topic in five different orchestration styles (round table, peer review, devil's advocate...). Each agent can use a different model. A cascading system tries cheap models first and only escalates to expensive ones when confidence is low. It also implements MCP as both server and client, so you can control EDDI from Claude Desktop or Cursor. And Google's A2A protocol for agents discovering each other across platforms. The whole thing runs in Java 25 on Quarkus, ships as a single Docker image, and installs with one command. Open source since 2017, Apache 2.0. Would love to hear thoughts on the architecture and feature set. And if you have ideas for what's missing or what you'd want from a system like this, I'm all ears. Always looking for good input on the roadmap. https://ift.tt/Rp83Xwo April 16, 2026 at 09:11PM
Show HN: CodeBurn – Analyze Claude Code token usage by task https://ift.tt/4pnvwDZ
Show HN: CodeBurn – Analyze Claude Code token usage by task Built this after realizing I was spending ~$1400/week on Claude Code with almost no visibility into what was actually consuming tokens. Tools like ccusage give a cost breakdown per model and per day, but I wanted to understand usage at the task level. CodeBurn reads the JSONL session transcripts that Claude Code stores locally (~/.claude/projects/) and classifies each turn into 13 categories based on tool usage patterns (no LLM calls involved). One surprising result: about 56% of my spend was on conversation turns with no tool usage. Actual coding (edits/writes) was only ~21%. The interface is an interactive terminal UI built with Ink (React for terminals), with gradient bar charts, responsive panels, and keyboard navigation. There’s also a SwiftBar menu bar integration for macOS. Happy to hear feedback or ideas. https://ift.tt/dbt8nS1 April 14, 2026 at 05:57AM
Wednesday, April 15, 2026
Show HN: Dependicus, a dashboard for your monorepo's dependencies https://ift.tt/AOLem6p
Show HN: Dependicus, a dashboard for your monorepo's dependencies Late last year, I was digging into some dependency-related tech debt, and struggling with how long it takes to run pnpm's introspection commands like 'pnpm why' in a medium-size monorepo. So I started working on a simple static site generator that would let me view the output of these expensive commands all at once, to make problems clearly visible instead of requiring deep exploration one at a time. Once I had that working, I realized I had enough data to add ticket tracking. It uses the data it gathers from the package manager to keep Linear or GitHub issues updated. And by auto-assigning those issues to coding agents, I get a Dependabot-but-better experience: agents keep up with API updates in addition to just bumping versions, and group related updates automatically. It's still early days, but it's working really well for us and I think people will find value in it, so I'm sharing here! https://descriptinc.github.io/dependicus/ April 16, 2026 at 12:02AM
Show HN: MCP server gives your agent a budget (save tokens, get smarter results) https://ift.tt/zD7ofRZ
Show HN: MCP server gives your agent a budget (save tokens, get smarter results) As a consultant I foot my own Cursor bills, and last month was $1,263. Opus is too good not to use, but there's no way to cap spending per session. After blowing through my Ultra limit, I realized how token-hungry Cursor + Opus really is. It spins up sub-agents, balloons the context window, and suddenly, a task I expected to cost $2 comes back at $8. My bill kept going up, but was I really going to switch to a worse model? No. So I built l6e: an MCP server that gives your agent the ability to budget. It works with Cursor, Claude Code, Windsurf, Openclaw, and every MCP-compatible application. Saving money was why I built it, but what surprised me was that the process of budgeting changed the agent's behavior. An agent that understands the limitations of the resources doesn't try to speculatively increase the context window with extra files. It doesn't try to reach every possible API. The agent plans ahead, sticks to it, and ends work when it should. It works, and we've been dogfooding it hard. After v1 shipped, the rest of l6e was all built with it. We launched the entire docs site using frontier models for $0.99. The kicker was every time l6e broke in development, I could feel the pain. The agent got sloppy, burned through context, and output quality dropped right along with it. Install: pip install l6e-mcp Docs: https://docs.l6e.ai GitHub: https://ift.tt/Ze3Aaxg Website: https://l6e.ai Happy to answer questions about the system design, calibration models, or why I can't go back to coding without it. https://l6e.ai April 15, 2026 at 10:38PM
Tuesday, April 14, 2026
Show HN: A Claude Code–driven tutor for learning algorithms in Go https://ift.tt/wTClJbg
Show HN: A Claude Code–driven tutor for learning algorithms in Go https://ift.tt/9o8brjG April 15, 2026 at 12:41AM
Show HN: LangAlpha – what if Claude Code was built for Wall Street? https://ift.tt/nqNr1Zw
Show HN: LangAlpha – what if Claude Code was built for Wall Street? Some technical context on what we ran into building this. MCP tools don't really work for financial data at scale. One tool call for five years of daily prices dumps tens of thousands of tokens into the context window. And data vendors pack dozens of tools into a single MCP server, schemas alone can eat 50k+ tokens before the agent does anything useful. So we auto-generate typed Python modules from the MCP schemas at workspace init and upload them into the sandbox. The agent just imports them like a normal library. Only a one-line summary per server stays in the prompt. We have around 80 tools across our servers and the prompt cost is the same whether a server has 3 tools or 30. This part isn't finance-specific, it works with any MCP server. The other big thing was making research actually persist across sessions. Most agents treat a single deliverable (a PDF, a spreadsheet) as the end goal. In investing that's day one. You update the model when earnings drop, re-run comps when a competitor reports, keep layering new analysis on old. But try doing that across agent sessions, files don't carry over, you re-paste context every time. So we built everything around workspaces. Each one maps to a persistent sandbox, one per research goal. The agent maintains its own memory file with findings and a file index that gets re-read before every LLM call. Come back a week later, start a new thread, it picks up where it left off. We also wanted the agent to have real domain context the way Claude Code has codebase context. Portfolio, watchlist, risk tolerance, financial data sources, all injected into every call. Existing AI investing platforms have some of that but nothing close to what a proper agent harness can do. We wanted both and couldn't find it, so we built it and open-sourced the whole thing. https://ift.tt/gE3LiU7 April 14, 2026 at 09:48PM
Subscribe to:
Comments (Atom)