The best Hacker News stories from Show from the past day

Go back

Latest posts:

Show HN: PageAgent, A GUI agent that lives inside your web app

Title: Show HN: PageAgent, A GUI agent that lives inside your web app<p>Hi HN,<p>I'm building PageAgent, an open-source (MIT) library that embeds an AI agent directly into your frontend.<p>I built this because I believe there's a massive design space for deploying general agents natively inside the web apps we already use, rather than treating the web merely as a dumb target for isolated bots.<p>Currently, most AI agents operate from external clients or server-side programs, effectively leaving web development out of the AI ecosystem. I'm experimenting with an "inside-out" paradigm instead. By dropping the library into a page, you get a client-side agent that interacts natively with the live DOM tree and inherits the user's active session out of the box, which works perfectly for SPAs.<p>To handle cross-page tasks, I built an optional browser extension that acts as a "bridge". This allows the web-page agent to control the entire browser with explicit user authorization. Instead of a desktop app controlling your browser, your web app is empowered to act as a general agent that can navigate the broader web.<p>I'd love to start a conversation about the viability of this architecture, and what you all think about the future of in-app general agents. Happy to answer any questions!

Show HN: PageAgent, A GUI agent that lives inside your web app

Title: Show HN: PageAgent, A GUI agent that lives inside your web app<p>Hi HN,<p>I'm building PageAgent, an open-source (MIT) library that embeds an AI agent directly into your frontend.<p>I built this because I believe there's a massive design space for deploying general agents natively inside the web apps we already use, rather than treating the web merely as a dumb target for isolated bots.<p>Currently, most AI agents operate from external clients or server-side programs, effectively leaving web development out of the AI ecosystem. I'm experimenting with an "inside-out" paradigm instead. By dropping the library into a page, you get a client-side agent that interacts natively with the live DOM tree and inherits the user's active session out of the box, which works perfectly for SPAs.<p>To handle cross-page tasks, I built an optional browser extension that acts as a "bridge". This allows the web-page agent to control the entire browser with explicit user authorization. Instead of a desktop app controlling your browser, your web app is empowered to act as a general agent that can navigate the broader web.<p>I'd love to start a conversation about the viability of this architecture, and what you all think about the future of in-app general agents. Happy to answer any questions!

Show HN: Poppy – A simple app to stay intentional with relationships

I built Poppy as a side project to help people keep in touch more intentionally. Would love feedback on onboarding, reminders, and overall UX. Happy to answer questions.

Show HN: Poppy – A simple app to stay intentional with relationships

I built Poppy as a side project to help people keep in touch more intentionally. Would love feedback on onboarding, reminders, and overall UX. Happy to answer questions.

Show HN: Jido 2.0, Elixir Agent Framework

Hi HN!<p>I'm the author of an Elixir Agent Framework called Jido. We reached our 2.0 release this week, shipping a production-hardened framework to build, manage and run Agents on the BEAM.<p>Jido now supports a host of Agentic features, including:<p>- Tool Calling and Agent Skills - Comprehensive multi-agent support across distributed BEAM processes with Supervision - Multiple reasoning strategies including ReAct, Chain of Thought, Tree of Thought, and more - Advanced workflow capabilities - Durability through a robust Storage and Persistence layer - Agentic Memory - MCP and Sensors to interface with external services - Deep observability and debugging capabilities, including full stack OTel<p>I know Agent Frameworks can be considered a bit stale, but there hasn't been a major release of a framework on the BEAM. With a growing realization that the architecture of the BEAM is a good match for Agentic workloads, the time was right to make the announcement.<p>My background is enterprise engineering, distributed systems and Open Source. We've got a strong and growing community of builders committed to the Jido ecosystem. We're looking forward to what gets built on top of Jido!<p>Come build agents with us!

Show HN: Jido 2.0, Elixir Agent Framework

Hi HN!<p>I'm the author of an Elixir Agent Framework called Jido. We reached our 2.0 release this week, shipping a production-hardened framework to build, manage and run Agents on the BEAM.<p>Jido now supports a host of Agentic features, including:<p>- Tool Calling and Agent Skills - Comprehensive multi-agent support across distributed BEAM processes with Supervision - Multiple reasoning strategies including ReAct, Chain of Thought, Tree of Thought, and more - Advanced workflow capabilities - Durability through a robust Storage and Persistence layer - Agentic Memory - MCP and Sensors to interface with external services - Deep observability and debugging capabilities, including full stack OTel<p>I know Agent Frameworks can be considered a bit stale, but there hasn't been a major release of a framework on the BEAM. With a growing realization that the architecture of the BEAM is a good match for Agentic workloads, the time was right to make the announcement.<p>My background is enterprise engineering, distributed systems and Open Source. We've got a strong and growing community of builders committed to the Jido ecosystem. We're looking forward to what gets built on top of Jido!<p>Come build agents with us!

Show HN: We want to displace Notion with collaborative Markdown files

Hi HN! We at Moment[1] are working on Notion alternative which is (1) rich and collaborative, but (2) also just plain-old Markdown files, stored in git (ok, technically in jj), on local disk. We think the era of rigid SaaS UI is, basically, over: coding agents (`claude`, `amp`, `copilot`, `opencode`, <i>etc</i>.) are good enough now that they instantly build custom UI that fits your needs exactly. The very best agents in the world are coding agents, and we want to allow people to simply use them, <i>e.g.</i>, to build little internal tools—but without compromising on collaboration.<p>Moment aims to cover this and other gaps: seamless collaborative editing for teams, more robust programming capabilities built in (including a from-scratch React integration), and tools for accessing private APIs.<p>A lot of our challenge is just in making the collaborative editing work really well. We have found this is a lot harder than simply slapping Yjs on the frontend and calling it a day. We wrote about this previously and the post[2] did pretty well on HN: Lies I was Told About Collaborative editing (352 upvotes as of this writing). Beyond that, in part 2, we'll talk about the reasons we found it hard to get collab to run at 60fps consistently—for one, the Yjs ProseMirror bindings completely tear down and re-create the entire document on every single collaborative keystroke.<p>We hope you will try it out! At this stage even negative feedback is helpful. :)<p>[1]: <a href="https://www.moment.dev/" rel="nofollow">https://www.moment.dev/</a><p>[2]: <a href="https://news.ycombinator.com/item?id=42343953">https://news.ycombinator.com/item?id=42343953</a>

Show HN: Effective Git

As many of us shift from being software engineers to software managers, tracking changes the right way is growing more important.<p>It’s time to truly understand and master Git.

Show HN: Explain Curl Commands

Show HN: I made a zero-copy coroutine tracer to find my scheduler's lost wakeups

Show HN: I made a zero-copy coroutine tracer to find my scheduler's lost wakeups

Show HN: Open-Source Article 12 Logging Infrastructure for the EU AI Act

EU legislation (which affects UK and US companies in many cases) requires being able to truly reconstruct agentic events.<p>I've worked in a number of regulated industries off & on for years, and recently hit this gap.<p>We already had strong observability, but if someone asked me to prove exactly what happened for a specific AI decision X months ago (and demonstrate that the log trail had not been altered), I could not.<p>The EU AI Act has already entered force, and its Article 12 kicks-in in August this year, requiring automatic event recording and six-month retention for high-risk systems, which many legal commentators have suggested reads more like an append-only ledger requirement than standard application logging.<p>With this in mind, we built a small free, open-source TypeScript library for Node apps using the Vercel AI SDK that captures inference as an append-only log.<p>It wraps the model in middleware, automatically logs every inference call to structured JSONL in your own S3 bucket, chains entries with SHA-256 hashes for tamper detection, enforces a 180-day retention floor, and provides a CLI to reconstruct a decision and verify integrity. There is also a coverage command that flags likely gaps (in practice omissions are a bigger risk than edits).<p>The library is deliberately simple: TS, targeting Vercel AI SDK middleware, S3 or local fs, linear hash chaining. It also works with Mastra (agentic framework), and I am happy to expand its integrations via PRs.<p>Blog post with link to repo: <a href="https://systima.ai/blog/open-source-article-12-audit-logging" rel="nofollow">https://systima.ai/blog/open-source-article-12-audit-logging</a><p>I'd value feedback, thoughts, and any critique.

Show HN: Rust compiler in PHP emitting x86-64 executables

Show HN: Rust compiler in PHP emitting x86-64 executables

Show HN: P0 – Yes, AI can ship complex features into real codebases

Show HN: Stacked Game of Life

<a href="https://github.com/vnglst/stacked-game-of-life" rel="nofollow">https://github.com/vnglst/stacked-game-of-life</a>

Show HN: Stacked Game of Life

<a href="https://github.com/vnglst/stacked-game-of-life" rel="nofollow">https://github.com/vnglst/stacked-game-of-life</a>

Show HN: Web Audio Studio – A Visual Debugger for Web Audio API Graphs

Hi HN,<p>I’ve been working on a browser-based tool for exploring and debugging Web Audio API graphs.<p>Web Audio Studio lets you write real Web Audio API code, run it, and see the runtime graph it produces as an interactive visual representation. Instead of mentally tracking connect() calls, you can inspect the actual structure of the graph, follow signal flow, and tweak parameters while the audio is playing.<p>It includes built-in visualizations for common node types — waveforms, filter responses, analyser time and frequency views, compressor transfer curves, waveshaper distortion, spatial positioning, delay timing, and more — so you can better understand what each part of the graph is doing. You can also insert an AnalyserNode between any two nodes to inspect the signal at that exact point in the chain.<p>There are around 20 templates (basic oscillator setups, FM/AM synthesis, convolution reverb, IIR filters, spatial audio, etc.), so you can start from working examples and modify them instead of building everything from scratch.<p>Everything runs fully locally in the browser — no signup, no backend.<p>The motivation came from working with non-trivial Web Audio graphs and finding it increasingly difficult to reason about structure and signal flow once things grow beyond simple examples. Most tutorials show small snippets, but real projects quickly become harder to inspect. I wanted something that stays close to the native Web Audio API while making the runtime graph visible and inspectable.<p>This is an early alpha and desktop-only for now.<p>I’d really appreciate feedback — especially from people who have used Web Audio API in production or built audio tools. You can leave comments here, or use the feedback button inside the app.<p><a href="https://webaudio.studio" rel="nofollow">https://webaudio.studio</a>

Show HN: Pianoterm – Run shell commands from your Piano. A Linux CLI tool

A little weekend project, made so I can pause/play/rewind directly on the piano, when learning a song by ear.

Show HN: Vibe Code your 3D Models

Hi HN,<p>I’m the creator of SynapsCAD, an open-source desktop application I've been building that combines an OpenSCAD code editor, a real-time 3D viewport, and an AI assistant.<p>You can write OpenSCAD code, compile it directly to a 3D mesh, and use an LLM (OpenAI, Claude, Gemini, ...) to modify the code through natural language.<p>Demo video: <a href="https://www.youtube.com/watch?v=cN8a5UozS5Q" rel="nofollow">https://www.youtube.com/watch?v=cN8a5UozS5Q</a><p>A bit about the architecture:<p>- It’s built entirely in Rust.<p>- The UI and 3D viewport are powered by Bevy 0.15 and egui.<p>- It uses a pure-Rust compilation pipeline (openscad-rs for parsing and csgrs for constructive solid geometry rendering) so there are no external tools or WASM required.<p>- Async AI network calls are handled by Tokio in the background to keep the Bevy render loop smooth.<p>Disclaimer: This is a very early prototype. The OpenSCAD parser/compiler doesn't support everything perfectly yet, so you will definitely hit some rough edges if you throw complex scripts at it.<p>I mostly just want to get this into the hands of people who tinker with CAD or Rust.<p>I'd be super happy for any feedback, architectural critiques, or bug reports—especially if you can drop specific OpenSCAD snippets that break the compiler in the GitHub issues!<p>GitHub (Downloads for Win/Mac/Linux): <a href="https://github.com/ierror/synaps-cad" rel="nofollow">https://github.com/ierror/synaps-cad</a><p>Happy to answer any questions about the tech stack or the roadmap!

< 1 2 3 4 ... 949 950 951 >