The best Hacker News stories from Show from the past day

Go back

Latest posts:

Show HN: Nogic – VS Code extension that visualizes your codebase as a graph

I built Nogic, a VSCode extension currently, because AI tools make code grow faster than developers can build a mental model by jumping between files. Exploring structure visually has been helping me onboard to unfamiliar codebases faster.<p>It’s early and rough, but usable. Would love feedback on whether this is useful and what relationships are most valuable to visualize.

Show HN: An iOS budget app I've been maintaining since 2011

I’ve been building and selling software since the early 2000s, starting with classic shareware. In 2011, I moved into the App Store world and built an iOS budget app because I needed a simple way to track my own expenses.<p>At the time, my plan was to replace a few larger shareware projects with several smaller apps to spread the risk. That didn’t quite work out — one app, MoneyControl, quickly grew so much that it became my main focus.<p>Fifteen years later, the app is still on the App Store, still actively developed, and still used by people who started with version 1.0. Many apps from that era are long gone.<p>Looking back, these are some of the things that mattered most:<p>Starting early helped, but wasn’t enough on its own. Early visibility made a difference, but long-term maintenance and reliability are what kept users.<p>Focus beat diversification. I wanted many small apps. I ended up with one large, long-lived product. Deep focus turned out to be more sustainable.<p>Long-term maintenance is most of the work. Adapting to new iOS versions, migrating data safely, handling edge cases, and keeping old data usable mattered more than flashy features.<p>Discoverability keeps getting harder. Reaching users on the App Store today is much more difficult than it was years ago. Prices are higher than in the old 99-cent days, but visibility hasn’t improved.<p>I’m a developer first, not a marketer. I work alone, with occasional help from freelancers. No employees, no growth team. The app could probably have grown more with better marketing, but that was never my strength.<p>You don’t need to get rich to build something sustainable. I didn’t build this for an exit. I’ve been able to make a living from my work for over 20 years, which feels like success to me.<p>Building things you actually use keeps you honest. Every product I built was something I personally needed. That authenticity mattered more than any roadmap.<p>This week I released version 10 with a new design and a major technical overhaul. It feels less like a milestone and more like preparing the app for the next phase.<p>Happy to answer questions about long-term app maintenance, indie development, or keeping a product alive across many iOS generations.

Show HN: An iOS budget app I've been maintaining since 2011

I’ve been building and selling software since the early 2000s, starting with classic shareware. In 2011, I moved into the App Store world and built an iOS budget app because I needed a simple way to track my own expenses.<p>At the time, my plan was to replace a few larger shareware projects with several smaller apps to spread the risk. That didn’t quite work out — one app, MoneyControl, quickly grew so much that it became my main focus.<p>Fifteen years later, the app is still on the App Store, still actively developed, and still used by people who started with version 1.0. Many apps from that era are long gone.<p>Looking back, these are some of the things that mattered most:<p>Starting early helped, but wasn’t enough on its own. Early visibility made a difference, but long-term maintenance and reliability are what kept users.<p>Focus beat diversification. I wanted many small apps. I ended up with one large, long-lived product. Deep focus turned out to be more sustainable.<p>Long-term maintenance is most of the work. Adapting to new iOS versions, migrating data safely, handling edge cases, and keeping old data usable mattered more than flashy features.<p>Discoverability keeps getting harder. Reaching users on the App Store today is much more difficult than it was years ago. Prices are higher than in the old 99-cent days, but visibility hasn’t improved.<p>I’m a developer first, not a marketer. I work alone, with occasional help from freelancers. No employees, no growth team. The app could probably have grown more with better marketing, but that was never my strength.<p>You don’t need to get rich to build something sustainable. I didn’t build this for an exit. I’ve been able to make a living from my work for over 20 years, which feels like success to me.<p>Building things you actually use keeps you honest. Every product I built was something I personally needed. That authenticity mattered more than any roadmap.<p>This week I released version 10 with a new design and a major technical overhaul. It feels less like a milestone and more like preparing the app for the next phase.<p>Happy to answer questions about long-term app maintenance, indie development, or keeping a product alive across many iOS generations.

Show HN: Self-host Reddit – 2.38B posts, works offline, yours forever

Reddit's API is effectively dead for archival. Third-party apps are gone. Reddit has threatened to cut off access to the Pushshift dataset multiple times. But 3.28TB of Reddit history exists as a torrent right now, and I built a tool to turn it into something you can browse on your own hardware.<p>The key point: This doesn't touch Reddit's servers. Ever. Download the Pushshift dataset, run my tool locally, get a fully browsable archive. Works on an air-gapped machine. Works on a Raspberry Pi serving your LAN. Works on a USB drive you hand to someone.<p>What it does: Takes compressed data dumps from Reddit (.zst), Voat (SQL), and Ruqqus (.7z) and generates static HTML. No JavaScript, no external requests, no tracking. Open index.html and browse. Want search? Run the optional Docker stack with PostgreSQL – still entirely on your machine.<p>API & AI Integration: Full REST API with 30+ endpoints – posts, comments, users, subreddits, full-text search, aggregations. Also ships with an MCP server (29 tools) so you can query your archive directly from AI tools.<p>Self-hosting options: - USB drive / local folder (just open the HTML files) - Home server on your LAN - Tor hidden service (2 commands, no port forwarding needed) - VPS with HTTPS - GitHub Pages for small archives<p>Why this matters: Once you have the data, you own it. No API keys, no rate limits, no ToS changes can take it away.<p>Scale: Tens of millions of posts per instance. PostgreSQL backend keeps memory constant regardless of dataset size. For the full 2.38B post dataset, run multiple instances by topic.<p>How I built it: Python, PostgreSQL, Jinja2 templates, Docker. Used Claude Code throughout as an experiment in AI-assisted development. Learned that the workflow is "trust but verify" – it accelerates the boring parts but you still own the architecture.<p>Live demo: <a href="https://online-archives.github.io/redd-archiver-example/" rel="nofollow">https://online-archives.github.io/redd-archiver-example/</a><p>GitHub: <a href="https://github.com/19-84/redd-archiver" rel="nofollow">https://github.com/19-84/redd-archiver</a> (Public Domain)<p>Pushshift torrent: <a href="https://academictorrents.com/details/1614740ac8c94505e4ecb9d88be8bed7b6afddd4" rel="nofollow">https://academictorrents.com/details/1614740ac8c94505e4ecb9d...</a>

Show HN: Self-host Reddit – 2.38B posts, works offline, yours forever

Reddit's API is effectively dead for archival. Third-party apps are gone. Reddit has threatened to cut off access to the Pushshift dataset multiple times. But 3.28TB of Reddit history exists as a torrent right now, and I built a tool to turn it into something you can browse on your own hardware.<p>The key point: This doesn't touch Reddit's servers. Ever. Download the Pushshift dataset, run my tool locally, get a fully browsable archive. Works on an air-gapped machine. Works on a Raspberry Pi serving your LAN. Works on a USB drive you hand to someone.<p>What it does: Takes compressed data dumps from Reddit (.zst), Voat (SQL), and Ruqqus (.7z) and generates static HTML. No JavaScript, no external requests, no tracking. Open index.html and browse. Want search? Run the optional Docker stack with PostgreSQL – still entirely on your machine.<p>API & AI Integration: Full REST API with 30+ endpoints – posts, comments, users, subreddits, full-text search, aggregations. Also ships with an MCP server (29 tools) so you can query your archive directly from AI tools.<p>Self-hosting options: - USB drive / local folder (just open the HTML files) - Home server on your LAN - Tor hidden service (2 commands, no port forwarding needed) - VPS with HTTPS - GitHub Pages for small archives<p>Why this matters: Once you have the data, you own it. No API keys, no rate limits, no ToS changes can take it away.<p>Scale: Tens of millions of posts per instance. PostgreSQL backend keeps memory constant regardless of dataset size. For the full 2.38B post dataset, run multiple instances by topic.<p>How I built it: Python, PostgreSQL, Jinja2 templates, Docker. Used Claude Code throughout as an experiment in AI-assisted development. Learned that the workflow is "trust but verify" – it accelerates the boring parts but you still own the architecture.<p>Live demo: <a href="https://online-archives.github.io/redd-archiver-example/" rel="nofollow">https://online-archives.github.io/redd-archiver-example/</a><p>GitHub: <a href="https://github.com/19-84/redd-archiver" rel="nofollow">https://github.com/19-84/redd-archiver</a> (Public Domain)<p>Pushshift torrent: <a href="https://academictorrents.com/details/1614740ac8c94505e4ecb9d88be8bed7b6afddd4" rel="nofollow">https://academictorrents.com/details/1614740ac8c94505e4ecb9d...</a>

Show HN: An LLM-optimized programming language

Show HN: 30k IKEA items in flat text

OP here.<p>I took the unofficial IKEA US dataset (originally scraped by jeffreyszhou) and converted all 30,511 products into a flat, markdown-like protocol called CommerceTXT.<p>The goal: See if a flatter structure is more efficient for LLM context windows.<p>The results: - Size: 30k products across 632 categories. - Efficiency: The text version uses ~24% fewer tokens (3.6M saved total) compared to the equivalent minified JSON. - Structure: Files are organized in folders (e.g. /products/category/), which helps with testing hierarchical retrieval routers.<p>The link goes to the dataset on Hugging Face which has the full benchmarks.<p>Parser code is here: <a href="https://github.com/commercetxt/commercetxt" rel="nofollow">https://github.com/commercetxt/commercetxt</a><p>Happy to answer questions about the conversion logic!

Show HN: Fall asleep by watching JavaScript load

Show HN: Fall asleep by watching JavaScript load

Show HN: Agent-of-empires: OpenCode and Claude Code session manager

Hi! I’m Nathan: an ML Engineer at Mozilla.ai: I built agent-of-empires (aoe): a CLI application to help you manage all of your running Claude Code/Opencode sessions and know when they are waiting for you.<p>- Written in rust and relies on tmux for security and reliability - Monitors state of cli sessions to tell you when an agent is running vs idle vs waiting for your input - Manage sessions by naming them, grouping them, configuring profiles for various settings<p>I'm passionate about getting self-hosted open-weight LLMs to be valid options to compete with proprietary closed models. One roadblock for me is that although tools like opencode allow you to connect to Local LLMs (Ollama, lm studio, etc), they generally run muuuuuch slower than models hosted by Anthropic and OpenAI. I would start a coding agent on a task, but then while I was sitting waiting for that task to complete, I would start opening new terminal windows to start multitasking. Pretty soon, I was spending a lot of time toggling between terminal windows to see which one needed me: like help in adding a clarification, approving a new command, or giving it a new task.<p>That’s why I build agent-of-empires (“aoe”). With aoe, I can launch a bunch of opencode and Claude Code sessions and quickly see their status or toggle between them, which helps me avoid having a lot of terminal windows open, or having to manually attach and detach from tmux sessions myself. It’s helping me give local LLMs a fair try, because them being slower is now much less of a bottleneck.<p>You can give it an install with<p>curl -fsSL <a href="https://raw.githubusercontent.com/njbrake/agent-of-empires/main/scripts/install.sh" rel="nofollow">https://raw.githubusercontent.com/njbrake/agent-of-empires/m...</a> | bash<p>Or brew install njbrake/aoe/aoe<p>And then launch by simply entering the command `aoe`.<p>I’m interested in what you think as well as what features you think would be useful to add!<p>I am planning to add some further features around sandboxing (with docker) as well as support for intuitive git worktrees and am curious if there are any opinions about what should or shouldn’t be in it.<p>I decided against MCP management or generic terminal usage, to help keep the tool focused on parts of agentic coding that I haven’t found a usable solution for.<p>I hit the character limit on this post which prevented me from including a view of the output, but the readme on the github link has a screenshot showing what it looks like.<p>Thanks!

Show HN: Agent-of-empires: OpenCode and Claude Code session manager

Hi! I’m Nathan: an ML Engineer at Mozilla.ai: I built agent-of-empires (aoe): a CLI application to help you manage all of your running Claude Code/Opencode sessions and know when they are waiting for you.<p>- Written in rust and relies on tmux for security and reliability - Monitors state of cli sessions to tell you when an agent is running vs idle vs waiting for your input - Manage sessions by naming them, grouping them, configuring profiles for various settings<p>I'm passionate about getting self-hosted open-weight LLMs to be valid options to compete with proprietary closed models. One roadblock for me is that although tools like opencode allow you to connect to Local LLMs (Ollama, lm studio, etc), they generally run muuuuuch slower than models hosted by Anthropic and OpenAI. I would start a coding agent on a task, but then while I was sitting waiting for that task to complete, I would start opening new terminal windows to start multitasking. Pretty soon, I was spending a lot of time toggling between terminal windows to see which one needed me: like help in adding a clarification, approving a new command, or giving it a new task.<p>That’s why I build agent-of-empires (“aoe”). With aoe, I can launch a bunch of opencode and Claude Code sessions and quickly see their status or toggle between them, which helps me avoid having a lot of terminal windows open, or having to manually attach and detach from tmux sessions myself. It’s helping me give local LLMs a fair try, because them being slower is now much less of a bottleneck.<p>You can give it an install with<p>curl -fsSL <a href="https://raw.githubusercontent.com/njbrake/agent-of-empires/main/scripts/install.sh" rel="nofollow">https://raw.githubusercontent.com/njbrake/agent-of-empires/m...</a> | bash<p>Or brew install njbrake/aoe/aoe<p>And then launch by simply entering the command `aoe`.<p>I’m interested in what you think as well as what features you think would be useful to add!<p>I am planning to add some further features around sandboxing (with docker) as well as support for intuitive git worktrees and am curious if there are any opinions about what should or shouldn’t be in it.<p>I decided against MCP management or generic terminal usage, to help keep the tool focused on parts of agentic coding that I haven’t found a usable solution for.<p>I hit the character limit on this post which prevented me from including a view of the output, but the readme on the github link has a screenshot showing what it looks like.<p>Thanks!

Show HN: Yolobox – Run AI coding agents with full sudo without nuking home dir

Show HN: Yolobox – Run AI coding agents with full sudo without nuking home dir

Show HN: AI in SolidWorks

Hey HN! We’re Will and Jorge, and we’ve built LAD (Language-Aided Design), a SolidWorks add-in that uses LLMs to create sketches, features, assemblies, and macros from conversational inputs (<a href="https://www.trylad.com/" rel="nofollow">https://www.trylad.com/</a>).<p>We come from software engineering backgrounds where tools like Claude Code and Cursor have come to dominate, but when poking around CAD systems a few months back we realized there's no way to go from a text prompt input to a modeling output in any of the major CAD systems. In our testing, the LLMs aren't as good at making 3D objects as they are are writing code, but we think they'll get a lot better in the upcoming months and years.<p>To bridge this gap, we've created LAD, an add-in in SolidWorks to turn conversational input and uploaded documents/images into parts, assemblies, and macros. It includes:<p>- Dozens of tools the LLM can call to create sketches, features, and other objects in parts.<p>- Assembly tools the LLM can call to turn parts into assemblies.<p>- File system tools the LLM can use to create, save, search, and read SolidWorks files and documentation.<p>- Macro writing/running tools plus a SolidWorks API documentation search so the LLM can use macros.<p>- Automatic screenshots and feature tree parsing to provide the LLM context on the current state.<p>- Checkpointing to roll back unwanted edits and permissioning to determine which commands wait for user permission.<p>You can try LAD at <a href="https://www.trylad.com/" rel="nofollow">https://www.trylad.com/</a> and let us know what features would make it more useful for your work. To be honest, the LLMs aren't great at CAD right now, but we're mostly curious to hear if people would want and use this if it worked well.

Show HN: AI in SolidWorks

Hey HN! We’re Will and Jorge, and we’ve built LAD (Language-Aided Design), a SolidWorks add-in that uses LLMs to create sketches, features, assemblies, and macros from conversational inputs (<a href="https://www.trylad.com/" rel="nofollow">https://www.trylad.com/</a>).<p>We come from software engineering backgrounds where tools like Claude Code and Cursor have come to dominate, but when poking around CAD systems a few months back we realized there's no way to go from a text prompt input to a modeling output in any of the major CAD systems. In our testing, the LLMs aren't as good at making 3D objects as they are are writing code, but we think they'll get a lot better in the upcoming months and years.<p>To bridge this gap, we've created LAD, an add-in in SolidWorks to turn conversational input and uploaded documents/images into parts, assemblies, and macros. It includes:<p>- Dozens of tools the LLM can call to create sketches, features, and other objects in parts.<p>- Assembly tools the LLM can call to turn parts into assemblies.<p>- File system tools the LLM can use to create, save, search, and read SolidWorks files and documentation.<p>- Macro writing/running tools plus a SolidWorks API documentation search so the LLM can use macros.<p>- Automatic screenshots and feature tree parsing to provide the LLM context on the current state.<p>- Checkpointing to roll back unwanted edits and permissioning to determine which commands wait for user permission.<p>You can try LAD at <a href="https://www.trylad.com/" rel="nofollow">https://www.trylad.com/</a> and let us know what features would make it more useful for your work. To be honest, the LLMs aren't great at CAD right now, but we're mostly curious to hear if people would want and use this if it worked well.

Show HN: Yellopages – New tab Chrome extension

Hey all- I just released a New tab replacement Chrome extension that makes browsing a lot easier - it also solves many of the annoyances with browser tabs. It's called Yellopages and it's free. Hope you'll give it a try.<p>* Groups all tabs from same domain. Makes it simple to kill all your Gmail tabs in one click (or keep just one).<p>* Groups all tabs playing audio. Toggle the sound for each one.<p>* Single text search for open tabs, bookmarks, and browsing history.<p>* Groups all tabs with new notifications (e.g. emails, likes, posts, replies, etc.)<p>* One click to kill all tabs (e.g. you're sharing screen in Zoom). A second click brings them all back.<p>I'm a solo web developer and I'm hoping to build an audience with my work. More at: <a href="https://buymeacoffee.com/kawaicheung" rel="nofollow">https://buymeacoffee.com/kawaicheung</a>

Show HN: mcpc – Universal command-line client for Model Context Protocol (MCP)

Show HN: VAM Seek – 2D video navigation grid, 15KB, zero server load

Hi HN! I built VAM Seek because I was frustrated with 1D seek bars – you never know where you're going until you get there.<p>VAM Seek renders a 2D thumbnail grid next to your video. Click any cell to jump. All frame extraction happens client-side via canvas – no server processing, no pre-generated thumbnails.<p>- 15KB, zero dependencies - One-line integration - Works with any <video> element<p>Live demo: <a href="https://haasiy.main.jp/vam_web/deploy/lolipop/index.html" rel="nofollow">https://haasiy.main.jp/vam_web/deploy/lolipop/index.html</a><p>Would love feedback!

Show HN: GlyphLang – An AI-first programming language

While working on a proof of concept project, I kept hitting Claude's token limit 30-60 minutes into their 5-hour sessions. The accumulating context from the codebase was eating through tokens fast. So I built a language designed to be generated by AI rather than written by humans.<p>GlyphLang<p>GlyphLang replaces verbose keywords with symbols that tokenize more efficiently:<p><pre><code> # Python @app.route('/users/<id>') def get_user(id): user = db.query("SELECT * FROM users WHERE id = ?", id) return jsonify(user) # GlyphLang @ GET /users/:id { $ user = db.query("SELECT * FROM users WHERE id = ?", id) > user } @ = route, $ = variable, > = return. Initial benchmarks show ~45% fewer tokens than Python, ~63% fewer than Java. </code></pre> In practice, that means more logic fits in context, and sessions stretch longer before hitting limits. The AI maintains a broader view of your codebase throughout.<p>Before anyone asks: no, this isn't APL with extra steps. APL, Perl, and Forth are symbol-heavy but optimized for mathematical notation, human terseness, or machine efficiency. GlyphLang is specifically optimized for how modern LLMs tokenize. It's designed to be generated by AI and reviewed by humans, not the other way around. That said, it's still readable enough to be written or tweaked if the occasion requires.<p>It's still a work in progress, but it's a usable language with a bytecode compiler, JIT, LSP, VS Code extension, PostgreSQL, WebSockets, async/await, generics.<p>Docs: <a href="https://glyphlang.dev/docs" rel="nofollow">https://glyphlang.dev/docs</a><p>GitHub: <a href="https://github.com/GlyphLang/GlyphLang" rel="nofollow">https://github.com/GlyphLang/GlyphLang</a>

Show HN: Epstein IM – Talk to Epstein clone in iMessage

< 1 2 3 4 ... 926 927 928 >