The best Hacker News stories from Show from the past day
Latest posts:
Show HN: Coroot – eBPF-based, open source observability with actionable insights
A common open source approach to observability will begin with databases and visualizations for telemetry - Grafana, Prometheus, Jaeger. But observability doesn’t begin and end here: these tools require configuration, dashboard customization, and may not actually pinpoint the data you need to mitigate system risks.<p>Coroot was designed to solve the problem of manual, time-consuming observability analysis: it handles the full observability journey — from collecting telemetry to turning it into actionable insights. We also strongly believe that simple observability should be an innovation everyone can benefit from: which is why our software is open source.<p>Features:<p>- Cost monitoring to track and minimise your cloud expenses (AWS, GCP, Azure.)<p>- SLO tracking with alerts to detect anomalies and compare them to your system’s baseline behaviour.<p>- 1-click application profiling: see the exact line of code that caused an anomaly.<p>- Mapped timeframes (stop digging through Grafana to find when the incident occurred.)<p>- eBPF automatically gathers logs, metrics, traces, and profiles for you.<p>- Service map to grasp a complete at-a-glance picture of your system.<p>- Automatic discovery and monitoring of every application deployment in your kubernetes cluster.<p>We welcome any feedback and hope the tool can improve your workflow!
Show HN: Coroot – eBPF-based, open source observability with actionable insights
A common open source approach to observability will begin with databases and visualizations for telemetry - Grafana, Prometheus, Jaeger. But observability doesn’t begin and end here: these tools require configuration, dashboard customization, and may not actually pinpoint the data you need to mitigate system risks.<p>Coroot was designed to solve the problem of manual, time-consuming observability analysis: it handles the full observability journey — from collecting telemetry to turning it into actionable insights. We also strongly believe that simple observability should be an innovation everyone can benefit from: which is why our software is open source.<p>Features:<p>- Cost monitoring to track and minimise your cloud expenses (AWS, GCP, Azure.)<p>- SLO tracking with alerts to detect anomalies and compare them to your system’s baseline behaviour.<p>- 1-click application profiling: see the exact line of code that caused an anomaly.<p>- Mapped timeframes (stop digging through Grafana to find when the incident occurred.)<p>- eBPF automatically gathers logs, metrics, traces, and profiles for you.<p>- Service map to grasp a complete at-a-glance picture of your system.<p>- Automatic discovery and monitoring of every application deployment in your kubernetes cluster.<p>We welcome any feedback and hope the tool can improve your workflow!
Show HN: Uncurl.dev – Convert curl commands to a shareable, executable UI
Hey HN,<p>I built uncurl.dev to scratch my own itch: I kept getting curl commands in API docs, bug reports, or Slack messages and wanted a quick way to visualize, run, and debug them without firing up Postman or writing code to dissect them. It was also sometimes painful to create a test page specifically for non-tech users to consume an API. I originally vibe-coded this over a weekend just to make it easier for myself to debug API requests shared as curl commands. It slowly grew into something I found surprisingly useful in my workflow, so I decided to clean it up and share it.<p>uncurl.dev takes a curl command and:
- Converts it into a visual representation
- Lets you edit and inspect all parts of the request
- Allows sharing via a unique link
- (Optionally) executes it from the server, so business or non-technical users can see results<p>Execution is currently behind login, with a cap (5/min) to avoid abuse and manage costs. Non-logged-in users can still build and share curl commands—they just can’t execute them. The server runs each request in a Docker sandbox with strict resource/time limits (cpu, memory, timeout, no network access outside the request).<p>It’s not meant to replace full-featured tools like Postman or Hoppscotch. It’s more of a “CLI-to-UI bridge” for fast sharing and debugging, especially in dev workflows where curl is the starting point. Think of it like Pastebin or JSFiddle, but for curl commands.<p>If you’ve ever copied a curl from an API doc and wanted a cleaner way to see it or send it to someone else, I’d love your feedback.<p>Thanks!
(You can try it without signup here: <a href="https://uncurl.dev" rel="nofollow">https://uncurl.dev</a>)
Show HN: Uncurl.dev – Convert curl commands to a shareable, executable UI
Hey HN,<p>I built uncurl.dev to scratch my own itch: I kept getting curl commands in API docs, bug reports, or Slack messages and wanted a quick way to visualize, run, and debug them without firing up Postman or writing code to dissect them. It was also sometimes painful to create a test page specifically for non-tech users to consume an API. I originally vibe-coded this over a weekend just to make it easier for myself to debug API requests shared as curl commands. It slowly grew into something I found surprisingly useful in my workflow, so I decided to clean it up and share it.<p>uncurl.dev takes a curl command and:
- Converts it into a visual representation
- Lets you edit and inspect all parts of the request
- Allows sharing via a unique link
- (Optionally) executes it from the server, so business or non-technical users can see results<p>Execution is currently behind login, with a cap (5/min) to avoid abuse and manage costs. Non-logged-in users can still build and share curl commands—they just can’t execute them. The server runs each request in a Docker sandbox with strict resource/time limits (cpu, memory, timeout, no network access outside the request).<p>It’s not meant to replace full-featured tools like Postman or Hoppscotch. It’s more of a “CLI-to-UI bridge” for fast sharing and debugging, especially in dev workflows where curl is the starting point. Think of it like Pastebin or JSFiddle, but for curl commands.<p>If you’ve ever copied a curl from an API doc and wanted a cleaner way to see it or send it to someone else, I’d love your feedback.<p>Thanks!
(You can try it without signup here: <a href="https://uncurl.dev" rel="nofollow">https://uncurl.dev</a>)
Show HN: Uncurl.dev – Convert curl commands to a shareable, executable UI
Hey HN,<p>I built uncurl.dev to scratch my own itch: I kept getting curl commands in API docs, bug reports, or Slack messages and wanted a quick way to visualize, run, and debug them without firing up Postman or writing code to dissect them. It was also sometimes painful to create a test page specifically for non-tech users to consume an API. I originally vibe-coded this over a weekend just to make it easier for myself to debug API requests shared as curl commands. It slowly grew into something I found surprisingly useful in my workflow, so I decided to clean it up and share it.<p>uncurl.dev takes a curl command and:
- Converts it into a visual representation
- Lets you edit and inspect all parts of the request
- Allows sharing via a unique link
- (Optionally) executes it from the server, so business or non-technical users can see results<p>Execution is currently behind login, with a cap (5/min) to avoid abuse and manage costs. Non-logged-in users can still build and share curl commands—they just can’t execute them. The server runs each request in a Docker sandbox with strict resource/time limits (cpu, memory, timeout, no network access outside the request).<p>It’s not meant to replace full-featured tools like Postman or Hoppscotch. It’s more of a “CLI-to-UI bridge” for fast sharing and debugging, especially in dev workflows where curl is the starting point. Think of it like Pastebin or JSFiddle, but for curl commands.<p>If you’ve ever copied a curl from an API doc and wanted a cleaner way to see it or send it to someone else, I’d love your feedback.<p>Thanks!
(You can try it without signup here: <a href="https://uncurl.dev" rel="nofollow">https://uncurl.dev</a>)
Show HN: Lux – A luxurious package manager for Lua
Show HN: Lux – A luxurious package manager for Lua
Show HN: Lux – A luxurious package manager for Lua
Show HN: Browser MCP – Automate your browser using Cursor, Claude, VS Code
Show HN: Browser MCP – Automate your browser using Cursor, Claude, VS Code
Show HN: Browser MCP – Automate your browser using Cursor, Claude, VS Code
Show HN: LocalScore – Local LLM Benchmark
Hey Folks!<p>I've been building an open source benchmark for measuring local LLM performance on your own hardware. The benchmarking tool is a CLI written on top of Llamafile to allow for portability across different hardware setups and operating systems. The website is a database of results from the benchmark, allowing you to explore the performance of different models and hardware configurations.<p>Please give it a try! Any feedback and contribution is much appreciated. I'd love for this to serve as a helpful resource for the local AI community.<p>For more check out:
- Website: <a href="https://localscore.ai" rel="nofollow">https://localscore.ai</a>
- Demo video: <a href="https://youtu.be/De6pA1bQsHU" rel="nofollow">https://youtu.be/De6pA1bQsHU</a>
- Blog post: <a href="https://localscore.ai/blog" rel="nofollow">https://localscore.ai/blog</a>
- CLI Github: <a href="https://github.com/Mozilla-Ocho/llamafile/tree/main/localscore" rel="nofollow">https://github.com/Mozilla-Ocho/llamafile/tree/main/localsco...</a>
- Website Github: <a href="https://github.com/cjpais/localscore" rel="nofollow">https://github.com/cjpais/localscore</a>
Show HN: LocalScore – Local LLM Benchmark
Hey Folks!<p>I've been building an open source benchmark for measuring local LLM performance on your own hardware. The benchmarking tool is a CLI written on top of Llamafile to allow for portability across different hardware setups and operating systems. The website is a database of results from the benchmark, allowing you to explore the performance of different models and hardware configurations.<p>Please give it a try! Any feedback and contribution is much appreciated. I'd love for this to serve as a helpful resource for the local AI community.<p>For more check out:
- Website: <a href="https://localscore.ai" rel="nofollow">https://localscore.ai</a>
- Demo video: <a href="https://youtu.be/De6pA1bQsHU" rel="nofollow">https://youtu.be/De6pA1bQsHU</a>
- Blog post: <a href="https://localscore.ai/blog" rel="nofollow">https://localscore.ai/blog</a>
- CLI Github: <a href="https://github.com/Mozilla-Ocho/llamafile/tree/main/localscore" rel="nofollow">https://github.com/Mozilla-Ocho/llamafile/tree/main/localsco...</a>
- Website Github: <a href="https://github.com/cjpais/localscore" rel="nofollow">https://github.com/cjpais/localscore</a>
Show HN: I built a Rust crate for running unsafe code safely
Show HN: I built a Rust crate for running unsafe code safely
Show HN: Pets for Cursor
I created an extension which adds a custom pet to your editor (Cursor/ VS Code/ Windsurf compatible).<p><a href="https://github.com/blairjordan/codachi" rel="nofollow">https://github.com/blairjordan/codachi</a><p>As you type, your pet hatches, acquires XP and evolves.<p>A playing card is generated when your pet evolves.<p>All pixel art is original and hand/ mouse drawn.
Show HN: iPhone 2005 weird "Blob Keyboard" simulator
Hi HN,<p>I teach tech design history, and one of the key stories I cover is the development of the original iPhone keyboard by Ken Kocienda. Reading about it in his book "Creative Selection" is great, but I wanted my students (and now you!) to actually <i>feel</i> this step in the process.<p>So, I built a web simulator of the "Blob Keyboard", Kocienda's very first attempt at a touchscreen keyboard that actually works, from September 2005:<p>Try the Blob Keyboard: <a href="https://juliendorra.github.io/blob-keyboard-simulator/blob-keyboard-simulator.html" rel="nofollow">https://juliendorra.github.io/blob-keyboard-simulator/blob-k...</a><p>- Tap for the middle letter<p>- Swipe left or right for the side letters<p>More on the github repo: <a href="https://github.com/juliendorra/blob-keyboard-simulator">https://github.com/juliendorra/blob-keyboard-simulator</a><p>The Blob Keyboard prototype emerged during a UX crisis for iPhone team (their software keyboard just didn't work at all, fingers being too big, and the Newton failure loomed over them), highlighting how innovation is rarely a straight path. It was developed on a tethered touchscreen display codenamed "Wallaby".<p>To make this simulator as authentic as possible, I referenced images from Kocienda's book and even got direct feedback and guidance from Ken Kocienda himself on Bluesky.<p>What to expect (or… what not to expect):<p>This is a reconstruction of a very early prototype with limitations reflecting that specific moment. The goal was to test first if typing with accuracy was even possible, as all the rest was moot if it failed!<p>It's NOT QWERTY: They were still hoping to get us out of QWERTY, but then familiarity won.<p>No Backspace: You can't delete.<p>No Cursor Movement: The text field is just a simple display.<p>No Caps or Numbers: Only lowercase letters.<p>No Smooth Animations: Keys just "pop" instantly when pressed. Kocienda noted that your eye fills in the gaps, giving a sense of movement.<p>Best Experience:<p>While it works with a mouse/trackpad on desktop, it's designed for touchscreens to better replicate the original Wallaby hardware interaction. Use it on your phone!<p>This project aims to provide a tangible glimpse into a turning point moment in iPhone development and the iterative nature of design. It's like stepping back in time and trying out that early demo on Kocienda's desk.<p>I would love to hear your reactions and thoughts on experiencing this piece of UI history! What other significant prototype do you wish you could experience?
Show HN: iPhone 2005 weird "Blob Keyboard" simulator
Hi HN,<p>I teach tech design history, and one of the key stories I cover is the development of the original iPhone keyboard by Ken Kocienda. Reading about it in his book "Creative Selection" is great, but I wanted my students (and now you!) to actually <i>feel</i> this step in the process.<p>So, I built a web simulator of the "Blob Keyboard", Kocienda's very first attempt at a touchscreen keyboard that actually works, from September 2005:<p>Try the Blob Keyboard: <a href="https://juliendorra.github.io/blob-keyboard-simulator/blob-keyboard-simulator.html" rel="nofollow">https://juliendorra.github.io/blob-keyboard-simulator/blob-k...</a><p>- Tap for the middle letter<p>- Swipe left or right for the side letters<p>More on the github repo: <a href="https://github.com/juliendorra/blob-keyboard-simulator">https://github.com/juliendorra/blob-keyboard-simulator</a><p>The Blob Keyboard prototype emerged during a UX crisis for iPhone team (their software keyboard just didn't work at all, fingers being too big, and the Newton failure loomed over them), highlighting how innovation is rarely a straight path. It was developed on a tethered touchscreen display codenamed "Wallaby".<p>To make this simulator as authentic as possible, I referenced images from Kocienda's book and even got direct feedback and guidance from Ken Kocienda himself on Bluesky.<p>What to expect (or… what not to expect):<p>This is a reconstruction of a very early prototype with limitations reflecting that specific moment. The goal was to test first if typing with accuracy was even possible, as all the rest was moot if it failed!<p>It's NOT QWERTY: They were still hoping to get us out of QWERTY, but then familiarity won.<p>No Backspace: You can't delete.<p>No Cursor Movement: The text field is just a simple display.<p>No Caps or Numbers: Only lowercase letters.<p>No Smooth Animations: Keys just "pop" instantly when pressed. Kocienda noted that your eye fills in the gaps, giving a sense of movement.<p>Best Experience:<p>While it works with a mouse/trackpad on desktop, it's designed for touchscreens to better replicate the original Wallaby hardware interaction. Use it on your phone!<p>This project aims to provide a tangible glimpse into a turning point moment in iPhone development and the iterative nature of design. It's like stepping back in time and trying out that early demo on Kocienda's desk.<p>I would love to hear your reactions and thoughts on experiencing this piece of UI history! What other significant prototype do you wish you could experience?
Show HN: OCR pipeline for ML training (tables, diagrams, math, multilingual)
Hi HN,<p>I’ve been working on an OCR pipeline specifically optimized for machine learning dataset preparation. It’s designed to process complex academic materials — including math formulas, tables, figures, and multilingual text — and output clean, structured formats like JSON and Markdown.<p>Some features:
• Multi-stage OCR combining DocLayout-YOLO, Google Vision, MathPix, and Gemini Pro Vision
• Extracts and understands diagrams, tables, LaTeX-style math, and multilingual text (Japanese/Korean/English)
• Highly tuned for ML training pipelines, including dataset generation and preprocessing for RAG or fine-tuning tasks<p>Sample outputs and real exam-based examples are included (EJU Biology, UTokyo Math, etc.)
Would love to hear any feedback or ideas for improvement.<p>GitHub: <a href="https://github.com/ses4255/Versatile-OCR-Program">https://github.com/ses4255/Versatile-OCR-Program</a>
Show HN: OCR pipeline for ML training (tables, diagrams, math, multilingual)
Hi HN,<p>I’ve been working on an OCR pipeline specifically optimized for machine learning dataset preparation. It’s designed to process complex academic materials — including math formulas, tables, figures, and multilingual text — and output clean, structured formats like JSON and Markdown.<p>Some features:
• Multi-stage OCR combining DocLayout-YOLO, Google Vision, MathPix, and Gemini Pro Vision
• Extracts and understands diagrams, tables, LaTeX-style math, and multilingual text (Japanese/Korean/English)
• Highly tuned for ML training pipelines, including dataset generation and preprocessing for RAG or fine-tuning tasks<p>Sample outputs and real exam-based examples are included (EJU Biology, UTokyo Math, etc.)
Would love to hear any feedback or ideas for improvement.<p>GitHub: <a href="https://github.com/ses4255/Versatile-OCR-Program">https://github.com/ses4255/Versatile-OCR-Program</a>