The best Hacker News stories from Show from the past day

Go back

Latest posts:

Show HN: Firecrawl-Simple – Stable fork of Firecrawl optimized for self-hosting

Firecrawl Simple is a stripped down and stable version of firecrawl optimized for self-hosting and ease of contribution.<p>The upstream firecrawl repo contains the following blurb:<p>>This repository is in development, and we're still integrating custom modules into the mono repo. It's not fully ready for self-hosted deployment yet, but you can run it locally.<p>Firecrawl's API surface and general functionality were ideal for our Trieve sitesearch product, but we needed a version ready for self-hosting that was easy to contribute to and scale on Kubernetes. Therefore, we decided to fork and begin maintaining a stripped down, stable version.<p>Fire-engine, Firecrawl's solution for anti-bot pages, being closed source is the biggest deal breaker requiring us to maintain this fork. Further, our purposes not requiring the SaaS and AI dependencies also pushes our use-case far enough away from Firecrawl's current mission that it doesn't seem like merging into the upstream is viable at this time.

Show HN: Term-Lisp – A Lisp, based on pattern matching and term rewriting

Show HN: Term-Lisp – A Lisp, based on pattern matching and term rewriting

Show HN: Term-Lisp – A Lisp, based on pattern matching and term rewriting

Show HN: SuperSplat – open-source 3D Gaussian Splat Editor

Show HN: SuperSplat – open-source 3D Gaussian Splat Editor

Show HN: SuperSplat – open-source 3D Gaussian Splat Editor

Show HN: Aide, an open-source AI native IDE

Hey HN, We are Sandeep and Naresh, the creators of Aide. We are happy to open source and invite the community to try out Aide which is a VSCode fork built with LLMs integrated.<p>To talk through the features, we engineered the following:<p>- A proactive agent<p>Agent which iterates on the linter errors (powered by the Language Server) and pulls in relevant context by doing go-to-definitions, go-to-references etc and propose fixes or ask for more files which might be missing in the context.<p>- Developer control<p>We encourage you to do edits on top of your coding sessions. To enable this, we built a VSCode native rollback feature which gets rid of all the edits made by the agent in a single click if there were mistakes, without messing up your changes from before.<p>- A combined chat+edit flow which you can use to brainstorm and edit<p>You can brainstorm a problem in chat by @’ting the files and then jump into edits (which can happen across multiple files) or go from a smaller set of edits and discuss the side-effects of it<p>- Inline editing widget<p>We took inspiration from the macos spotlight widget and created a similar one inside the editor, you can highlight part of the code, do Cmd+K and just give your instructions freely<p>- Local running AI brain<p>We ship a binary called sidecar which takes care of talking to the LLM providers, preparing the prompts and using the editor for the LLM. All of this is local first and you get full control over the prompts/responses without anything leaking to our end (unless you choose to use your subscription and share the data with us)<p>We spent the last 15 months learning about the internals of VSCode (its a non-trivial codebase) and also powering up our AI game, the framework is also at the top of swebench-lite with 43% score. On top of this, since the whole AI side of the logic runs locally on your machine you have complete control over the data, from the prompt to the responses and you can use your own API Keys as well (can be any LLM provider) and talk to them directly.<p>There’s still a whole lot to build and we are at 1% of the journey. Right now the editor feels robust and does not break on any of the flows which we aimed to solve for.<p>Let us know if there’s anything else you would like to see us build. We also want to empower extensibility and work together with the community to build the next set of features and set a new milestone of AI native editors.

Show HN: Aide, an open-source AI native IDE

Hey HN, We are Sandeep and Naresh, the creators of Aide. We are happy to open source and invite the community to try out Aide which is a VSCode fork built with LLMs integrated.<p>To talk through the features, we engineered the following:<p>- A proactive agent<p>Agent which iterates on the linter errors (powered by the Language Server) and pulls in relevant context by doing go-to-definitions, go-to-references etc and propose fixes or ask for more files which might be missing in the context.<p>- Developer control<p>We encourage you to do edits on top of your coding sessions. To enable this, we built a VSCode native rollback feature which gets rid of all the edits made by the agent in a single click if there were mistakes, without messing up your changes from before.<p>- A combined chat+edit flow which you can use to brainstorm and edit<p>You can brainstorm a problem in chat by @’ting the files and then jump into edits (which can happen across multiple files) or go from a smaller set of edits and discuss the side-effects of it<p>- Inline editing widget<p>We took inspiration from the macos spotlight widget and created a similar one inside the editor, you can highlight part of the code, do Cmd+K and just give your instructions freely<p>- Local running AI brain<p>We ship a binary called sidecar which takes care of talking to the LLM providers, preparing the prompts and using the editor for the LLM. All of this is local first and you get full control over the prompts/responses without anything leaking to our end (unless you choose to use your subscription and share the data with us)<p>We spent the last 15 months learning about the internals of VSCode (its a non-trivial codebase) and also powering up our AI game, the framework is also at the top of swebench-lite with 43% score. On top of this, since the whole AI side of the logic runs locally on your machine you have complete control over the data, from the prompt to the responses and you can use your own API Keys as well (can be any LLM provider) and talk to them directly.<p>There’s still a whole lot to build and we are at 1% of the journey. Right now the editor feels robust and does not break on any of the flows which we aimed to solve for.<p>Let us know if there’s anything else you would like to see us build. We also want to empower extensibility and work together with the community to build the next set of features and set a new milestone of AI native editors.

Show HN: Aide, an open-source AI native IDE

Hey HN, We are Sandeep and Naresh, the creators of Aide. We are happy to open source and invite the community to try out Aide which is a VSCode fork built with LLMs integrated.<p>To talk through the features, we engineered the following:<p>- A proactive agent<p>Agent which iterates on the linter errors (powered by the Language Server) and pulls in relevant context by doing go-to-definitions, go-to-references etc and propose fixes or ask for more files which might be missing in the context.<p>- Developer control<p>We encourage you to do edits on top of your coding sessions. To enable this, we built a VSCode native rollback feature which gets rid of all the edits made by the agent in a single click if there were mistakes, without messing up your changes from before.<p>- A combined chat+edit flow which you can use to brainstorm and edit<p>You can brainstorm a problem in chat by @’ting the files and then jump into edits (which can happen across multiple files) or go from a smaller set of edits and discuss the side-effects of it<p>- Inline editing widget<p>We took inspiration from the macos spotlight widget and created a similar one inside the editor, you can highlight part of the code, do Cmd+K and just give your instructions freely<p>- Local running AI brain<p>We ship a binary called sidecar which takes care of talking to the LLM providers, preparing the prompts and using the editor for the LLM. All of this is local first and you get full control over the prompts/responses without anything leaking to our end (unless you choose to use your subscription and share the data with us)<p>We spent the last 15 months learning about the internals of VSCode (its a non-trivial codebase) and also powering up our AI game, the framework is also at the top of swebench-lite with 43% score. On top of this, since the whole AI side of the logic runs locally on your machine you have complete control over the data, from the prompt to the responses and you can use your own API Keys as well (can be any LLM provider) and talk to them directly.<p>There’s still a whole lot to build and we are at 1% of the journey. Right now the editor feels robust and does not break on any of the flows which we aimed to solve for.<p>Let us know if there’s anything else you would like to see us build. We also want to empower extensibility and work together with the community to build the next set of features and set a new milestone of AI native editors.

Show HN: Hacker News frontpage as a print newspaper that you can personalize

Show HN: Hacker News frontpage as a print newspaper that you can personalize

Show HN: Hacker News frontpage as a print newspaper that you can personalize

Show HN: Hacker News frontpage as a print newspaper that you can personalize

Rd-TableBench – Accurately evaluating table extraction

Hey HN!<p>A ton of document parsing solutions have been coming out lately, each claiming SOTA with little evidence. A lot of these turned out to be LLM or LVM wrappers that hallucinate frequently on complex tables.<p>We just released RD-TableBench, an open benchmark to help teams evaluate extraction performance for complex tables. The benchmark includes a variety of challenging scenarios including scanned tables, handwriting, language detection, merged cells, and more.<p>We employed an independent team of PhD-level human labelers who manually annotated 1000 complex table images from a diverse set of publicly available documents.<p>Alongside this, we also release a new bioinformatics inspired algorithm for grading table similarity. Would love to hear any feedback!<p>-Raunak

Show HN: Varse – Simple remote application config

Hey HN,<p>We wanted a simple way to update application configs, without redeploying. We wanted to remotely set variables, and read them in our application. We created Varse to do this.<p>Varse has a dashboard for creating key - value pairs and an SDK to read them. It's un-opinionated and allows for strings, booleans, or even json objects to be stored.<p>We have instructions for running it yourself. We also have a hosted version where you can create an account and manage variables.<p>This is our first time building an open source project. We'd love feedback on how to do it right.<p>Github: <a href="https://github.com/varse-io/varse">https://github.com/varse-io/varse</a><p>Hosted Version: <a href="https://app.varse.io/signup" rel="nofollow">https://app.varse.io/signup</a><p>Website: <a href="https://www.varse.io/" rel="nofollow">https://www.varse.io/</a><p>Contact: izak@varse.io

Show HN: Varse – Simple remote application config

Hey HN,<p>We wanted a simple way to update application configs, without redeploying. We wanted to remotely set variables, and read them in our application. We created Varse to do this.<p>Varse has a dashboard for creating key - value pairs and an SDK to read them. It's un-opinionated and allows for strings, booleans, or even json objects to be stored.<p>We have instructions for running it yourself. We also have a hosted version where you can create an account and manage variables.<p>This is our first time building an open source project. We'd love feedback on how to do it right.<p>Github: <a href="https://github.com/varse-io/varse">https://github.com/varse-io/varse</a><p>Hosted Version: <a href="https://app.varse.io/signup" rel="nofollow">https://app.varse.io/signup</a><p>Website: <a href="https://www.varse.io/" rel="nofollow">https://www.varse.io/</a><p>Contact: izak@varse.io

Show HN: rallyup – Lightweight Wake-on-LAN Scheduler

Hi HN,<p>I’ve wanted a simple solution to handle Wake-on-LAN sequences for my home and work labs to boot up servers in the right order. I was already dabbling in Rust and thought this would be an interesting project to dive deeper and see if it could work well for this kind of network tool. The result is rallyup.<p>rallyup lets you set up server dependencies in a YAML file, so each service (e.g., firewalls, storage, VM hosts) comes online in the right order. It verifies each server’s status before moving to the next.<p>Features:<p>- Dependency-based WOL with VLAN support - Built-in health checks (HTTP, open ports, shell commands) - Lightweight enough to run on a Raspberry Pi or similar device<p>Would love any feedback. Thanks for taking a look!

Show HN: rallyup – Lightweight Wake-on-LAN Scheduler

Hi HN,<p>I’ve wanted a simple solution to handle Wake-on-LAN sequences for my home and work labs to boot up servers in the right order. I was already dabbling in Rust and thought this would be an interesting project to dive deeper and see if it could work well for this kind of network tool. The result is rallyup.<p>rallyup lets you set up server dependencies in a YAML file, so each service (e.g., firewalls, storage, VM hosts) comes online in the right order. It verifies each server’s status before moving to the next.<p>Features:<p>- Dependency-based WOL with VLAN support - Built-in health checks (HTTP, open ports, shell commands) - Lightweight enough to run on a Raspberry Pi or similar device<p>Would love any feedback. Thanks for taking a look!

Show HN: Whirlwind – Async concurrent hashmap for Rust

Hey HN, this is Will and David from Fortress (<a href="https://news.ycombinator.com/item?id=41426998">https://news.ycombinator.com/item?id=41426998</a>).<p>We use a lot of async Rust internally, and created this library out of a need for an async-aware concurrent hashmap since there weren’t many available in the Rust ecosystem.<p>Whirlwind is a sharded HashMap with a fully asynchronous API. Just as dashmap is a replacement for std::sync::RwLock<HashMap>, whirlwind aims to be a replacement for tokio::sync::RwLock<HashMap>. It has a similar design and performance characteristics to dashmap, but seems to perform better in read-heavy workloads with tokio's green threading.<p>Benchmarks are in the readme! We used an asyncified version of dashmap's benchmark suite. The project is in a pretty early stage and I'm sure there are flaws, but I'm pretty happy with the performance.<p>There is some unsafe involved, but we run Miri in ci to (hopefully) catch undefined behavior well before it's in an actual release.<p>We'd appreciate any feedback! Thanks in advance :)

< 1 2 3 ... 153 154 155 156 157 ... 864 865 866 >