The best Hacker News stories from All from the past day

Go back

Latest posts:

Hurl, the Exceptional Language

The t-test was invented at the Guinness brewery

To the brain, reading computer code is not the same as reading language (2020)

How Home Assistant is being used to protect from missile and drone attacks

What the damaged Svalbard cable looked like

What the damaged Svalbard cable looked like

Google Meet rolls out multi-device adaptive audio merging

Helen Keller on her life before self-consciousness (1908)

Cloudflare took down our website

Google just updated its algorithm. The Internet will never be the same

Abusing Go's Infrastructure

Abusing Go's Infrastructure

Tmux is worse-is-better

Voxel Displacement Renderer – Modernizing the Retro 3D Aesthetic

The hikikomori in Asia: A life within four walls

Google scrambles to manually remove weird AI answers in search

Mp3tag – Universal Tag Editor

Mp3tag – Universal Tag Editor

Show HN: We open sourced our entire text-to-SQL product

Long story short: We (Dataherald) just open-sourced our entire codebase, including the core engine, the clients that interact with it and the backend application layer for authentication and RBAC. You can now use the full solution to build text-to-SQL into your product.<p>The Problem: modern LLMs write syntactically correct SQL, but they struggle with real-world relational data. This is because real world data and schema is messy, natural language can often be ambiguous and LLMs are not trained on your specific dataset.<p>Solution: The core NL-to-SQL engine in Dataherald is an LLM based agent which uses Chain of Thought (CoT) reasoning and a number of different tools to generate high accuracy SQL from a given user prompt. The engine achieves this by:<p>- Collecting context at configuration from the database and sources such as data dictionaries and unstructured documents which are stored in a data store or a vector DB and injected if relevant<p>- Allowing users to upload sample NL <> SQL pairs (golden SQL) which can be used in few shot prompting or to fine-tune an NL-to-SQL LLM for that specific dataset<p>- Executing the SQL against the DB to get a few sample rows and recover from errors<p>- Using an evaluator to assign a confidence score to the generated SQL<p>The repo includes four services <a href="https://github.com/Dataherald/dataherald/tree/main/services">https://github.com/Dataherald/dataherald/tree/main/services</a>:<p>1- Engine: The core service which includes the LLM agent, vector stores and DB connectors.<p>2- Admin Console: a NextJS front-end for configuring the engine and observability.<p>3- Enterprise Backend: Wraps the core engine, adding authentication, caching, and APIs for the frontend.<p>4- Slackbot: Integrate Dataherald directly into your Slack workflow for on-the-fly data exploration.<p>Would love to hear from the community on building natural language interfaces to relational data. Anyone live in production without a human in the loop? Thoughts on how to improve performance without spending weeks on model training?

Show HN: We open sourced our entire text-to-SQL product

Long story short: We (Dataherald) just open-sourced our entire codebase, including the core engine, the clients that interact with it and the backend application layer for authentication and RBAC. You can now use the full solution to build text-to-SQL into your product.<p>The Problem: modern LLMs write syntactically correct SQL, but they struggle with real-world relational data. This is because real world data and schema is messy, natural language can often be ambiguous and LLMs are not trained on your specific dataset.<p>Solution: The core NL-to-SQL engine in Dataherald is an LLM based agent which uses Chain of Thought (CoT) reasoning and a number of different tools to generate high accuracy SQL from a given user prompt. The engine achieves this by:<p>- Collecting context at configuration from the database and sources such as data dictionaries and unstructured documents which are stored in a data store or a vector DB and injected if relevant<p>- Allowing users to upload sample NL <> SQL pairs (golden SQL) which can be used in few shot prompting or to fine-tune an NL-to-SQL LLM for that specific dataset<p>- Executing the SQL against the DB to get a few sample rows and recover from errors<p>- Using an evaluator to assign a confidence score to the generated SQL<p>The repo includes four services <a href="https://github.com/Dataherald/dataherald/tree/main/services">https://github.com/Dataherald/dataherald/tree/main/services</a>:<p>1- Engine: The core service which includes the LLM agent, vector stores and DB connectors.<p>2- Admin Console: a NextJS front-end for configuring the engine and observability.<p>3- Enterprise Backend: Wraps the core engine, adding authentication, caching, and APIs for the frontend.<p>4- Slackbot: Integrate Dataherald directly into your Slack workflow for on-the-fly data exploration.<p>Would love to hear from the community on building natural language interfaces to relational data. Anyone live in production without a human in the loop? Thoughts on how to improve performance without spending weeks on model training?

< 1 2 3 ... 205 206 207 208 209 ... 829 830 831 >