The best Hacker News stories from Show from the past day

Go back

Latest posts:

Show HN: AgentMail – Email infra for AI agents

Hey HN, we're Haakam, Michael, and Adi. We're building AgentMail (<a href="https://agentmail.to/">https://agentmail.to/</a>), an API to give AI agents their own email inboxes. We’re not talking about AI for your email, this is email for your AI.<p>We started building email agents because they can converse with users in their inboxes, automate email-based workflows, and authenticate with third-party applications. Given these unique capabilities, we think email will be a core interface for agents.<p>But we were building on top of Gmail, which was a struggle: poor API support, expensive subscriptions, rate limits, sending limits, GCP Pub/Sub, OAuth, crappy keyword search, and an overall terrible developer experience.<p>Gmail and other providers didn’t work for us. So we decided to bite the bullet and build our own.<p>AgentMail is like Gmail, but API-first, with programmatic inbox creation, events over webhooks and websockets, simple API key auth, organization-wide semantic search, structured data extraction, and usage-based pricing that scales with emails sent/received.<p>Here’s a demo of building an email agent: <a href="https://youtu.be/1V7BISeFUTM" rel="nofollow">https://youtu.be/1V7BISeFUTM</a>, and here’s a demo of a voice agent with its own email inbox: <a href="https://youtu.be/eG2fCsRK4RY" rel="nofollow">https://youtu.be/eG2fCsRK4RY</a><p>So far AgentMail has been deployed to use cases such as apps with dedicated inboxes for each user, voice agents that receive documents in real time, automated account provisioning and QA testing, cold outbound platforms with thousands of inboxes, automations for processing invoices, and agents that coordinate work with humans and other agents.<p>We would love to hear your thoughts and feedback. You can try our playground at <a href="https://chat.agentmail.to">https://chat.agentmail.to</a>

Show HN: Mcp-use – Connect any LLM to any MCP

Hey Pietro and Luigi here, we are the authors of mcp-use (<a href="https://github.com/mcp-use/mcp-use">https://github.com/mcp-use/mcp-use</a>).<p>When the first MCP servers came out we were very excited about the technology, but as soon as we wanted to get our hands dirty, we found out that MCP could be used only through Claude Desktop or Cursor. As engineers, we did not like that. MCP seemed like something you wanted to use to build products and applications yourself, not something to hide behind a closed source application.<p>So we approached the SDK but were pretty dissatisfied with the developer experience (double async loops, lots of boilerplate). We decided to write mcp-use to make our lives easier.<p>mcp-use lets you connect any LLM to any MCP server in just 6 lines of code. We provide a high level abstraction over the official MCP SDK that makes your life easier and supports all the functionalities of the protocol.<p>Demo video here: <a href="https://www.youtube.com/watch?v=nL_B6LZAsp4" rel="nofollow">https://www.youtube.com/watch?v=nL_B6LZAsp4</a>.<p>The key abstractions we provide are called MCPClient and MCPAgent.<p>MCPClient takes in a set of server configurations, automatically detects the transport type and creates a background task which handles the stream from/to the server.<p>MCPAgent is a combination of the MCPClient, an LLM, and a custom system prompt. It consumes the MCP client by transforming the tools, resources and prompts into model agnostic tools that can be called by the LLM.<p>The library also contains some cool utilities:<p>- secure sandboxed execution of MCP servers (we know the protocol doesn't shine for security)<p>- meta-tools that allow the agent to search over available servers and tools (to avoid context flooding) and connect dynamically to the server it needs (you could create the omnipotent agent with this).<p>Some cool things we did with this: - write an agent that can use a browser and create/read linear tickets updated with latest information on the internet<p>- write an agent that has access to the metrics of our company to automatically create weekly reports.<p>- I connected an agent to an IKEA curtain I hacked an MCP on to adapt the lighting of my room from images of the lighting situation.<p>- recreated am open source claude code like CLI, with full MCP capability but with custom models and BYOK (<a href="https://github.com/mcp-use/mcp-use-cli">https://github.com/mcp-use/mcp-use-cli</a>).<p>We recently crossed 100,000 download and we are used by many organizations, including NASA!<p>We’d love to hear what you think of it, most importantly how we can improve it! We are happy to answer any questions and look forward to your comments.

Show HN: Mcp-use – Connect any LLM to any MCP

Hey Pietro and Luigi here, we are the authors of mcp-use (<a href="https://github.com/mcp-use/mcp-use">https://github.com/mcp-use/mcp-use</a>).<p>When the first MCP servers came out we were very excited about the technology, but as soon as we wanted to get our hands dirty, we found out that MCP could be used only through Claude Desktop or Cursor. As engineers, we did not like that. MCP seemed like something you wanted to use to build products and applications yourself, not something to hide behind a closed source application.<p>So we approached the SDK but were pretty dissatisfied with the developer experience (double async loops, lots of boilerplate). We decided to write mcp-use to make our lives easier.<p>mcp-use lets you connect any LLM to any MCP server in just 6 lines of code. We provide a high level abstraction over the official MCP SDK that makes your life easier and supports all the functionalities of the protocol.<p>Demo video here: <a href="https://www.youtube.com/watch?v=nL_B6LZAsp4" rel="nofollow">https://www.youtube.com/watch?v=nL_B6LZAsp4</a>.<p>The key abstractions we provide are called MCPClient and MCPAgent.<p>MCPClient takes in a set of server configurations, automatically detects the transport type and creates a background task which handles the stream from/to the server.<p>MCPAgent is a combination of the MCPClient, an LLM, and a custom system prompt. It consumes the MCP client by transforming the tools, resources and prompts into model agnostic tools that can be called by the LLM.<p>The library also contains some cool utilities:<p>- secure sandboxed execution of MCP servers (we know the protocol doesn't shine for security)<p>- meta-tools that allow the agent to search over available servers and tools (to avoid context flooding) and connect dynamically to the server it needs (you could create the omnipotent agent with this).<p>Some cool things we did with this: - write an agent that can use a browser and create/read linear tickets updated with latest information on the internet<p>- write an agent that has access to the metrics of our company to automatically create weekly reports.<p>- I connected an agent to an IKEA curtain I hacked an MCP on to adapt the lighting of my room from images of the lighting situation.<p>- recreated am open source claude code like CLI, with full MCP capability but with custom models and BYOK (<a href="https://github.com/mcp-use/mcp-use-cli">https://github.com/mcp-use/mcp-use-cli</a>).<p>We recently crossed 100,000 download and we are used by many organizations, including NASA!<p>We’d love to hear what you think of it, most importantly how we can improve it! We are happy to answer any questions and look forward to your comments.

Show HN: Mcp-use – Connect any LLM to any MCP

Hey Pietro and Luigi here, we are the authors of mcp-use (<a href="https://github.com/mcp-use/mcp-use">https://github.com/mcp-use/mcp-use</a>).<p>When the first MCP servers came out we were very excited about the technology, but as soon as we wanted to get our hands dirty, we found out that MCP could be used only through Claude Desktop or Cursor. As engineers, we did not like that. MCP seemed like something you wanted to use to build products and applications yourself, not something to hide behind a closed source application.<p>So we approached the SDK but were pretty dissatisfied with the developer experience (double async loops, lots of boilerplate). We decided to write mcp-use to make our lives easier.<p>mcp-use lets you connect any LLM to any MCP server in just 6 lines of code. We provide a high level abstraction over the official MCP SDK that makes your life easier and supports all the functionalities of the protocol.<p>Demo video here: <a href="https://www.youtube.com/watch?v=nL_B6LZAsp4" rel="nofollow">https://www.youtube.com/watch?v=nL_B6LZAsp4</a>.<p>The key abstractions we provide are called MCPClient and MCPAgent.<p>MCPClient takes in a set of server configurations, automatically detects the transport type and creates a background task which handles the stream from/to the server.<p>MCPAgent is a combination of the MCPClient, an LLM, and a custom system prompt. It consumes the MCP client by transforming the tools, resources and prompts into model agnostic tools that can be called by the LLM.<p>The library also contains some cool utilities:<p>- secure sandboxed execution of MCP servers (we know the protocol doesn't shine for security)<p>- meta-tools that allow the agent to search over available servers and tools (to avoid context flooding) and connect dynamically to the server it needs (you could create the omnipotent agent with this).<p>Some cool things we did with this: - write an agent that can use a browser and create/read linear tickets updated with latest information on the internet<p>- write an agent that has access to the metrics of our company to automatically create weekly reports.<p>- I connected an agent to an IKEA curtain I hacked an MCP on to adapt the lighting of my room from images of the lighting situation.<p>- recreated am open source claude code like CLI, with full MCP capability but with custom models and BYOK (<a href="https://github.com/mcp-use/mcp-use-cli">https://github.com/mcp-use/mcp-use-cli</a>).<p>We recently crossed 100,000 download and we are used by many organizations, including NASA!<p>We’d love to hear what you think of it, most importantly how we can improve it! We are happy to answer any questions and look forward to your comments.

Show HN: Mcp-use – Connect any LLM to any MCP

Hey Pietro and Luigi here, we are the authors of mcp-use (<a href="https://github.com/mcp-use/mcp-use">https://github.com/mcp-use/mcp-use</a>).<p>When the first MCP servers came out we were very excited about the technology, but as soon as we wanted to get our hands dirty, we found out that MCP could be used only through Claude Desktop or Cursor. As engineers, we did not like that. MCP seemed like something you wanted to use to build products and applications yourself, not something to hide behind a closed source application.<p>So we approached the SDK but were pretty dissatisfied with the developer experience (double async loops, lots of boilerplate). We decided to write mcp-use to make our lives easier.<p>mcp-use lets you connect any LLM to any MCP server in just 6 lines of code. We provide a high level abstraction over the official MCP SDK that makes your life easier and supports all the functionalities of the protocol.<p>Demo video here: <a href="https://www.youtube.com/watch?v=nL_B6LZAsp4" rel="nofollow">https://www.youtube.com/watch?v=nL_B6LZAsp4</a>.<p>The key abstractions we provide are called MCPClient and MCPAgent.<p>MCPClient takes in a set of server configurations, automatically detects the transport type and creates a background task which handles the stream from/to the server.<p>MCPAgent is a combination of the MCPClient, an LLM, and a custom system prompt. It consumes the MCP client by transforming the tools, resources and prompts into model agnostic tools that can be called by the LLM.<p>The library also contains some cool utilities:<p>- secure sandboxed execution of MCP servers (we know the protocol doesn't shine for security)<p>- meta-tools that allow the agent to search over available servers and tools (to avoid context flooding) and connect dynamically to the server it needs (you could create the omnipotent agent with this).<p>Some cool things we did with this: - write an agent that can use a browser and create/read linear tickets updated with latest information on the internet<p>- write an agent that has access to the metrics of our company to automatically create weekly reports.<p>- I connected an agent to an IKEA curtain I hacked an MCP on to adapt the lighting of my room from images of the lighting situation.<p>- recreated am open source claude code like CLI, with full MCP capability but with custom models and BYOK (<a href="https://github.com/mcp-use/mcp-use-cli">https://github.com/mcp-use/mcp-use-cli</a>).<p>We recently crossed 100,000 download and we are used by many organizations, including NASA!<p>We’d love to hear what you think of it, most importantly how we can improve it! We are happy to answer any questions and look forward to your comments.

Show HN: A high-altitude low-power flight computer for high-altitude balloons

I've been working on this for a while now, and I'm happy to share!<p>I've been into launching weather balloons for a few years. One aspect of the hobby that really drew me in was the tracking side of things. Tracking systems let you follow the balloon's position throughout the flight and, most importantly, know exactly where it lands so you can recover the instrumentation. APRS is what I started out using during my first few years in the hobby, after I got my amateur radio license in 2020 (W0MXX). I designed a few small boards using the trackuino (<a href="https://github.com/trackuino/trackuino">https://github.com/trackuino/trackuino</a>) firmware (while breaking 3 $70 radio modules along the way).<p>I then got into recovering radiosondes, which are launched twice per day by the NWS and can be reprogrammed using RS41ng (<a href="https://github.com/mikaelnousiainen/RS41ng">https://github.com/mikaelnousiainen/RS41ng</a>) to run many amateur radio tracking protocols. I was a bit dissatisfied with how large and heavy the radiosonde trackers were, so I designed my own tracking system, called Tiny4FSK.<p>Tiny4FSK is a flight computer with built-in tracking using the Horus Binary v2 tracking system. This protocol was developed by the Project Horus team specifically for high-altitude balloons, and it brings features like high transmit rates, forward error correction, and excellent weak-signal performance in an open source package. It's designed to be as compact as possible and can run on a single AA battery for upwards of 17 hours.<p>The main board comes with header rows that allow for out-of-the-box expansion. I developed a shield that supports the BME280 environmental sensor, the ICM-20948 9-axis IMU, and more via the Qwiic connector. It also features an OLED display for basic diagnostics.<p>While I've pretty much polished the main tracking procedures (and have tested on multiple flights), I'm still developing the IMU code using a lightweight Kalman filter. Additionally, there isn't yet a wide network of Horus Binary decoding stations like the APRS network has (I-gates), but I hope that by promoting this protocol, more stations will pop up. This means that if you're not in an area with many receive stations, you'll need to set up your own using either Horus-GUI (<a href="https://github.com/projecthorus/horus-gui">https://github.com/projecthorus/horus-gui</a>) or horusdemodlib (<a href="https://github.com/projecthorus/horusdemodlib">https://github.com/projecthorus/horusdemodlib</a>).<p>One issue I’m still working on is improving RF signal strength. Although the protocol is decodable in very low-noise environments, the transmit power appears to be lower than that of a typical radiosonde. This could be due to several factors: limited current on a weak power source (signal is stronger when powered from a bench supply), off-tuned filtering/matching, or not paying enough attention to the antenna. I'm planning to run more simulations to figure this out. That said, the signal is still decodable from the ground even at max altitude (~100,000 feet).<p>On the more technical side, Tiny4FSK uses: - the SAMD21 microcontroller, which is an ARM Cortex-M0+ MCU - the TPS61200 boost converter, which is adjusted to output 3.3v - Si4063 radio module, which I use on the 70cm band - ATGM336H gps module - pretty cheap GPS module which works in airborne mode (>18km) - integrated BME280 temperature, pressure, and humidity sensor The code uses the Arduino framework to make it accessible to beginners.<p>All flights using Horus Binary v2, including reprogrammed radiosondes, other custom trackers, and Tiny4FSK show up in real-time on Sondehub Amateur (<a href="https://amateur.sondehub.org" rel="nofollow">https://amateur.sondehub.org</a>). Flight data can be found in the /Media/Data folder on Github (there's several missing flights on there though).<p>Thanks for reading, hope I didn’t mess anything up too badly in the post! -Max

Show HN: A high-altitude low-power flight computer for high-altitude balloons

I've been working on this for a while now, and I'm happy to share!<p>I've been into launching weather balloons for a few years. One aspect of the hobby that really drew me in was the tracking side of things. Tracking systems let you follow the balloon's position throughout the flight and, most importantly, know exactly where it lands so you can recover the instrumentation. APRS is what I started out using during my first few years in the hobby, after I got my amateur radio license in 2020 (W0MXX). I designed a few small boards using the trackuino (<a href="https://github.com/trackuino/trackuino">https://github.com/trackuino/trackuino</a>) firmware (while breaking 3 $70 radio modules along the way).<p>I then got into recovering radiosondes, which are launched twice per day by the NWS and can be reprogrammed using RS41ng (<a href="https://github.com/mikaelnousiainen/RS41ng">https://github.com/mikaelnousiainen/RS41ng</a>) to run many amateur radio tracking protocols. I was a bit dissatisfied with how large and heavy the radiosonde trackers were, so I designed my own tracking system, called Tiny4FSK.<p>Tiny4FSK is a flight computer with built-in tracking using the Horus Binary v2 tracking system. This protocol was developed by the Project Horus team specifically for high-altitude balloons, and it brings features like high transmit rates, forward error correction, and excellent weak-signal performance in an open source package. It's designed to be as compact as possible and can run on a single AA battery for upwards of 17 hours.<p>The main board comes with header rows that allow for out-of-the-box expansion. I developed a shield that supports the BME280 environmental sensor, the ICM-20948 9-axis IMU, and more via the Qwiic connector. It also features an OLED display for basic diagnostics.<p>While I've pretty much polished the main tracking procedures (and have tested on multiple flights), I'm still developing the IMU code using a lightweight Kalman filter. Additionally, there isn't yet a wide network of Horus Binary decoding stations like the APRS network has (I-gates), but I hope that by promoting this protocol, more stations will pop up. This means that if you're not in an area with many receive stations, you'll need to set up your own using either Horus-GUI (<a href="https://github.com/projecthorus/horus-gui">https://github.com/projecthorus/horus-gui</a>) or horusdemodlib (<a href="https://github.com/projecthorus/horusdemodlib">https://github.com/projecthorus/horusdemodlib</a>).<p>One issue I’m still working on is improving RF signal strength. Although the protocol is decodable in very low-noise environments, the transmit power appears to be lower than that of a typical radiosonde. This could be due to several factors: limited current on a weak power source (signal is stronger when powered from a bench supply), off-tuned filtering/matching, or not paying enough attention to the antenna. I'm planning to run more simulations to figure this out. That said, the signal is still decodable from the ground even at max altitude (~100,000 feet).<p>On the more technical side, Tiny4FSK uses: - the SAMD21 microcontroller, which is an ARM Cortex-M0+ MCU - the TPS61200 boost converter, which is adjusted to output 3.3v - Si4063 radio module, which I use on the 70cm band - ATGM336H gps module - pretty cheap GPS module which works in airborne mode (>18km) - integrated BME280 temperature, pressure, and humidity sensor The code uses the Arduino framework to make it accessible to beginners.<p>All flights using Horus Binary v2, including reprogrammed radiosondes, other custom trackers, and Tiny4FSK show up in real-time on Sondehub Amateur (<a href="https://amateur.sondehub.org" rel="nofollow">https://amateur.sondehub.org</a>). Flight data can be found in the /Media/Data folder on Github (there's several missing flights on there though).<p>Thanks for reading, hope I didn’t mess anything up too badly in the post! -Max

Show HN: An AI agent that learns your product and guides your users

Hey HN! My name is Christian, and I’m the co-founder of <a href="https://frigade.ai">https://frigade.ai</a>. We’ve built an AI agent that automatically learns how to use any web-based product, and in turn guides users directly in the UI, automatically generates documentation, and even takes actions on a user’s behalf. Think of it as Clippy from the old MS Office. But on steroids. And actually helpful.<p>You can see the agent and tool-calling SDK in action here: <a href="https://www.youtube.com/watch?v=UPe0t3A1Vpg" rel="nofollow">https://www.youtube.com/watch?v=UPe0t3A1Vpg</a><p>How is this different from other AI customer support products?<p>Most AI "copilots" are really just glorified chatbots. They skim your help center and spit out some nonspecific bullet points. Basically some ‘hopes and prayers’ that your users will figure it out. Ultimately, this puts the burden on the user to follow through. And assumes companies are keeping their help center up-to-date with every product change. That means constant screenshots of new product UI or features for accurate instructions.These solutions leverage only a fraction of what’s possible with AI, which can now reason about software interfaces extensively.<p>With Frigade AI, we guide the user directly in the product and build on-demand tours based on the current user’s state and context. The agents can also take actions immediately on a user’s behalf, e.g. inviting a colleague to a workspace or retrieving billing information (via our tool calling SDK).<p>This was only made possible recently. The latest frontier models (GPT 4.1, Claude 4, Gemini 2.5, etc.) are able to reason about UIs and workflows in a way that simply didn’t work just 6 months ago. That’s why we’re so excited to bring this technology to the forefront of complex legacy SaaS applications that are not yet AI enabled.<p>How does it work?<p>1. Invite agent@frigade.ai to your product. You can send multiple invitations based on distinct roles.<p>2. Our agent automatically explores and reasons about your application.<p>3. Attach any existing help center resources or training documentation to supplement the agent’s understanding. Totally optional.<p>4. Install the agent assistant Javascript snippet (just a few lines).<p>5. That’s it. Your users can now start asking questions and get on demand product tours and questions answered in real time without any overhead.<p>This process takes only a few minutes. Once running, you can improve the agent by rating and providing feedback to the responses it provides. If you want to integrate further, you can also hook up your own code to our tool calling SDK to enable the agent to look up customer info, issue refunds, etc. directly. These calls can be made with just a few lines of code by describing the tool and its parameters in natural language and passing a single Javascript promise (e.g. make an API call, call a function in your app, etc.).<p>Would love to hear what the HN crowd thinks about this approach! Are you building your own AI agent from scratch, or looking to embed one off the shelf?

Show HN: An AI agent that learns your product and guides your users

Hey HN! My name is Christian, and I’m the co-founder of <a href="https://frigade.ai">https://frigade.ai</a>. We’ve built an AI agent that automatically learns how to use any web-based product, and in turn guides users directly in the UI, automatically generates documentation, and even takes actions on a user’s behalf. Think of it as Clippy from the old MS Office. But on steroids. And actually helpful.<p>You can see the agent and tool-calling SDK in action here: <a href="https://www.youtube.com/watch?v=UPe0t3A1Vpg" rel="nofollow">https://www.youtube.com/watch?v=UPe0t3A1Vpg</a><p>How is this different from other AI customer support products?<p>Most AI "copilots" are really just glorified chatbots. They skim your help center and spit out some nonspecific bullet points. Basically some ‘hopes and prayers’ that your users will figure it out. Ultimately, this puts the burden on the user to follow through. And assumes companies are keeping their help center up-to-date with every product change. That means constant screenshots of new product UI or features for accurate instructions.These solutions leverage only a fraction of what’s possible with AI, which can now reason about software interfaces extensively.<p>With Frigade AI, we guide the user directly in the product and build on-demand tours based on the current user’s state and context. The agents can also take actions immediately on a user’s behalf, e.g. inviting a colleague to a workspace or retrieving billing information (via our tool calling SDK).<p>This was only made possible recently. The latest frontier models (GPT 4.1, Claude 4, Gemini 2.5, etc.) are able to reason about UIs and workflows in a way that simply didn’t work just 6 months ago. That’s why we’re so excited to bring this technology to the forefront of complex legacy SaaS applications that are not yet AI enabled.<p>How does it work?<p>1. Invite agent@frigade.ai to your product. You can send multiple invitations based on distinct roles.<p>2. Our agent automatically explores and reasons about your application.<p>3. Attach any existing help center resources or training documentation to supplement the agent’s understanding. Totally optional.<p>4. Install the agent assistant Javascript snippet (just a few lines).<p>5. That’s it. Your users can now start asking questions and get on demand product tours and questions answered in real time without any overhead.<p>This process takes only a few minutes. Once running, you can improve the agent by rating and providing feedback to the responses it provides. If you want to integrate further, you can also hook up your own code to our tool calling SDK to enable the agent to look up customer info, issue refunds, etc. directly. These calls can be made with just a few lines of code by describing the tool and its parameters in natural language and passing a single Javascript promise (e.g. make an API call, call a function in your app, etc.).<p>Would love to hear what the HN crowd thinks about this approach! Are you building your own AI agent from scratch, or looking to embed one off the shelf?

Show HN: An AI agent that learns your product and guides your users

Hey HN! My name is Christian, and I’m the co-founder of <a href="https://frigade.ai">https://frigade.ai</a>. We’ve built an AI agent that automatically learns how to use any web-based product, and in turn guides users directly in the UI, automatically generates documentation, and even takes actions on a user’s behalf. Think of it as Clippy from the old MS Office. But on steroids. And actually helpful.<p>You can see the agent and tool-calling SDK in action here: <a href="https://www.youtube.com/watch?v=UPe0t3A1Vpg" rel="nofollow">https://www.youtube.com/watch?v=UPe0t3A1Vpg</a><p>How is this different from other AI customer support products?<p>Most AI "copilots" are really just glorified chatbots. They skim your help center and spit out some nonspecific bullet points. Basically some ‘hopes and prayers’ that your users will figure it out. Ultimately, this puts the burden on the user to follow through. And assumes companies are keeping their help center up-to-date with every product change. That means constant screenshots of new product UI or features for accurate instructions.These solutions leverage only a fraction of what’s possible with AI, which can now reason about software interfaces extensively.<p>With Frigade AI, we guide the user directly in the product and build on-demand tours based on the current user’s state and context. The agents can also take actions immediately on a user’s behalf, e.g. inviting a colleague to a workspace or retrieving billing information (via our tool calling SDK).<p>This was only made possible recently. The latest frontier models (GPT 4.1, Claude 4, Gemini 2.5, etc.) are able to reason about UIs and workflows in a way that simply didn’t work just 6 months ago. That’s why we’re so excited to bring this technology to the forefront of complex legacy SaaS applications that are not yet AI enabled.<p>How does it work?<p>1. Invite agent@frigade.ai to your product. You can send multiple invitations based on distinct roles.<p>2. Our agent automatically explores and reasons about your application.<p>3. Attach any existing help center resources or training documentation to supplement the agent’s understanding. Totally optional.<p>4. Install the agent assistant Javascript snippet (just a few lines).<p>5. That’s it. Your users can now start asking questions and get on demand product tours and questions answered in real time without any overhead.<p>This process takes only a few minutes. Once running, you can improve the agent by rating and providing feedback to the responses it provides. If you want to integrate further, you can also hook up your own code to our tool calling SDK to enable the agent to look up customer info, issue refunds, etc. directly. These calls can be made with just a few lines of code by describing the tool and its parameters in natural language and passing a single Javascript promise (e.g. make an API call, call a function in your app, etc.).<p>Would love to hear what the HN crowd thinks about this approach! Are you building your own AI agent from scratch, or looking to embed one off the shelf?

Show HN: Dlg – Zero-cost printf-style debugging for Go

Hey HN,<p>I tend to use printf-style debugging as my primary troubleshooting method and only resort to gdb as a last resort.<p>While I like its ease of use printf debugging isn't without its annoyances, namely removing the print statements once you're done.<p>I used to use trace-level logging from proper logging libraries but adding trace calls in every corner quickly gets out of control and results in an overwhelming amount of output.<p>To scratch my own itch I created dlg - a minimal debugging library that disappears completely from production builds. Its API exposes just a single function, Printf [1].<p>dlg is optimized for performance in debug builds and, most importantly, when compiled without the dlg build tag, all calls are eliminated by the Go linker as if dlg was never imported.<p>For debug builds it adds optional stack trace generation configurable via environment variables or linker flags.<p>GitHub: <a href="https://github.com/vvvvv/dlg">https://github.com/vvvvv/dlg</a><p>Any feedback is much appreciated.<p>[1]: Actually two functions - there's also SetOutput.

Show HN: Dlg – Zero-cost printf-style debugging for Go

Hey HN,<p>I tend to use printf-style debugging as my primary troubleshooting method and only resort to gdb as a last resort.<p>While I like its ease of use printf debugging isn't without its annoyances, namely removing the print statements once you're done.<p>I used to use trace-level logging from proper logging libraries but adding trace calls in every corner quickly gets out of control and results in an overwhelming amount of output.<p>To scratch my own itch I created dlg - a minimal debugging library that disappears completely from production builds. Its API exposes just a single function, Printf [1].<p>dlg is optimized for performance in debug builds and, most importantly, when compiled without the dlg build tag, all calls are eliminated by the Go linker as if dlg was never imported.<p>For debug builds it adds optional stack trace generation configurable via environment variables or linker flags.<p>GitHub: <a href="https://github.com/vvvvv/dlg">https://github.com/vvvvv/dlg</a><p>Any feedback is much appreciated.<p>[1]: Actually two functions - there's also SetOutput.

Show HN: Open-source alternative to ChatGPT Agents for browsing

Hey HN,<p>We are Winston, Edward, and James, and we built Meka Agent, an open-source framework that lets vision-based LLMs execute tasks directly on a computer, just like a person would.<p>Backstory:<p>In the last few months, we've been building computer-use agents that have been used by various teams for QA testing, but realized that the underlying browsing frameworks aren't quite good enough yet.<p>As such, we've been working on a browsing agent.<p>We achieved 72.7% on WebArena compared to the previous state of the art set by OpenAI's new ChatGPT agent at 65.4%. You can read more about it here: <a href="https://github.com/trymeka/webarena_evals">https://github.com/trymeka/webarena_evals</a>.<p>Today, we are open sourcing Meka, our state of the art agent, to allow anyone to build their own powerful, vision-based agents from scratch. We provide the groundwork for the hard parts, so you don't have to:<p>* True vision-based control: Meka doesn't just read HTML. It looks at the screen, identifies interactive elements, and decides where to click, type, and scroll.<p>* Full computer access: It's not sandboxed in a browser. Meka operates with OS-level controls, allowing it to handle system dialogues, file uploads, and other interactions that browser-only automation tools can't.<p>* Extensible by design: We've made it easy to plug in your own LLMs and computer providers.<p>* State-of-the-art performance: 72.7% on WebArena<p>Our goal is to enable developers to create repeatable, robust tasks on any computer just by prompting an agent, without worrying about the implementation details.<p>We’d love to get your feedback on how this tool could fit into your automation workflows. Try it out and let us know what you think.<p>You can find the repo on GitHub and get started quickly with our hosted platform, <a href="https://app.withmeka.com/" rel="nofollow">https://app.withmeka.com/</a>.<p>Thanks, Winston, Edward, and James

Show HN: Open-source alternative to ChatGPT Agents for browsing

Hey HN,<p>We are Winston, Edward, and James, and we built Meka Agent, an open-source framework that lets vision-based LLMs execute tasks directly on a computer, just like a person would.<p>Backstory:<p>In the last few months, we've been building computer-use agents that have been used by various teams for QA testing, but realized that the underlying browsing frameworks aren't quite good enough yet.<p>As such, we've been working on a browsing agent.<p>We achieved 72.7% on WebArena compared to the previous state of the art set by OpenAI's new ChatGPT agent at 65.4%. You can read more about it here: <a href="https://github.com/trymeka/webarena_evals">https://github.com/trymeka/webarena_evals</a>.<p>Today, we are open sourcing Meka, our state of the art agent, to allow anyone to build their own powerful, vision-based agents from scratch. We provide the groundwork for the hard parts, so you don't have to:<p>* True vision-based control: Meka doesn't just read HTML. It looks at the screen, identifies interactive elements, and decides where to click, type, and scroll.<p>* Full computer access: It's not sandboxed in a browser. Meka operates with OS-level controls, allowing it to handle system dialogues, file uploads, and other interactions that browser-only automation tools can't.<p>* Extensible by design: We've made it easy to plug in your own LLMs and computer providers.<p>* State-of-the-art performance: 72.7% on WebArena<p>Our goal is to enable developers to create repeatable, robust tasks on any computer just by prompting an agent, without worrying about the implementation details.<p>We’d love to get your feedback on how this tool could fit into your automation workflows. Try it out and let us know what you think.<p>You can find the repo on GitHub and get started quickly with our hosted platform, <a href="https://app.withmeka.com/" rel="nofollow">https://app.withmeka.com/</a>.<p>Thanks, Winston, Edward, and James

Show HN: Online Ruler – Measuring in inches/centimeters

Show HN: ELF Injector

The ELF Injector allows you to "inject" arbitrary-sized relocatable code chunks into ELF executables. The code chunks will run before the original entry point of the executable runs.<p>Included in the project are sample chunks as well as a step-by-step tutorial on how it works.<p>It's a mix of C and assembly and currently runs on 32-bit ARM though it's easy to port to other architectures.

Show HN: 433 – How to make a font that says nothing

Show HN: CUDA Fractal Renderer

Show HN: Color Me Same – A new kind of logic game

Show HN: A GitHub Action that quizzes you on a pull request

A little idea I got from playing with AI SWE Agents. Can AI help make sure we understand the code that our AIs write?<p>PR Quiz uses AI to generate a quiz from a pull request and blocks you from merging until the quiz is passed. You can configure various options like the LLM model to use, max number of attempts to pass the quiz or min diff size to generate a quiz for. I found that the reasoning models, while more expensive, generated better questions from my limited testing.<p>Privacy: This GitHub Action runs a local webserver and uses ngrok to serve the quiz through a temporary url. Your code is only sent to the model provider (OpenAI).

< 1 2 3 ... 21 22 23 24 25 ... 864 865 866 >