The best Hacker News stories from Show from the past day
Latest posts:
Show HN: Detect if an audio file was generated by NotebookLM
Show HN: Detect if an audio file was generated by NotebookLM
Show HN: Brisa Framework – Unifying server and client using the Web Platform
Show HN: Brisa Framework – Unifying server and client using the Web Platform
Show HN: Brisa Framework – Unifying server and client using the Web Platform
Show HN: Open-source real-time talk-to-AI wearable device for few $
1. In the US, about 1/5 children are hospitalized each year that don’t have a caregiver. Caregiver such as play therapists and parents’ stress can also affect children's emotions.<p>2. Not everyone is good at making friends and not everyone has a BFF to talk to. Imagine you're having a hard time in life or at work, and you can't tell your parents or friends.<p>So, we built an open-source project Starmoon and are using affordable hardware components to bring AI characters to real-word objects like toys and plushies to help people emotional growth.<p>We believe this is a complement tool and it is not intended to replace anyone. Please leave any opinion.
Show HN: Open-source real-time talk-to-AI wearable device for few $
1. In the US, about 1/5 children are hospitalized each year that don’t have a caregiver. Caregiver such as play therapists and parents’ stress can also affect children's emotions.<p>2. Not everyone is good at making friends and not everyone has a BFF to talk to. Imagine you're having a hard time in life or at work, and you can't tell your parents or friends.<p>So, we built an open-source project Starmoon and are using affordable hardware components to bring AI characters to real-word objects like toys and plushies to help people emotional growth.<p>We believe this is a complement tool and it is not intended to replace anyone. Please leave any opinion.
Show HN: Open-source real-time talk-to-AI wearable device for few $
1. In the US, about 1/5 children are hospitalized each year that don’t have a caregiver. Caregiver such as play therapists and parents’ stress can also affect children's emotions.<p>2. Not everyone is good at making friends and not everyone has a BFF to talk to. Imagine you're having a hard time in life or at work, and you can't tell your parents or friends.<p>So, we built an open-source project Starmoon and are using affordable hardware components to bring AI characters to real-word objects like toys and plushies to help people emotional growth.<p>We believe this is a complement tool and it is not intended to replace anyone. Please leave any opinion.
Show HN: Screensavers for your terminal (Bevy/Ratatui)
Show HN: Screensavers for your terminal (Bevy/Ratatui)
Show HN: A tool for creating chord charts on the go
Author here - one of the most notable facts about the app is that it's made entirely in Godot Game Engine! I think it's great for apps like this because it makes it especialy easy to iterate on GUI designs such as this one.<p>Feel free to ask me anything about Chord Chart Memo, or my experience with Godot.
Show HN: A tool for creating chord charts on the go
Author here - one of the most notable facts about the app is that it's made entirely in Godot Game Engine! I think it's great for apps like this because it makes it especialy easy to iterate on GUI designs such as this one.<p>Feel free to ask me anything about Chord Chart Memo, or my experience with Godot.
Show HN: FFmpeg-over-IP – Connect to remote FFmpeg servers
Dear HN,<p>I’m excited to show case a personal project. It has helped me quite a bit with my home lab, I hope it can help you with yours too! ffmpeg-over-ip has two components, a server and a client. You can run the server in an environment with access to a GPU and a locally installed version of ffmpeg, the client only needs network access to the server and no GPU or ffmpeg locally.<p>Both and client and the server need a shared filesystem for this to work (so the server can write output to it, and client can read from it). In my usecase, smb works well if your (GPU) server is a windows machine, nfs works really well for linux setups.<p>This utility can be useful in a number of scenarios:<p>- You find passing through a (v)GPU to your virtual machines complicated<p>- You want to use the same GPU for ffmpeg in multiple virtual machines<p>- Your server has a weak GPU so you want to use the GPU from your gaming machine<p>- Your GPU drivers in one OS are not as good as another (AMD RX6400 never worked for me in linux, but did so in windows)<p>I’ve posted some instructions in the Github package README, please let me know if they are unclear in any way and I’ll try to help!<p>Here's the link: <a href="https://github.com/steelbrain/ffmpeg-over-ip">https://github.com/steelbrain/ffmpeg-over-ip</a>
Show HN: FFmpeg-over-IP – Connect to remote FFmpeg servers
Dear HN,<p>I’m excited to show case a personal project. It has helped me quite a bit with my home lab, I hope it can help you with yours too! ffmpeg-over-ip has two components, a server and a client. You can run the server in an environment with access to a GPU and a locally installed version of ffmpeg, the client only needs network access to the server and no GPU or ffmpeg locally.<p>Both and client and the server need a shared filesystem for this to work (so the server can write output to it, and client can read from it). In my usecase, smb works well if your (GPU) server is a windows machine, nfs works really well for linux setups.<p>This utility can be useful in a number of scenarios:<p>- You find passing through a (v)GPU to your virtual machines complicated<p>- You want to use the same GPU for ffmpeg in multiple virtual machines<p>- Your server has a weak GPU so you want to use the GPU from your gaming machine<p>- Your GPU drivers in one OS are not as good as another (AMD RX6400 never worked for me in linux, but did so in windows)<p>I’ve posted some instructions in the Github package README, please let me know if they are unclear in any way and I’ll try to help!<p>Here's the link: <a href="https://github.com/steelbrain/ffmpeg-over-ip">https://github.com/steelbrain/ffmpeg-over-ip</a>
Show HN: kew – A Terminal Music Player for Linux
Hi HN,<p>I created kew, a music player for the Linux terminal.<p>This started when I asked myself: what if I could just type something like "play nirvana" in the terminal and have the rest taken care of automatically? That got the ball rolling and I kept adding stuff: covers in ascii and then as sixel images, a playlist view, a visualizer, a library view and finally search.<p>While kew can be used as a commandline tool, it has evolved into a TUI app.<p>Here are some example commands:<p>kew nirvana # Plays all of your Nirvana songs, shuffled<p>kew nevermind # Plays the "Nevermind" album in order<p>kew spirit # Plays "Smells Like Teen Spirit"<p>kew all # Plays all your music, shuffled<p>kew albums # Plays one album after the other in random order<p>It works best when your music library is organized like this:
Artist/Album(s)/Track(s)<p>kew is written in C and licensed under GPLv2.<p>Source and screenshot: <a href="https://github.com/ravachol/kew">https://github.com/ravachol/kew</a>
Show HN: kew – A Terminal Music Player for Linux
Hi HN,<p>I created kew, a music player for the Linux terminal.<p>This started when I asked myself: what if I could just type something like "play nirvana" in the terminal and have the rest taken care of automatically? That got the ball rolling and I kept adding stuff: covers in ascii and then as sixel images, a playlist view, a visualizer, a library view and finally search.<p>While kew can be used as a commandline tool, it has evolved into a TUI app.<p>Here are some example commands:<p>kew nirvana # Plays all of your Nirvana songs, shuffled<p>kew nevermind # Plays the "Nevermind" album in order<p>kew spirit # Plays "Smells Like Teen Spirit"<p>kew all # Plays all your music, shuffled<p>kew albums # Plays one album after the other in random order<p>It works best when your music library is organized like this:
Artist/Album(s)/Track(s)<p>kew is written in C and licensed under GPLv2.<p>Source and screenshot: <a href="https://github.com/ravachol/kew">https://github.com/ravachol/kew</a>
Show HN: kew – A Terminal Music Player for Linux
Hi HN,<p>I created kew, a music player for the Linux terminal.<p>This started when I asked myself: what if I could just type something like "play nirvana" in the terminal and have the rest taken care of automatically? That got the ball rolling and I kept adding stuff: covers in ascii and then as sixel images, a playlist view, a visualizer, a library view and finally search.<p>While kew can be used as a commandline tool, it has evolved into a TUI app.<p>Here are some example commands:<p>kew nirvana # Plays all of your Nirvana songs, shuffled<p>kew nevermind # Plays the "Nevermind" album in order<p>kew spirit # Plays "Smells Like Teen Spirit"<p>kew all # Plays all your music, shuffled<p>kew albums # Plays one album after the other in random order<p>It works best when your music library is organized like this:
Artist/Album(s)/Track(s)<p>kew is written in C and licensed under GPLv2.<p>Source and screenshot: <a href="https://github.com/ravachol/kew">https://github.com/ravachol/kew</a>
Show HN: Open source framework OpenAI uses for Advanced Voice
Hey HN, we've been working with OpenAI for the past few months on the new Realtime API.<p>The goal is to give everyone access to the same stack that underpins Advanced Voice in the ChatGPT app.<p>Under the hood it works like this:
- A user's speech is captured by a LiveKit client SDK in the ChatGPT app
- Their speech is streamed using WebRTC to OpenAI’s voice agent
- The agent relays the speech prompt over websocket to GPT-4o
- GPT-4o runs inference and streams speech packets (over websocket) back to the agent
- The agent relays generated speech using WebRTC back to the user’s device<p>The Realtime API that OpenAI launched is the websocket interface to GPT-4o. This backend framework covers the voice agent portion. Besides having additional logic like function calling, the agent fundamentally proxies WebRTC to websocket.<p>The reason for this is because websocket isn’t the best choice for client-server communication. The vast majority of packet loss occurs between a server and client device and websocket doesn’t provide programmatic control or intervention in lossy network environments like WiFi or cellular. Packet loss leads to higher latency and choppy or garbled audio.
Show HN: Open source framework OpenAI uses for Advanced Voice
Hey HN, we've been working with OpenAI for the past few months on the new Realtime API.<p>The goal is to give everyone access to the same stack that underpins Advanced Voice in the ChatGPT app.<p>Under the hood it works like this:
- A user's speech is captured by a LiveKit client SDK in the ChatGPT app
- Their speech is streamed using WebRTC to OpenAI’s voice agent
- The agent relays the speech prompt over websocket to GPT-4o
- GPT-4o runs inference and streams speech packets (over websocket) back to the agent
- The agent relays generated speech using WebRTC back to the user’s device<p>The Realtime API that OpenAI launched is the websocket interface to GPT-4o. This backend framework covers the voice agent portion. Besides having additional logic like function calling, the agent fundamentally proxies WebRTC to websocket.<p>The reason for this is because websocket isn’t the best choice for client-server communication. The vast majority of packet loss occurs between a server and client device and websocket doesn’t provide programmatic control or intervention in lossy network environments like WiFi or cellular. Packet loss leads to higher latency and choppy or garbled audio.
Show HN: Open source framework OpenAI uses for Advanced Voice
Hey HN, we've been working with OpenAI for the past few months on the new Realtime API.<p>The goal is to give everyone access to the same stack that underpins Advanced Voice in the ChatGPT app.<p>Under the hood it works like this:
- A user's speech is captured by a LiveKit client SDK in the ChatGPT app
- Their speech is streamed using WebRTC to OpenAI’s voice agent
- The agent relays the speech prompt over websocket to GPT-4o
- GPT-4o runs inference and streams speech packets (over websocket) back to the agent
- The agent relays generated speech using WebRTC back to the user’s device<p>The Realtime API that OpenAI launched is the websocket interface to GPT-4o. This backend framework covers the voice agent portion. Besides having additional logic like function calling, the agent fundamentally proxies WebRTC to websocket.<p>The reason for this is because websocket isn’t the best choice for client-server communication. The vast majority of packet loss occurs between a server and client device and websocket doesn’t provide programmatic control or intervention in lossy network environments like WiFi or cellular. Packet loss leads to higher latency and choppy or garbled audio.