The best Hacker News stories from Show from the past day

Go back

Latest posts:

Show HN: Open-source enterprise SSO – integrate SAML with a few lines of code

Show HN: Open-source enterprise SSO – integrate SAML with a few lines of code

Show HN: Stand out with your Resume. Use Resoume

My Indie hacking journey continues~<p>After a bit over 2 years as a solo dev, I am proud of how far I have come with Resoume. Still all alone :D<p>Resoume is now an AI-powered resume optimization tool to help you analyze your CV. It then provides you with suggestions on how to improve your CV. On top, I also got a great set of templates with a lot of customization to make every CV unique.<p>Resoume covers your job-seeking journey from design to content. And helps your CV to be perfectly tailored toward the job you want to apply for!<p>That is it from me. Would love to hear your thoughts Enjoy your day! Gobie

Show HN: Stand out with your Resume. Use Resoume

My Indie hacking journey continues~<p>After a bit over 2 years as a solo dev, I am proud of how far I have come with Resoume. Still all alone :D<p>Resoume is now an AI-powered resume optimization tool to help you analyze your CV. It then provides you with suggestions on how to improve your CV. On top, I also got a great set of templates with a lot of customization to make every CV unique.<p>Resoume covers your job-seeking journey from design to content. And helps your CV to be perfectly tailored toward the job you want to apply for!<p>That is it from me. Would love to hear your thoughts Enjoy your day! Gobie

Show HN: Contribution Graph as a Git Command

Show HN: Contribution Graph as a Git Command

Show HN: Contribution Graph as a Git Command

Show HN: A Reddit style site to discuss podcast episodes

Show HN: A Reddit style site to discuss podcast episodes

Show HN: A Reddit style site to discuss podcast episodes

Show HN: A Reddit style site to discuss podcast episodes

Show HN: Metagration – Migrate Postgres with Postgres

Show HN: Never forget what you've learned

Save All uses quizzes, notifications, and emails to stop you forgetting what you've learned. Unlike other spaced repetition apps, we've (really) prioritised making Save All as simple as possible - minimalist design, no distractions, and minimal cognitive load placed on users.<p>We achieve this simplicity by using machine learning (e.g. large language models like BERT, GPT) to reduce the number of decisions users have to make. You don't have to decide whether you remembered a card, we know whether you did. You don't have to decide how to be quizzed on your information, we'll decide for you.<p>We're VC funded & growing fast but would love to hear HN's critical feedback. Tell it to us how it is!

Show HN: Run Pi-hole on a local Kubernetes/K3s cluster

Show HN: I turned my face rec system into a video codec

Before the pandemic, my tiny startup was doing quite well selling Edge AI systems, based on our own lightweight AI inference engine, with object detection and face recognition for smart city and smart retail & food service applications.<p>When the real world shut down, there was suddenly nothing to monitor on streets and in restaurants, so I set out to try and evolve our real time face recognition system into a video codec for high quality face-to-face online interactions, as I was not satisfied with the quality of Zoom and friends. I got it to work, and the first release for IOS was just approved on Apple's app store, link: <a href="https://apps.apple.com/app/vertigo-focus/id1540073203" rel="nofollow">https://apps.apple.com/app/vertigo-focus/id1540073203</a><p>The way it works is that you create a meeting URL, which you can share out-of-band, for instance via slack or text message. You can also share as a QR code which the app can scan to join a call. You then place your device on a surface in front of you so that the front camera can see you, and it will recognize you face and assign you to your own session, which is broadcast to the meeting channel. If more than one person is in view, both of you will be broadcast but with separate session ids, like if you were on separate cameras. Other meeting participants will show up on your screen and you can start talking. It is optimized for eye contact, meaning that the eyes will actually make it through to the other side as more than just dark pixel clouds, so thinks should feel a bit more personal than the standard Zoom/Teams/or Google Meet call.<p>Because it uses face rec, you can ONLY show your face, and if you disappear from view your audio will stop after a while, to avoid situations like when you need to go the the restroom but forget to mute. This also solves dick-pics etc.<p>The CODEC is not based on H26[45], but is pure AI that runs on the GPU. There is a neural network that compresses the video in real time, and another one decompressing on the receiving end. Finding a tight network architecture that would do this in real time with acceptable quality was a major part of the effort. There are several quality settings possible, but right now it is set fairly high and for 20FPS maxes out around 700kbit/s, though typically uses about half. I've demonstrated good results down to around 200kbit/s, so in theory it should work over satellite links or even Bluetooth. The protocol is UDP with no congestion control but with (Wirehair) FEC to protect against mild packet loss, future versions will detect packet loss and adapt to available bandwidth.<p>The audio just uses OPUS and may click a little bit, I blame AudioEngine or the fact that the last time I wrote audio code was for the game I published for the Amiga in 1994.<p>If you don't have a friend around or multiple devices to play with, there is an "echo test" server mode that allows you to be in a meeting with yourself. Traffic will be peer-to-peer if possible, but otherwise you will be relaying through my tiny Raspberry PI server, so YMMV. I plan to try to switch to something like fly.io soon to improve scalability.<p>There is also a MacOS version coming very soon, and the underlying AI engine also runs on Windows & Linux. Android support is planned.<p>Please take a look and let me know what you think.

Show HN: I turned my face rec system into a video codec

Before the pandemic, my tiny startup was doing quite well selling Edge AI systems, based on our own lightweight AI inference engine, with object detection and face recognition for smart city and smart retail & food service applications.<p>When the real world shut down, there was suddenly nothing to monitor on streets and in restaurants, so I set out to try and evolve our real time face recognition system into a video codec for high quality face-to-face online interactions, as I was not satisfied with the quality of Zoom and friends. I got it to work, and the first release for IOS was just approved on Apple's app store, link: <a href="https://apps.apple.com/app/vertigo-focus/id1540073203" rel="nofollow">https://apps.apple.com/app/vertigo-focus/id1540073203</a><p>The way it works is that you create a meeting URL, which you can share out-of-band, for instance via slack or text message. You can also share as a QR code which the app can scan to join a call. You then place your device on a surface in front of you so that the front camera can see you, and it will recognize you face and assign you to your own session, which is broadcast to the meeting channel. If more than one person is in view, both of you will be broadcast but with separate session ids, like if you were on separate cameras. Other meeting participants will show up on your screen and you can start talking. It is optimized for eye contact, meaning that the eyes will actually make it through to the other side as more than just dark pixel clouds, so thinks should feel a bit more personal than the standard Zoom/Teams/or Google Meet call.<p>Because it uses face rec, you can ONLY show your face, and if you disappear from view your audio will stop after a while, to avoid situations like when you need to go the the restroom but forget to mute. This also solves dick-pics etc.<p>The CODEC is not based on H26[45], but is pure AI that runs on the GPU. There is a neural network that compresses the video in real time, and another one decompressing on the receiving end. Finding a tight network architecture that would do this in real time with acceptable quality was a major part of the effort. There are several quality settings possible, but right now it is set fairly high and for 20FPS maxes out around 700kbit/s, though typically uses about half. I've demonstrated good results down to around 200kbit/s, so in theory it should work over satellite links or even Bluetooth. The protocol is UDP with no congestion control but with (Wirehair) FEC to protect against mild packet loss, future versions will detect packet loss and adapt to available bandwidth.<p>The audio just uses OPUS and may click a little bit, I blame AudioEngine or the fact that the last time I wrote audio code was for the game I published for the Amiga in 1994.<p>If you don't have a friend around or multiple devices to play with, there is an "echo test" server mode that allows you to be in a meeting with yourself. Traffic will be peer-to-peer if possible, but otherwise you will be relaying through my tiny Raspberry PI server, so YMMV. I plan to try to switch to something like fly.io soon to improve scalability.<p>There is also a MacOS version coming very soon, and the underlying AI engine also runs on Windows & Linux. Android support is planned.<p>Please take a look and let me know what you think.

Show HN: I turned my face rec system into a video codec

Before the pandemic, my tiny startup was doing quite well selling Edge AI systems, based on our own lightweight AI inference engine, with object detection and face recognition for smart city and smart retail & food service applications.<p>When the real world shut down, there was suddenly nothing to monitor on streets and in restaurants, so I set out to try and evolve our real time face recognition system into a video codec for high quality face-to-face online interactions, as I was not satisfied with the quality of Zoom and friends. I got it to work, and the first release for IOS was just approved on Apple's app store, link: <a href="https://apps.apple.com/app/vertigo-focus/id1540073203" rel="nofollow">https://apps.apple.com/app/vertigo-focus/id1540073203</a><p>The way it works is that you create a meeting URL, which you can share out-of-band, for instance via slack or text message. You can also share as a QR code which the app can scan to join a call. You then place your device on a surface in front of you so that the front camera can see you, and it will recognize you face and assign you to your own session, which is broadcast to the meeting channel. If more than one person is in view, both of you will be broadcast but with separate session ids, like if you were on separate cameras. Other meeting participants will show up on your screen and you can start talking. It is optimized for eye contact, meaning that the eyes will actually make it through to the other side as more than just dark pixel clouds, so thinks should feel a bit more personal than the standard Zoom/Teams/or Google Meet call.<p>Because it uses face rec, you can ONLY show your face, and if you disappear from view your audio will stop after a while, to avoid situations like when you need to go the the restroom but forget to mute. This also solves dick-pics etc.<p>The CODEC is not based on H26[45], but is pure AI that runs on the GPU. There is a neural network that compresses the video in real time, and another one decompressing on the receiving end. Finding a tight network architecture that would do this in real time with acceptable quality was a major part of the effort. There are several quality settings possible, but right now it is set fairly high and for 20FPS maxes out around 700kbit/s, though typically uses about half. I've demonstrated good results down to around 200kbit/s, so in theory it should work over satellite links or even Bluetooth. The protocol is UDP with no congestion control but with (Wirehair) FEC to protect against mild packet loss, future versions will detect packet loss and adapt to available bandwidth.<p>The audio just uses OPUS and may click a little bit, I blame AudioEngine or the fact that the last time I wrote audio code was for the game I published for the Amiga in 1994.<p>If you don't have a friend around or multiple devices to play with, there is an "echo test" server mode that allows you to be in a meeting with yourself. Traffic will be peer-to-peer if possible, but otherwise you will be relaying through my tiny Raspberry PI server, so YMMV. I plan to try to switch to something like fly.io soon to improve scalability.<p>There is also a MacOS version coming very soon, and the underlying AI engine also runs on Windows & Linux. Android support is planned.<p>Please take a look and let me know what you think.

Show HN: I turned my face rec system into a video codec

Before the pandemic, my tiny startup was doing quite well selling Edge AI systems, based on our own lightweight AI inference engine, with object detection and face recognition for smart city and smart retail & food service applications.<p>When the real world shut down, there was suddenly nothing to monitor on streets and in restaurants, so I set out to try and evolve our real time face recognition system into a video codec for high quality face-to-face online interactions, as I was not satisfied with the quality of Zoom and friends. I got it to work, and the first release for IOS was just approved on Apple's app store, link: <a href="https://apps.apple.com/app/vertigo-focus/id1540073203" rel="nofollow">https://apps.apple.com/app/vertigo-focus/id1540073203</a><p>The way it works is that you create a meeting URL, which you can share out-of-band, for instance via slack or text message. You can also share as a QR code which the app can scan to join a call. You then place your device on a surface in front of you so that the front camera can see you, and it will recognize you face and assign you to your own session, which is broadcast to the meeting channel. If more than one person is in view, both of you will be broadcast but with separate session ids, like if you were on separate cameras. Other meeting participants will show up on your screen and you can start talking. It is optimized for eye contact, meaning that the eyes will actually make it through to the other side as more than just dark pixel clouds, so thinks should feel a bit more personal than the standard Zoom/Teams/or Google Meet call.<p>Because it uses face rec, you can ONLY show your face, and if you disappear from view your audio will stop after a while, to avoid situations like when you need to go the the restroom but forget to mute. This also solves dick-pics etc.<p>The CODEC is not based on H26[45], but is pure AI that runs on the GPU. There is a neural network that compresses the video in real time, and another one decompressing on the receiving end. Finding a tight network architecture that would do this in real time with acceptable quality was a major part of the effort. There are several quality settings possible, but right now it is set fairly high and for 20FPS maxes out around 700kbit/s, though typically uses about half. I've demonstrated good results down to around 200kbit/s, so in theory it should work over satellite links or even Bluetooth. The protocol is UDP with no congestion control but with (Wirehair) FEC to protect against mild packet loss, future versions will detect packet loss and adapt to available bandwidth.<p>The audio just uses OPUS and may click a little bit, I blame AudioEngine or the fact that the last time I wrote audio code was for the game I published for the Amiga in 1994.<p>If you don't have a friend around or multiple devices to play with, there is an "echo test" server mode that allows you to be in a meeting with yourself. Traffic will be peer-to-peer if possible, but otherwise you will be relaying through my tiny Raspberry PI server, so YMMV. I plan to try to switch to something like fly.io soon to improve scalability.<p>There is also a MacOS version coming very soon, and the underlying AI engine also runs on Windows & Linux. Android support is planned.<p>Please take a look and let me know what you think.

Show HN: I brought the benefits of LaTeX to non technical users

Show HN: Dress Circle – IMDB for UK Theatre

< 1 2 3 ... 634 635 636 637 638 ... 902 903 904 >