The best Hacker News stories from Show from the past day
Latest posts:
Show HN: Archivy – Extensible Self Hosted Knowledge Base – v1 Release
Show HN: Archivy – Extensible Self Hosted Knowledge Base – v1 Release
Launch HN: Zaraz (YC W20) – Use third-party tools without slowing your website
Hi HN! We're Yair and Yo'av of Zaraz (<a href="https://zaraz.com" rel="nofollow">https://zaraz.com</a>). We make websites faster by loading their third-party stack in an optimized way. By “third-party” we mean utilities or additional products you add to your website (eg. analytics), not things you build your website with (eg. React).<p>Before we started this we worked on opposite sides of this battle for third-party inclusion: Yair was working for the folks asking to implement just-one-more analytics tool, while Yo'av was a developer trying (and often failing) to push back. Avoiding bloat to begin with would be preferable, but anyone working for even a medium-sized company knows how hard that is - usually when a higher up agrees to try or add a new tool, resistance is futile. Hence the question becomes, can you do it without harming your performance?<p>The average US top 5,000 website loads 22 different third-party tools - analytics, customer success, marketing and whatnot. We wrote a bot that scanned these websites and discovered that third-parties account for 40% of their “Time to Interactive”, and other metrics like TBT, FCP, FID and CLS were hurt in a similar way. From the user perspective, the page usually behaves exactly the same without these tools (...except 40% faster).<p>These new metrics are becoming more popular for two reasons. Firstly, users actually feel them - unlike events such as "DOMContentLoad" & "Load" that can be triggered long before the user can actually do anything, these metrics provide a much better proxy to the real user experience. Secondly, with Google soon penalizing slow websites, they're becoming more and more important for SEO. We see the growing popularity of these metrics as a good thing. We want a faster web.<p>Nowadays, the most common way to integrate a third-party into your website is either to just paste its `<script>` snippet somewhere in your code, or use some "Tag Management" software (awful name!) like Google Tag Manager, Tealium, or even a tool like Segment. All these options pretty much come with the same cost - everything loads by default together with your page, and users just have to wait and wait. If all this slowness doesn't feel so bad on your devices now, remember that much of the world accesses the internet through devices that are probably a lot slower than yours.<p>We built Zaraz to be a performance-first third-party manager. Each tool is different, but the concept is to run whatever we can on our backend instead of in the browser, leaving it to focus on loading your website. While other solutions serve all your visitors with the same script and then evaluate it in the browser (should we run this conversion pixel? Should we load this analytics tool?) - we do this on our backend. But the real magic is that we created an environment living inside a Cloudflare Worker, that executes the actual third-party scripts instead of having them run in the browser. Google Analytics, Reddit conversion pixel, LinkedIn Insight, you name it - we’re turning all those things into miniature server-side applications that your visitors’ browsers need not to worry about. If a certain tool still needs to fire a request from the browser (eg. it needs to set a cookie), only the resulting URL from evaluating its script will be sent back to the user browser. It’s a server side environment executing third-party code, that you have 100% control over. When we measure the speed of a website optimized with Zaraz, third-parties have close to zero effect on it, because the browser almost does nothing.<p>We are already serving a few customers in production, and we’re seeing huge improvements in speed (and revenues!) with all of them. Zaraz is probably the easiest way you can make your website faster, today (try our analyzer to see how we can improve your website: <a href="https://zaraz.com/analyze" rel="nofollow">https://zaraz.com/analyze</a>). Aside from performance, since we have total control on what data is revealed to our isolated environment, we are using it to help companies protect the privacy of their visitors by masking PIIs, hiding IP addresses, disabling fingerprinting etc. We designed our infrastructure as a set of serverless, storageless and stateless functions - to make sure your visitors data is never saved, not even by accident.<p>We are currently onboarding mostly enterprises and high-traffic websites, but we plan to introduce a free tier after we are done creating a self-onboarding flow. We are on a mission to make the web faster and we want all websites to benefit from it!<p>We would be thrilled to hear what you think, and if you have more ideas on how to make websites faster please do share them with us. Thank you!
Launch HN: Zaraz (YC W20) – Use third-party tools without slowing your website
Hi HN! We're Yair and Yo'av of Zaraz (<a href="https://zaraz.com" rel="nofollow">https://zaraz.com</a>). We make websites faster by loading their third-party stack in an optimized way. By “third-party” we mean utilities or additional products you add to your website (eg. analytics), not things you build your website with (eg. React).<p>Before we started this we worked on opposite sides of this battle for third-party inclusion: Yair was working for the folks asking to implement just-one-more analytics tool, while Yo'av was a developer trying (and often failing) to push back. Avoiding bloat to begin with would be preferable, but anyone working for even a medium-sized company knows how hard that is - usually when a higher up agrees to try or add a new tool, resistance is futile. Hence the question becomes, can you do it without harming your performance?<p>The average US top 5,000 website loads 22 different third-party tools - analytics, customer success, marketing and whatnot. We wrote a bot that scanned these websites and discovered that third-parties account for 40% of their “Time to Interactive”, and other metrics like TBT, FCP, FID and CLS were hurt in a similar way. From the user perspective, the page usually behaves exactly the same without these tools (...except 40% faster).<p>These new metrics are becoming more popular for two reasons. Firstly, users actually feel them - unlike events such as "DOMContentLoad" & "Load" that can be triggered long before the user can actually do anything, these metrics provide a much better proxy to the real user experience. Secondly, with Google soon penalizing slow websites, they're becoming more and more important for SEO. We see the growing popularity of these metrics as a good thing. We want a faster web.<p>Nowadays, the most common way to integrate a third-party into your website is either to just paste its `<script>` snippet somewhere in your code, or use some "Tag Management" software (awful name!) like Google Tag Manager, Tealium, or even a tool like Segment. All these options pretty much come with the same cost - everything loads by default together with your page, and users just have to wait and wait. If all this slowness doesn't feel so bad on your devices now, remember that much of the world accesses the internet through devices that are probably a lot slower than yours.<p>We built Zaraz to be a performance-first third-party manager. Each tool is different, but the concept is to run whatever we can on our backend instead of in the browser, leaving it to focus on loading your website. While other solutions serve all your visitors with the same script and then evaluate it in the browser (should we run this conversion pixel? Should we load this analytics tool?) - we do this on our backend. But the real magic is that we created an environment living inside a Cloudflare Worker, that executes the actual third-party scripts instead of having them run in the browser. Google Analytics, Reddit conversion pixel, LinkedIn Insight, you name it - we’re turning all those things into miniature server-side applications that your visitors’ browsers need not to worry about. If a certain tool still needs to fire a request from the browser (eg. it needs to set a cookie), only the resulting URL from evaluating its script will be sent back to the user browser. It’s a server side environment executing third-party code, that you have 100% control over. When we measure the speed of a website optimized with Zaraz, third-parties have close to zero effect on it, because the browser almost does nothing.<p>We are already serving a few customers in production, and we’re seeing huge improvements in speed (and revenues!) with all of them. Zaraz is probably the easiest way you can make your website faster, today (try our analyzer to see how we can improve your website: <a href="https://zaraz.com/analyze" rel="nofollow">https://zaraz.com/analyze</a>). Aside from performance, since we have total control on what data is revealed to our isolated environment, we are using it to help companies protect the privacy of their visitors by masking PIIs, hiding IP addresses, disabling fingerprinting etc. We designed our infrastructure as a set of serverless, storageless and stateless functions - to make sure your visitors data is never saved, not even by accident.<p>We are currently onboarding mostly enterprises and high-traffic websites, but we plan to introduce a free tier after we are done creating a self-onboarding flow. We are on a mission to make the web faster and we want all websites to benefit from it!<p>We would be thrilled to hear what you think, and if you have more ideas on how to make websites faster please do share them with us. Thank you!
Launch HN: Zaraz (YC W20) – Use third-party tools without slowing your website
Hi HN! We're Yair and Yo'av of Zaraz (<a href="https://zaraz.com" rel="nofollow">https://zaraz.com</a>). We make websites faster by loading their third-party stack in an optimized way. By “third-party” we mean utilities or additional products you add to your website (eg. analytics), not things you build your website with (eg. React).<p>Before we started this we worked on opposite sides of this battle for third-party inclusion: Yair was working for the folks asking to implement just-one-more analytics tool, while Yo'av was a developer trying (and often failing) to push back. Avoiding bloat to begin with would be preferable, but anyone working for even a medium-sized company knows how hard that is - usually when a higher up agrees to try or add a new tool, resistance is futile. Hence the question becomes, can you do it without harming your performance?<p>The average US top 5,000 website loads 22 different third-party tools - analytics, customer success, marketing and whatnot. We wrote a bot that scanned these websites and discovered that third-parties account for 40% of their “Time to Interactive”, and other metrics like TBT, FCP, FID and CLS were hurt in a similar way. From the user perspective, the page usually behaves exactly the same without these tools (...except 40% faster).<p>These new metrics are becoming more popular for two reasons. Firstly, users actually feel them - unlike events such as "DOMContentLoad" & "Load" that can be triggered long before the user can actually do anything, these metrics provide a much better proxy to the real user experience. Secondly, with Google soon penalizing slow websites, they're becoming more and more important for SEO. We see the growing popularity of these metrics as a good thing. We want a faster web.<p>Nowadays, the most common way to integrate a third-party into your website is either to just paste its `<script>` snippet somewhere in your code, or use some "Tag Management" software (awful name!) like Google Tag Manager, Tealium, or even a tool like Segment. All these options pretty much come with the same cost - everything loads by default together with your page, and users just have to wait and wait. If all this slowness doesn't feel so bad on your devices now, remember that much of the world accesses the internet through devices that are probably a lot slower than yours.<p>We built Zaraz to be a performance-first third-party manager. Each tool is different, but the concept is to run whatever we can on our backend instead of in the browser, leaving it to focus on loading your website. While other solutions serve all your visitors with the same script and then evaluate it in the browser (should we run this conversion pixel? Should we load this analytics tool?) - we do this on our backend. But the real magic is that we created an environment living inside a Cloudflare Worker, that executes the actual third-party scripts instead of having them run in the browser. Google Analytics, Reddit conversion pixel, LinkedIn Insight, you name it - we’re turning all those things into miniature server-side applications that your visitors’ browsers need not to worry about. If a certain tool still needs to fire a request from the browser (eg. it needs to set a cookie), only the resulting URL from evaluating its script will be sent back to the user browser. It’s a server side environment executing third-party code, that you have 100% control over. When we measure the speed of a website optimized with Zaraz, third-parties have close to zero effect on it, because the browser almost does nothing.<p>We are already serving a few customers in production, and we’re seeing huge improvements in speed (and revenues!) with all of them. Zaraz is probably the easiest way you can make your website faster, today (try our analyzer to see how we can improve your website: <a href="https://zaraz.com/analyze" rel="nofollow">https://zaraz.com/analyze</a>). Aside from performance, since we have total control on what data is revealed to our isolated environment, we are using it to help companies protect the privacy of their visitors by masking PIIs, hiding IP addresses, disabling fingerprinting etc. We designed our infrastructure as a set of serverless, storageless and stateless functions - to make sure your visitors data is never saved, not even by accident.<p>We are currently onboarding mostly enterprises and high-traffic websites, but we plan to introduce a free tier after we are done creating a self-onboarding flow. We are on a mission to make the web faster and we want all websites to benefit from it!<p>We would be thrilled to hear what you think, and if you have more ideas on how to make websites faster please do share them with us. Thank you!
Show HN: A retrainable subtitle synchronizer you can now build your own
Show HN: A retrainable subtitle synchronizer you can now build your own
Show HN: Deploy ML Models on a Budget
Show HN: Deploy ML Models on a Budget
Launch HN: Albedo (YC W21) – Highest resolution satellite imagery
Hey HN! I’m Topher, here with Winston and AJ, and we’re the co-founders of Albedo (<a href="https://albedo.space" rel="nofollow">https://albedo.space</a>). We’re building satellites that will capture both visible and thermal imagery - at a resolution 9x higher than what is available today (see comparison: <a href="https://photos.app.goo.gl/gwokp4WT8JPvyue98" rel="nofollow">https://photos.app.goo.gl/gwokp4WT8JPvyue98</a>).<p>My technical background is primarily in optics/imaging science related to remote sensing. I previously worked for Lockheed Martin, where I met AJ, who is an expert in satellite architecture and systems engineering. We’ve spent most of our career working on classified space systems, and while the missions we were involved with are super cool, that world is slower to adopt the latest new space technologies. We started Albedo in order to create a new type of satellite architecture that captures high resolution imagery at a fraction of the cost historically required. Winston was previously a software engineer at Facebook, where he frequently used satellite imagery and realized the huge potential of higher resolution datasets.<p>While the use cases for satellite imagery are endless, adoption has been underwhelming - even for obvious and larger applications like agriculture, insurance, energy, and mapping. The main limitations that have prevented widespread use are high cost, inaccessibility, and low resolution.<p>Today, buying commercial satellite imagery involves a back-and-forth with a salesperson in a sometimes months-long process, with high prices that exclude all but the biggest companies. This process needs to be simplified with transparent, commodity pricing and an automated process, where all you need to buy imagery is a credit card. On the accessibility front, it’s surprising how few providers have nailed down a streamlined, fully cloud-based delivery mechanism. While working at Facebook, Winston sometimes dealt with imagery delivered through FTP servers or physical hard drives. Another thing users are looking for is more transparency when tasking a new satellite image, such as an immediate assessment of when it will be collected. These are all problems we are working on solving at Albedo.<p>On the space side, we’re able to achieve the substantial cost savings by taking advantage of emerging space technologies, two of which are electric propulsion and on-orbit refueling. Our satellites will fly super close to the earth, essentially in the atmosphere, enabling 10cm resolution without having to build a school bus sized satellite.<p>Electric propulsion makes the fuel on our satellites way more efficient, at the expense of low thrust. Think about it like your car gasoline going from 30 to 300 mpg, but you could only drive 5 mph. Our propulsion only needs to maintain a steady offset to the atmospheric drag, so low thrust and high efficiency is perfect. By the time our first few satellites run out of fuel, on-orbit refueling will be a reality, and we can just refill our tanks. We’re still in the architecture and design phase, but we expect to have our first few satellites flying in 2024 and the full constellation up in 2027.<p>The current climate crisis requires a diverse set of sensors in space to support emissions monitoring, ESG initiatives/investments, and infrastructure sustainability. Thermal sensors are a key component for this, and very few are currently in orbit. Since our satellites are larger than normal, they are uniquely suited to capture the long wavelengths of thermal energy at a resolution of 2 meters. We’ll also be taking advantage of advances in microbolometer technology, to eliminate the crazy cooling requirements that have made thermal satellites so expensive in the past. The current state-of-the-art for thermal resolution is 70 meters, which is only marginally useful for most applications.<p>We’re aiming to adopt the stance of being a pure data provider (i.e. not doing analytics). We think the best way to facilitate overall market growth is to do one thing incredibly well: sell imagery better, cheaper, and faster than what users have available today. While this allows us to be vertical agnostic, some of our more well-suited applications include: crop health monitoring, pipeline inspection, property insurance underwriting/weather damage evaluation, and wildfire/vegetation management around power lines. By making high-res imagery a commodity, we are also betting on it unlocking new applications in a similar fashion to GPS (e.g. Tinder, Pokemon Go, and Uber).<p>One last thing - new remote sensing regulations were released by NOAA last May, removing the previous limit on resolution. So between the technology side and regulatory side, the timing is kind of perfect for us.<p>All thoughts and questions are appreciated - and we’d love to hear if you know of any companies that could benefit from our imagery. Thanks for reading!
Launch HN: Albedo (YC W21) – Highest resolution satellite imagery
Hey HN! I’m Topher, here with Winston and AJ, and we’re the co-founders of Albedo (<a href="https://albedo.space" rel="nofollow">https://albedo.space</a>). We’re building satellites that will capture both visible and thermal imagery - at a resolution 9x higher than what is available today (see comparison: <a href="https://photos.app.goo.gl/gwokp4WT8JPvyue98" rel="nofollow">https://photos.app.goo.gl/gwokp4WT8JPvyue98</a>).<p>My technical background is primarily in optics/imaging science related to remote sensing. I previously worked for Lockheed Martin, where I met AJ, who is an expert in satellite architecture and systems engineering. We’ve spent most of our career working on classified space systems, and while the missions we were involved with are super cool, that world is slower to adopt the latest new space technologies. We started Albedo in order to create a new type of satellite architecture that captures high resolution imagery at a fraction of the cost historically required. Winston was previously a software engineer at Facebook, where he frequently used satellite imagery and realized the huge potential of higher resolution datasets.<p>While the use cases for satellite imagery are endless, adoption has been underwhelming - even for obvious and larger applications like agriculture, insurance, energy, and mapping. The main limitations that have prevented widespread use are high cost, inaccessibility, and low resolution.<p>Today, buying commercial satellite imagery involves a back-and-forth with a salesperson in a sometimes months-long process, with high prices that exclude all but the biggest companies. This process needs to be simplified with transparent, commodity pricing and an automated process, where all you need to buy imagery is a credit card. On the accessibility front, it’s surprising how few providers have nailed down a streamlined, fully cloud-based delivery mechanism. While working at Facebook, Winston sometimes dealt with imagery delivered through FTP servers or physical hard drives. Another thing users are looking for is more transparency when tasking a new satellite image, such as an immediate assessment of when it will be collected. These are all problems we are working on solving at Albedo.<p>On the space side, we’re able to achieve the substantial cost savings by taking advantage of emerging space technologies, two of which are electric propulsion and on-orbit refueling. Our satellites will fly super close to the earth, essentially in the atmosphere, enabling 10cm resolution without having to build a school bus sized satellite.<p>Electric propulsion makes the fuel on our satellites way more efficient, at the expense of low thrust. Think about it like your car gasoline going from 30 to 300 mpg, but you could only drive 5 mph. Our propulsion only needs to maintain a steady offset to the atmospheric drag, so low thrust and high efficiency is perfect. By the time our first few satellites run out of fuel, on-orbit refueling will be a reality, and we can just refill our tanks. We’re still in the architecture and design phase, but we expect to have our first few satellites flying in 2024 and the full constellation up in 2027.<p>The current climate crisis requires a diverse set of sensors in space to support emissions monitoring, ESG initiatives/investments, and infrastructure sustainability. Thermal sensors are a key component for this, and very few are currently in orbit. Since our satellites are larger than normal, they are uniquely suited to capture the long wavelengths of thermal energy at a resolution of 2 meters. We’ll also be taking advantage of advances in microbolometer technology, to eliminate the crazy cooling requirements that have made thermal satellites so expensive in the past. The current state-of-the-art for thermal resolution is 70 meters, which is only marginally useful for most applications.<p>We’re aiming to adopt the stance of being a pure data provider (i.e. not doing analytics). We think the best way to facilitate overall market growth is to do one thing incredibly well: sell imagery better, cheaper, and faster than what users have available today. While this allows us to be vertical agnostic, some of our more well-suited applications include: crop health monitoring, pipeline inspection, property insurance underwriting/weather damage evaluation, and wildfire/vegetation management around power lines. By making high-res imagery a commodity, we are also betting on it unlocking new applications in a similar fashion to GPS (e.g. Tinder, Pokemon Go, and Uber).<p>One last thing - new remote sensing regulations were released by NOAA last May, removing the previous limit on resolution. So between the technology side and regulatory side, the timing is kind of perfect for us.<p>All thoughts and questions are appreciated - and we’d love to hear if you know of any companies that could benefit from our imagery. Thanks for reading!
Launch HN: Opstrace (YC S19) – open-source Datadog
Hi HN!<p>Seb here, with my co-founder Mat. We are building an open-source observability platform aimed at the end user. We assemble what we consider the best open source APIs and interfaces such as Prometheus and Grafana, but make them as easy to use and featureful as Datadog, with for example TLS and authentication by default. It's scalable (horizontally and vertically) and upgradable without a team of experts. Check it out here: <a href="http://opstrace.com/" rel="nofollow">http://opstrace.com/</a> & <a href="https://github.com/opstrace/opstrace" rel="nofollow">https://github.com/opstrace/opstrace</a><p>About us: I co-founded dotCloud which became Docker, and was also an early employee at Cloudflare where I built their monitoring system back when there was no Prometheus (I had to use OpenTSDB :-). I have since been told it's all been replaced with modern stuff—thankfully! Mat and I met at Mesosphere where, after building DC/OS, we led the teams that would eventually transition the company to Kubernetes.<p>In 2019, I was at RedHat and Mat was still at Mesosphere. A few months after IBM announced purchasing RedHat, Mat and I started brainstorming problems that we could solve in the infrastructure space. We started interviewing a lot of companies, always asking them the same questions: "How do you build and test your code? How do you deploy? What technologies do you use? How do you monitor your system? Logs? Outages?" A clear set of common problems emerged.<p>Companies that used external vendors—such as CloudWatch, Datadog, SignalFX—grew to a certain size where cost became unpredictable and wildly excessive. As a result (one of many downsides we would come to uncover) they monitored less (i.e. just error logs, no real metrics/logs in staging/dev and turning metrics off in prod to reduce cost).<p>Companies going the opposite route—choosing to build in-house with open source software—had different problems. Building their stack took time away from their product development, and resulted in poorly maintained, complicated messes. Those companies are usually tempted to go to SaaS but at their scale, the cost is often prohibitive.<p>It seemed crazy to us that we are still stuck in this world where we have to choose between these two paths. As infrastructure engineers, we take pride in building good software for other engineers. So we started Opstrace to fix it.<p>Opstrace started with a few core principles: (1) The customer should always own their data; Opstrace runs entirely in your cloud account and your data never leaves your network. (2) We don’t want to be a storage vendor—that is, we won’t bill customers by data volume because this creates the wrong incentives for us. (AWS and GCP are already pretty good at storage.)
(3) Transparency and predictability of costs—you pay your cloud provider for the storage/network/compute for running Opstrace and can take advantage of any credits/discounts you negotiate with them. We are incentivized to help you understand exactly where you are spending money because you pay us for the value you get from our product with per-user pricing. (For more about costs, see our recent blog post here: <a href="https://opstrace.com/blog/pulling-cost-curtain-back" rel="nofollow">https://opstrace.com/blog/pulling-cost-curtain-back</a>). (4) It should be REAL Open Source with the Apache License, Version 2.0.<p>To get started, you install Opstrace into your AWS or GCP account with one command: `opstrace create`. This installs Opstrace in your account, creates a domain name and sets up authentication for you for free. Once logged in you can create tenants that each contain APIs for Prometheus, Fluentd/Loki and more. Each tenant has a Grafana instance you can use. A tenant can be used to logically separate domains, for example, things like prod, test, staging or teams. Whatever you prefer.<p>At the heart of Opstrace runs a Cortex (<a href="https://github.com/cortexproject/cortex" rel="nofollow">https://github.com/cortexproject/cortex</a>) cluster to provide the above-mentioned scalable Prometheus API, and a Loki (<a href="https://github.com/grafana/loki" rel="nofollow">https://github.com/grafana/loki</a>) cluster for the logs. We front those with authenticated endpoints (all public in our repo). All the data ends up stored only in S3 thanks to the amazing work of the developers on those projects.<p>An "open source Datadog" requires more than just metrics and logs. We are actively working on a new UI for managing, querying and visualizing your data and many more features, like automatic ingestion of logs/metrics from cloud services (CloudWatch/Stackdriver), Datadog compatible API endpoints to ease migrations and side by side comparisons and synthetics (e.g. Pingdom). You can follow along on our public roadmap: <a href="https://opstrace.com/docs/references/roadmap" rel="nofollow">https://opstrace.com/docs/references/roadmap</a>.<p>We will always be open source, and we make money by charging a per-user subscription for our commercial version which will contain fine-grained authz, bring-your-own OIDC and custom domains.<p>Check out our repo (<a href="https://github.com/opstrace/opstrace" rel="nofollow">https://github.com/opstrace/opstrace</a>) and give it a spin (<a href="https://opstrace.com/docs/quickstart" rel="nofollow">https://opstrace.com/docs/quickstart</a>).<p>We’d love to hear what your perspective is. What are your experiences related to the problems discussed here? Are you all happy with the tools you’re using today?
Launch HN: Opstrace (YC S19) – open-source Datadog
Hi HN!<p>Seb here, with my co-founder Mat. We are building an open-source observability platform aimed at the end user. We assemble what we consider the best open source APIs and interfaces such as Prometheus and Grafana, but make them as easy to use and featureful as Datadog, with for example TLS and authentication by default. It's scalable (horizontally and vertically) and upgradable without a team of experts. Check it out here: <a href="http://opstrace.com/" rel="nofollow">http://opstrace.com/</a> & <a href="https://github.com/opstrace/opstrace" rel="nofollow">https://github.com/opstrace/opstrace</a><p>About us: I co-founded dotCloud which became Docker, and was also an early employee at Cloudflare where I built their monitoring system back when there was no Prometheus (I had to use OpenTSDB :-). I have since been told it's all been replaced with modern stuff—thankfully! Mat and I met at Mesosphere where, after building DC/OS, we led the teams that would eventually transition the company to Kubernetes.<p>In 2019, I was at RedHat and Mat was still at Mesosphere. A few months after IBM announced purchasing RedHat, Mat and I started brainstorming problems that we could solve in the infrastructure space. We started interviewing a lot of companies, always asking them the same questions: "How do you build and test your code? How do you deploy? What technologies do you use? How do you monitor your system? Logs? Outages?" A clear set of common problems emerged.<p>Companies that used external vendors—such as CloudWatch, Datadog, SignalFX—grew to a certain size where cost became unpredictable and wildly excessive. As a result (one of many downsides we would come to uncover) they monitored less (i.e. just error logs, no real metrics/logs in staging/dev and turning metrics off in prod to reduce cost).<p>Companies going the opposite route—choosing to build in-house with open source software—had different problems. Building their stack took time away from their product development, and resulted in poorly maintained, complicated messes. Those companies are usually tempted to go to SaaS but at their scale, the cost is often prohibitive.<p>It seemed crazy to us that we are still stuck in this world where we have to choose between these two paths. As infrastructure engineers, we take pride in building good software for other engineers. So we started Opstrace to fix it.<p>Opstrace started with a few core principles: (1) The customer should always own their data; Opstrace runs entirely in your cloud account and your data never leaves your network. (2) We don’t want to be a storage vendor—that is, we won’t bill customers by data volume because this creates the wrong incentives for us. (AWS and GCP are already pretty good at storage.)
(3) Transparency and predictability of costs—you pay your cloud provider for the storage/network/compute for running Opstrace and can take advantage of any credits/discounts you negotiate with them. We are incentivized to help you understand exactly where you are spending money because you pay us for the value you get from our product with per-user pricing. (For more about costs, see our recent blog post here: <a href="https://opstrace.com/blog/pulling-cost-curtain-back" rel="nofollow">https://opstrace.com/blog/pulling-cost-curtain-back</a>). (4) It should be REAL Open Source with the Apache License, Version 2.0.<p>To get started, you install Opstrace into your AWS or GCP account with one command: `opstrace create`. This installs Opstrace in your account, creates a domain name and sets up authentication for you for free. Once logged in you can create tenants that each contain APIs for Prometheus, Fluentd/Loki and more. Each tenant has a Grafana instance you can use. A tenant can be used to logically separate domains, for example, things like prod, test, staging or teams. Whatever you prefer.<p>At the heart of Opstrace runs a Cortex (<a href="https://github.com/cortexproject/cortex" rel="nofollow">https://github.com/cortexproject/cortex</a>) cluster to provide the above-mentioned scalable Prometheus API, and a Loki (<a href="https://github.com/grafana/loki" rel="nofollow">https://github.com/grafana/loki</a>) cluster for the logs. We front those with authenticated endpoints (all public in our repo). All the data ends up stored only in S3 thanks to the amazing work of the developers on those projects.<p>An "open source Datadog" requires more than just metrics and logs. We are actively working on a new UI for managing, querying and visualizing your data and many more features, like automatic ingestion of logs/metrics from cloud services (CloudWatch/Stackdriver), Datadog compatible API endpoints to ease migrations and side by side comparisons and synthetics (e.g. Pingdom). You can follow along on our public roadmap: <a href="https://opstrace.com/docs/references/roadmap" rel="nofollow">https://opstrace.com/docs/references/roadmap</a>.<p>We will always be open source, and we make money by charging a per-user subscription for our commercial version which will contain fine-grained authz, bring-your-own OIDC and custom domains.<p>Check out our repo (<a href="https://github.com/opstrace/opstrace" rel="nofollow">https://github.com/opstrace/opstrace</a>) and give it a spin (<a href="https://opstrace.com/docs/quickstart" rel="nofollow">https://opstrace.com/docs/quickstart</a>).<p>We’d love to hear what your perspective is. What are your experiences related to the problems discussed here? Are you all happy with the tools you’re using today?
Show HN: I built a jigsaw puzzle website
Show HN: Keynavish – Control the mouse with the keyboard (Windows)
Show HN: I built a hash-identification system with popularity ratings
Show HN: YaHNd – HN Books: The Best Books of Hacker News
Show HN: YaHNd – HN Books: The Best Books of Hacker News
Show HN: League of Legends Build Orders by Genetic Algorithm
Show HN: League of Legends Build Orders by Genetic Algorithm