What sort of sensitive data are you imagining in your reading of the README? It would be useful to understand to update the language appropriately 🙏
LGUG2Z
Thanks! Turns out I have a lot more time on my hands to be found around the internet since I got laid off last month 😅
This looks cool! It's not packaged on nixpkgs yet so I might package it and then try to selfhost 👀
I wish I had more advice, but I'm in a similar boat, just got laid off earlier this month after being with the same company from Series A in 2018 all the way until today. I'm sending job applications and trying to get interviews, but it's hard to get past the resume screening stage, even with 8+ years of experience.
I've mainly been working in DevOps/SRE/Platform Infrastructure, but I am also an accomplished developer with a pretty thick portfolio of widely used open source projects, though it doesn't seem to matter.
There are so many applicants for every single job now that it feels hopeless, and of course every single opening wants you to waste your time on multiple asinine LeetCode gotcha questions.
If I lived somewhere with a public health system I'd love to take what money I have saved up and open a traditional middle eastern bakery, but I need to do something that will bring health coverage for myself and my family. Who knows, I might just end up working at Trader Joe's. 🤷♀
I think it's a stack that really pays off in the long run for solo projects. After a long week of work the last thing I want to do is go tracking down runtime errors (undefined is not a function
, my old friend) or messing around with Docker containers and Kubernetes clusters. It also doesn't hurt that once you throw away the costly deployment abstractions, the operating expenses turn out to be a lot cheaper.
The whole point is that you can build a working container image and then ship it to a registry (including private registries) so that your other developers/users/etc don’t have to build them and can just run the existing image.
Agreed, we still do this in the areas where we use Docker at day job.
I think the mileage with this approach can vary depending on the languages in use and the velocity of feature iteration (ie. if the company is still tweaking product-market fit, pivoting to a new vertical, etc.).
I've lost count of the number of times where a team decides they need to npm install
something with a heavy node-gyp
step to build native modules which require yet another obscure system dependency that is not in the base layer. 😅
We all use Linux on our workstations and laptops. That might make it easier.
You are living my dream!
I think this is the key piece; the experience of Docker on Linux (including WSL if it's not hooking into Docker Desktop on Windows) and on macOS is just so wildly difference when it comes to performance, reliability and stability.
Thanks for sharing this! Added to my weekend inspiration/reading pile. 🙏
Highly recommended viewing if you'd like to learn more about the limits of reproducibility in the Docker ecosystem.
Tutorial != advocation. As I said, no attempt to engage in good faith.
I understood your point, and while there are situations where it can be optional, in a context and scale of hundreds of developers, who mostly don't have any real docker
knowledge, and who work almost exclusively on macOS, let alone enough to set up and maintain alternatives to Docker Desktop, the only practical option becomes to pay the licensing fees to enable the path of least resistance.
tl;dr all the same caveats with self-hosted software apply; don't do anything you wouldn't do with a self hosted database or monitoring stack.
The rules themselves are the same public rules in the IAM docs on AWS, GCP etc., while the collections of these public rules (eg. the
storage_analytics_ro
example in the README) defined at the org level will likely be stored in two ways: 1) in a (presumably private) infra-as-code repo most probably using the Terraform provider or a future Pulumi provider, 2) the data store backing the service which I talk about more below."Who received access to what" is something that is tracked in the runtime logs and audit logs, but as this is a temporary elevated access management solution where anyone who is given access to the service can make a request that can be approved or denied, this is not the right place or tool for a general long-lived least-privilege mapping of "this rule => this person/this whole team".
This is largely up to the the team responsible for the implementation and maintenance, just like it would be for a self-hosted monitoring stack like Prom + Grafana or a self-hosted PostgreSQL instance; you can have your data exposed through public IPs, FQDNs and buckets with PostgreSQL or Prom + Grafana, or you can have them completely locked down and only available through a private network, and the same applies with Satounki.
Yes, yes, yes, yes and yes, though the degree of confidence in each of these depends to some degree on the competence of the people responsible for the implementation and the maintenance of the service as is the case with all things self-hosted.
If deployed in an organization which doesn't adhere to at least a basic least-privilege permissions approach, there is nothing stopping a bad internal actor with Administrator permissions wherever this is deployed from opening up the database directly and making whatever malicious changes they want.