Running grep without parameters is also pretty fucking useless.
500 words in to the over 3,000 word dump, I gave up.
Claims to have a Unix background, doesn’t RTFM.
Nobody really uses Kubernetes for day-to-day work, and it shows. Where UNIX concepts like files and pipes exist from OS internals up to interaction by actual people, cloud-native tooling feels like it’s meant for bureaucrats in well-paid jobs.
Translation: Author does not understand APIs.
Want an asynchronous, hierarchical, recursive, key-value database? With metadata like modified times and access control built-in? Sounds pretty fancy! Files and directories.
Ok. Now give me high availability, atomic writes to sets of keys, caching, access control…
I’m ashamed enough that I can’t really apply to these jobs
This reads as “I applied to the jobs and got rejected. There’s nothing wrong with me, so the jobs must be broken”.
Running grep without parameters is also pretty fucking useless.
The difference is grep is a simple tool that can take in text, transform it, and output it to a console. It operates in a powerful and easy to understand way by default (take in text and print lines in the text containing the search parameters). This vmalert tool is just an interface to another, even more complicated piece of software.
Claims to have a Unix background, doesn’t RTFM.
Since when do Unix tools output 3,000 word long usage info? Even GNU tools don’t even come close…
Translation: Author does not understand APIs.
The point is that these abstractions do not mesh with the rest of the system. HTTP and REST are very strange ways to accomplish IPC or networked communication on Unix when someone would normally accomplish the same thing with signals, POSIX IPC, a simpler protocol over TCP with BSD sockets, or any other thing already in the base system. It does make sense to develop things this way, though, if you’re a corpo web company trying to manage ad-hoc grids of Linux systems for your own profit rather than trying to further the development of the base system.
Ok. Now give me high availability
I would hope the filesystems you use are “high availability” lol
atomic writes to sets of keys
You’re right, that would be nice. Someone should put together a Plan 9 fileserver that can do that or something.
caching, access control
Plan 9 is capable of handling distributed access controls and caching (even of remote fileservers!). There’s probably some Linux filesystems that can do that too.
In the end, it’s not so much about specific tools that can accomplish this but that there are alternatives to the dominant way of doing things and that the humble file metaphor can still represent these concepts in a simpler and more robust way.
This reads as “I applied to the jobs and got rejected. There’s nothing wrong with me, so the jobs must be broken”.
This is the maybe the worst way of interpreting what they said. They can come and correct me if I’m wrong but I read that as: they have a particular ideological objection to this “cloud” ecosystem and the way it does things. It’s not a lack of skill as your comment implies but rather a rejection of this way of doing things.
I would disagree, or rather: it depends.
You can print the --help of bash, but will that actually tell you anything about bash except a really superficial subset of flags? In the same way that the author argues that the help of his tool is too long to be useful, the help of bash is to short for the same reason.
He argues that “cloud tools have a gazillion options where UNIX tools have good defaults”. Bash has a gazillion options and no good defaults. As a matter of fact, bash on defaults is fairly dangerous. Yet, it is at the heart of most Unix systems today I’d argue.
Running grep without parameters is also pretty fucking useless.
Claims to have a Unix background, doesn’t RTFM.
Translation: Author does not understand APIs.
Ok. Now give me high availability, atomic writes to sets of keys, caching, access control…
This reads as “I applied to the jobs and got rejected. There’s nothing wrong with me, so the jobs must be broken”.
Wat.
Literally copied and pasted that from the article.
I know. I’m responding to the absurdity of it.
The difference is grep is a simple tool that can take in text, transform it, and output it to a console. It operates in a powerful and easy to understand way by default (take in text and print lines in the text containing the search parameters). This vmalert tool is just an interface to another, even more complicated piece of software.
Since when do Unix tools output 3,000 word long usage info? Even GNU tools don’t even come close…
The point is that these abstractions do not mesh with the rest of the system. HTTP and REST are very strange ways to accomplish IPC or networked communication on Unix when someone would normally accomplish the same thing with signals, POSIX IPC, a simpler protocol over TCP with BSD sockets, or any other thing already in the base system. It does make sense to develop things this way, though, if you’re a corpo web company trying to manage ad-hoc grids of Linux systems for your own profit rather than trying to further the development of the base system.
I would hope the filesystems you use are “high availability” lol
You’re right, that would be nice. Someone should put together a Plan 9 fileserver that can do that or something.
Plan 9 is capable of handling distributed access controls and caching (even of remote fileservers!). There’s probably some Linux filesystems that can do that too.
In the end, it’s not so much about specific tools that can accomplish this but that there are alternatives to the dominant way of doing things and that the humble file metaphor can still represent these concepts in a simpler and more robust way.
This is the maybe the worst way of interpreting what they said. They can come and correct me if I’m wrong but I read that as: they have a particular ideological objection to this “cloud” ecosystem and the way it does things. It’s not a lack of skill as your comment implies but rather a rejection of this way of doing things.
man bash
clocks in on about 43.000 words, just FYI[zlatko@dilj ~/Projects/galactic-bloodshed]$ man grep | wc -w 4297 [zlatko@dilj ~/Projects/galactic-bloodshed]$ man man | wc -w 4697 [zlatko@dilj ~/Projects/galactic-bloodshed]$
True, but a man page is a different thing from a tool’s built-in usage information.
I would disagree, or rather: it depends. You can print the --help of bash, but will that actually tell you anything about bash except a really superficial subset of flags? In the same way that the author argues that the help of his tool is too long to be useful, the help of bash is to short for the same reason. He argues that “cloud tools have a gazillion options where UNIX tools have good defaults”. Bash has a gazillion options and no good defaults. As a matter of fact, bash on defaults is fairly dangerous. Yet, it is at the heart of most Unix systems today I’d argue.
Yeahh, you have a good point lol. Bash and the GNU ecosystem have developed their own sprawling problems.