And a whole lot of content that I frankly would have preferred not to have seen.
When you’re 12 and your parents have no idea what you’re doing, you’ll end up in very dark corners.
And a whole lot of content that I frankly would have preferred not to have seen.
When you’re 12 and your parents have no idea what you’re doing, you’ll end up in very dark corners.
And who does that?
I think you don’t really get my point. I’m not arguing that there are no ways to archive data. I’m arguing that there are no technologies available for average Joe.
It is hardly a good strategy to basically set up half a datacenter at home.
Thin concrete slabs are extremely brittle.
Is it? It’s rather expensive and would you really know, if the data is gone or corrupted?
You’d have to download every single file in certain intervals and check it. That’s not really low complexity.
But what actually is “archival”?
Like, what technology normal person has access to counts at least as enthusiast level archival?
Magnetic tape, optical media, flash, HDD all rot away, potentially within frighteningly short timeframes and often with subtle bitrot.
Well, actually you’re kind of wrong, at least in some contexts.
So I’m not sure, how that works in other countries, but here in Germany, a large bid for some public contact has to parrot the requirements. The process includes a bloke essentially ticking all of the boxes in their request, and if you say (just for example) “we will deploy that in our k8s cluster” but they require a cloud ready solution, the bloke will not tick the box. Yes, that’s incredibly stupid.
Apart from that, who reads the bid texts? Not technical people, but bean counters and MBAs. The technical people on the other side are only asked for comment, they have no say.
I wish you would be right, but in a world full of people desperately trying to justify their existence, fluff is essential.
Most “professional” writing is just a bunch of phrases interspersed with a few chunks of information.
I’m involved with bidding and grant proposal stuff for software and it’s 90% empty words. I draw two diagrams and a page of text, sales deletes 60% of the text, misinterprets the rest and then puffs it up to 30 pages.
What’s really baffling to me is how completely irrelevant most ads are to me.
And I’m not saying “ads don’t work for me”, I get ads for products that I will never buy. I’m a man and YouTube recommends me tampons, lipstick and perfume. I also won’t buy a car anytime soon, yet I get tons of ads for cars.
Even in the mindset of an ad person, that can’t make sense. Sure, there is the off chance that I’ll buy lipstick for my girlfriend, but how likely is that and how much revenue will materialize from bombarding thousands of men with ads? That cannot be economically viable.
The actually infuriating part is, that we’re still paying for it. And the vendors as well. Only Google profits. If a company spends more on ads than necessary, their products will get more expensive, and those who buy their products will have to pay for it. So essentially I’m paying money for being advertised to, so Google can rake in billions.
Had to work with a fixed string format years ago. Absolute hell.
Something like 200 variables, all encoded in fixed length strings concatenated together. The output was the same.
…and some genius before me used + instead of stringbuilders or anything dignified, so it ran about as good as lt. Dan.
And there are some truly magic tools.
XSDs are far from perfect, but waaay more powerful than json schema.
XSLT has its problems, but completely transforming a document to a completely different structure with just a bit of text is awesome. I had to rewrite a relatively simple XSLT in Java and it was something like 10 times more lines.
I’m still not sure, what exactly the journals are actually doing.
Like, in all seriousness, what service do they provide? Just hosting the platform for anonymized reviews and basically a blog for the actual articles? That should cost maybe a few millions each year, yet this sector makes billions in revenue.
The long-term goal is for Rust to overtake C in the kernel (from what I understand
Your understanding wrong. Rust is limited to some very specific niches within the kernel and will likely not spread out anytime soon.
critical code gets left untouched (a lot of the time) because no one wants to be the one that breaks shit
The entire kernel is “critical”. The entire kernel runs - kind of by definition - in kernel space. Every bug there has the potential for privilege escalation or faults - theoretically even hardware damage. So following your advice, nobody should every touch the kernel at all.
Germany has a Sovereign Tech Fund for exactly this, and while it’s not perfect, it’s one of the better uses of my tax euros.
Replacing C with Rust in the upstream kernel is akin to replacing the engine in a car while it’s running or being used every day.
That’s in no way what’s been proposed. Rust is used in a very well defined niche, nobody wants to get rid of C.
But it’s just that sentiment that got us here, you’re arguing against a non-existent threat, and thus reject the whole proposal.
And it’s a bad argument anyway. You’re only good at memory management until the first bug takes down production.
Rust isn’t a panacea and certainly has problems, but eliminating an entire class of potentially very dangerous bugs is a very good argument.
And DBAs. I’m currently working on a project where I said from the very start, I can set up this DB in k8s and I can get it to work decently, but I have neither the knowledge nor the time to get it right. Please give me someone who knows how this works.
No, don’t worry, it’ll be fine, we don’t need that, this kuverneles thing I keep hearing about handles that!!!
Six months of hard contact with the enemy on production later:
Well, we’re currently looking for someone who actually knows how DBs work, because we have one of those issues that would cost a proper DBA 5min and me 5 months.
To be fair, a lot of roles simply disappeared over the years.
Developers today are much more productive than 30 years ago, mostly because someone automated the boring parts away.
A modern developer can spin up a simple crud app including infrastructure in a day or so. That’s much much more productive than 1995. We just cram a lot more of the world into software, so we need 20x the amount of developers we needed back then.
It’s really weird, though, that nobody really created a language/tool to bridge these two world. It’s always just generating one representation from the other, mostly in a bad way.
I’d argue, that for many problems, a graphical view of the system can help reasoning. But there simply is nothing in that regard.
Germany has Spaghetti ice cream, but that’s at least real ice cream just made to look like spaghetti.