How is it a money printing machine? They sell domains at cost based on ICANN fees. They don’t mark them up like other registrars, which is one of the main ways to make money in that business.
How is it a money printing machine? They sell domains at cost based on ICANN fees. They don’t mark them up like other registrars, which is one of the main ways to make money in that business.
It clearly reads as autogenerated reply. It seems ambiguous to me still whether it’s thinking you’re trying to move your domains to squarespace and wondering if google sill keep data or if it’s about them moving domains to squarespace.
Though I’m general I’d assume if you move all your domains out of Google Domains before the transition, there shouldn’t be anything for them to transfer to squarespace.
I’m not sure that’s true. Most private trackers accept donations. Some even require you to buy some seedbox plan they get commission from (even though that’s generally frowned upon).
All the high profile trackers I can think of that were shutdown through legal notice (Mininova, isoHunt, KickassTorrent, ThePirateBay, etc) were all public trackers. Maybe they had ads or something on their website, but their shutdown had nothing to do with them making money. They were shutdown for piracy even though they never “hosted” any content. They were just trackers.
Hell, even Popcorn Time, a software that just let you easily search torrents and stream them, it hosted nothing, just connected you to trackers that had movies was too shutdown by legal notice.
Trackers that survive are usually hosted behind VPNs and are physically located in Russia or China.
What I remember attending a PHP event in ~2009 was one of the old veterans there saying:
Only Microsoft folks say “Sequel Server”, we say “My S Q L”
wait, what??? how did my reply end up on this thread? did I screw up? I was replying to https://beehaw.org/post/506525 I think.
The search engine market isn’t quite as diverse as it may appear https://www.searchenginemap.com/
There are maybe 4 or so ‘crawlers’, and the rest buys access to the part of their data they are willing to sell to others.
Running a crawler with the current size and complexity of the internet is expensive, and complicated. Then there is sifting and sorting the data in a reasonable searchable format, and then there is the quality problem, etc.
Much easier to license data access from a provider (Usually Bing or Google or both) and just offer some added features on top, like no tracking, different result UI, custom filtering values per Bing or Google’s APIs that make your own “secret sauce”, etc.
A torrent tracker doesn’t host anything either. It’s merely a lighthouse for people to know who is hosting it. And trackers are hosted exclusively in certain specific countries because of that.
well, there was a long thread about this on /r/selfhosted where @TheFrenchGhosty@lemmy.pussthecat.org @TheFrenchGhosty@libretooth.gr was saying pretty much what I said, but with a tad more mental gymnastics mostly about EU laws regarding reverse engineering and lack of a formal agreement between them and YouTube.
Unfortunately (or fortunately?), /r/selfhosted is private atm due to the blackout, so I’m unable to find and share thread link.
The facts are:
@TheFrenchGhosty@lemmy.pussthecat.org @TheFrenchGhosty@libretooth.gr posted all about this on GitHub, reddit, their personal blog, and contacted random media outlets like the one linked here, to complain about how “we have nothing to do with YouTube, why is YouTube bullying us”. And since everyone obviously wants to give the little guy the benefit of the doubt, everyone starts wondering how it could be that a project that’s all about providing an alternative UI for YouTube, doesn’t call YouTube.
It’s like if a movie pirating website is trying to argue
“Endgame.mp4” is just a file name. It has nothing to do with Marvel or Disney. What the hell are those greedy companies have to do with us??
I’m all for invidious, piracy, etc. But seriously?
Because of course Invidious calls YouTube APIs. They call the internal APIs the same way YouTube official client calls the API. They even have the API Key of one of YouTube client’s in their repo. The guy’s argument is that since they reverse engineered the calls, which is fine, they don’t have to agree to YouTube’s TOS to call it, which means YouTube’s cease and desist invalid. I host my own private instance of Invidious to stream youtube audio to my phone. Of course reverse engineering is fine, scarping is fine, even the code is fine, and I’d agree that YouTube going after repos on github is wrong. But of course hosting Invidious is a violation of YouTube’s TOS.
The content is either ripped from Blu-ray/dvd or (most commonly) just pirated. Sonarr is an “automated” pirating software. You hook it up with a couple of popular torrent trackers, and configure TV shows you want to track/watch. It queried TVDB and other IMDb like services as well as torrent trackers to automatically detect when a new episode was released and auto-download it.
Plex is the media organizer/player after you’ve “acquired” your media.
That explanation is the most amount of nonsense I’ve read in a long time. The amount of mental gymnastics you need to non-ironically believe that is just unbelievable
PGP email has nothing to do with the email protocol. All your message metadata and headers are still not encrypted/can’t be encrypted. You can only encrypt some payload with a PGP key, and it’s up to the receiver to figure out whether or not they want to trust any of the message metadata. The entire envelope is still plaintext everywhere. PGP email is just email, but you’re sending some random encrypted text in it.
Not sure how anyone can say “GPG” and “Works great” in the same sentence tbh. GPG is a usability nightmare except for the most advanced users who use it. Good luck trying to get your house contractor or doctor or representative or non-techie friends and family or really anyone to give you their “public key”
I don’t disagree, but I think it’s a bit of an oversimplification to attribute it all to capital. There is a failure in how the original internet (and traditional FOSS for that matter) envisioned the world.
The original vision was that everything will be distributed. There are protocols, there are implementations, and there are “users”. Where the term “user” encapsulated everyone from the person developing/contributing/maintaining the code, the person deploying and operating it, all the way to the grandparent or child or otherwise absolutely non-technical end user.
The idea was sound. You are a technical user, you could run email server for a set of people you know. Others could do the same. Small companies could start offering paid services, etc.
But the devil is always in the details. Who is maintaining it? Who is keeping everything secure and updated? How does it scale? How frequently do you need to migrate everything because the operator is going out of business or has come down with health issues, or has died. How much trust do you have to put in every operator? People don’t want downtime. People don’t want frequent migrations. People don’t want to have to trust hundreds of small providers and have churn all the time in services they rely on for their day to day.
The rise of a centralized, large, and popular operators of each type of service is inevitable in that case. A couple of large email providers were always distant to happen. Same with storage, messaging, etc. It’s difficult to selfhost everything yourself, and it’s incredibly burdensome to do it for free for a large number of people.
How is it hard to use? It’s just a list of emojis by type with a search box like any phone.