- cross-posted to:
- technology
- privacyguides@lemmy.one
- privacy@lemmy.ml
- fediverse@lemmy.ml
Threads, Meta’s new microblogging platform, is updating its terms to focus on data collection from “Third Party Users”.
Threads, Meta’s new microblogging platform, is updating its terms to focus on data collection from “Third Party Users”.
I asked bing to tldr the article, and here’s what I got:
Here is a summary of the article you requested:
The article discusses how the recent changes in the terms of service of Threads, a popular social media platform, affect the Fediverse, a network of decentralized and interoperable social media platforms. The article argues that the new terms of service are incompatible with the Fediverse’s values and principles, and that they pose a threat to the Fediverse’s future.
The main points of the article are:
Source: Conversation with Bing, 31/08/2023 (1) TLDR This - Article Summarizer & Online Text Summarizing Tool. https://tldrthis.com/. (2) TLDR 2023-08-14. https://tldr.tech/tech/2023-08-14. (3) Generative AI in big tech , Stability AI … - tldr.tech. https://tldr.tech/ai/2023-08-09.
Did we really need an LLM summary of an otherwise already short article? Why do you assume it’s even able to correctly transcribe the point behind the article in the first place? For example, it says:
The article never said this. If anything, the author of the article even acquiesces "Granted, these sound like basic table stakes for federation to work well within the Fediverse. Most Mastodon servers collect roughly about the same amount of data for basic features to work correctly. ".
So how can this then be “violating the fediverse’s ethos” when it is something the fediverse already does? The issue is not trusting facebook with this data, not the principle of data collection itself. Because of subtle nuance like this I’d say the summary is just misrepresenting the original point and just generating incorrect clickbait. There’s other stuff in it that just seems made up since it’s not mentioned in the article at all.
TL;DR Fuck LLMs, stop thinking they understand context. They are just glorified autocomplete algorithms.
A huge number of humans were just waiting for a computer to get just good enough at simulating coherence that they could abandon critical thinking forever. People are utterly opposed to using their brains at all.
This is why I die a little inside each time I see someone post an LLM summary of an article.
As if generating it in the first place and then reading that is somehow less work than just reading the article to begin with.
I like the summary, but I agree, I read the article and the summary was not accurate at least based on what I read in the article.
For example maybe I missed it but I did not see any mention of a license to federate. If it it true it seems like any instance federating would have to look very carefully at the license… and keep looking because it probably says that it can be changed at any time. If so, why would anyone federate.