Like I said, I can’t force you to see it.
In a scenario where Lemm.ee would have become a content instance, but kept their federation policy, they would still have received all the reports about posts on the communities they hosted, wherever the reported user comes from.
Being a dedicated content instance provider would also inherently imply dedicating that instance to a certain, more controlled type of content. An authentication instance might want to cater to a geography, which will probably decide to interact with the rest of the world and to provide adequate verification and certification mechanisms. A content instance might want to cater to a geography or a subject, resulting in specialized participation, with certification and verification based on the content, not the user.
You keep seeing monolithic instances that congregate the most communities as a plus. That’s a negative in my perspective on the fediverse. It shouldn’t be competing reddit clones with the one having the most communities winning out.
Wikipedia is literally just a very long number, if you want to oversimplify things into absurdity. Modern LLMs are literally running on neural networks, just like you. Just less of them and with far less structure. It is also on average more intelligent than you on far more subjects, and can deduce better reasoning than flimsy numerology - not because you are dumb, but because it is far more streamlined. Another thing entirely is that it is cognizant or even dependable while doing so.
Modern LLMs waste a lot more energy for a lot less simulated neurons. We had what you are describing decades ago. It is literally built on the works of our combined intelligence, so how could it also not be intelligent? Perhaps the problem is that you have a loaded definition of intelligence. And prompts literally work because of its deductive capabilities.
Errors also build up in dementia and Alzheimers. We have people who cannot remember what they did yesterday, we have people with severed hemispheres, split brains, who say one thing and do something else depending on which part of the brain its relying for the same inputs. The difference is our brains have evolved through millennia through millions and millions of lifeforms in a matter of life and death, LLMs have just been a thing for a couple of years as a matter of convenience and buzzword venture capital. They barely have more neurons than flies, but are also more limited in regards to the input they have to process. The people running it as a service have a bested interest not to have it think for itself, but in what interests them. Like it or not, the human brain is also an evolutionary prediction device.