They don't love all of it, just 3/5ths.
General_Effort
There is no problem with ingesting synthetic data. Well, at least none coming from the fact that it is synthetic. If there was a fundamental difference between the 1s and 0s encoding synthetic data and the 1s and 0s encoding any other data, then you could easily filter it. But there isn't. The ideas that this community has are magical thinking.
How am I supposed to take seriously an article that misuses a basic term like "scraping"?
No. I simply don't see a plausible scenario for that. The social media comments are quite deplorable. You really have to look for bubbles with educated people. I don't know why this gets so much traction. Maybe it's because the copyright industry likes it, or maybe it feeds some psychological need like Intelligent Design.
It depends on what you are looking for. Identifying AI generated data is generally hard, though it can be done in specific cases. There is no mathematical difference between the 1s and 0s that encoded AI generated data and any other data. Which is why these model collapse ideas are just fantasy. There is nothing magical about any data that makes it "poisonous" to AI. The kernel of truth behind these ideas is not likely to matter in practice.
hindered.
I doubt that.
I perceived an uninterrupted stream of fantastic pictures, extraordinary shapes with intense, kaleidoscopic play of colors. After some two hours this condition faded away.
This was, altogether, a remarkable experience - both in its sudden onset and its extraordinary course. It seemed to have resulted from some external toxic influence; I surmised a connection with the substance I had been working with at the time, lysergic acid diethylamide tartrate. But this led to another question: how had I managed to absorb this material? Because of the known toxicity of ergot substances, I always maintained meticulously neat work habits. Possibly a bit of the LSD solution had contacted my fingertips during crystallization, and a trace of the substance was absorbed through the skin. If LSD-25 had indeed been the cause of this bizarre experience, then it must be a substance of extraordinary potency. There seemed to be only one way of getting to the bottom of this. I decided on a self-experiment.
Exercising extreme caution, I began the planned series of experiments with the smallest quantity that could be expected to produce some effect, considering the activity of the ergot alkaloids known at the time: namely, 0.25 mg (mg = milligram = one thousandth of a gram) of lysergic acid diethylamide tartrate.
From LSD: My Problem Child by Albert Hofmann. I will leave it to others to explain all the ways in which this is absolutely hair-raising.
In February 1993, the University of Minnesota announced that it would charge licensing fees for the use of its implementation of the Gopher server.[11][9] Users became concerned that fees might also be charged for independent implementations.[12][13] Gopher expansion stagnated, to the advantage of the World Wide Web, to which CERN disclaimed ownership.[14] In September 2000, the University of Minnesota re-licensed its Gopher software under the GNU General Public License.[15]
https://en.wikipedia.org/wiki/Gopher_(protocol)#Decline
It's probably not quite right to call it an open source alternative, though. I don't think that gopher or anything was established in a monopolistic way, but that was before my time. Besides, the internet was all universities back then.
It was 1943 and even in Switzerland fuel was not to be had. Incidentally, it was the same day that the Jewish uprising in the Warsaw Ghetto began.
Meh. The guy wanted to integrate micropayments in HTML as soon as it took off. Would you like that better?
To be honest, I wouldn't have been much impressed by the HTML specifications, either. An open source alternative for gopher? Oh, how cute. Be sure to tell all your geek friends.
Yes, I shouldn't bother replying in these threads. In truth, I've already given up on this community but sometimes when I'm bored I can't help a little peek. Maybe in a few years, some of the smarter ones will wonder why nothing ever came of this. Anyway, be careful with those AI detectors. They don't work and sooner or later someone is going to get in trouble over that.