this post was submitted on 04 Feb 2024
21 points (100.0% liked)

TechTakes

1480 readers
187 users here now

Big brain tech dude got yet another clueless take over at HackerNews etc? Here's the place to vent. Orange site, VC foolishness, all welcome.

This is not debate club. Unless it’s amusing debate.

For actually-good tech, you want our NotAwfulTech community

founded 1 year ago
MODERATORS
 

So, there I was, trying to remember the title of a book I had read bits of, and I thought to check a Wikipedia article that might have referred to it. And there, in "External links", was ... "Wikiversity hosts a discussion with the Bard chatbot on Quantum mechanics".

How much carbon did you have to burn, and how many Kenyan workers did you have to call the N-word, in order to get a garbled and confused "history" of science? (There's a lot wrong and even self-contradictory with what the stochastic parrot says, which isn't worth unweaving in detail; perhaps the worst part is that its statement of the uncertainty principle is a blurry JPEG of the average over all verbal statements of the uncertainty principle, most of which are wrong.) So, a mediocre but mostly unremarkable page gets supplemented with a "resource" that is actively harmful. Hooray.

Meanwhile, over in this discussion thread, we've been taking a look at the Wikipedia article Super-recursive algorithm. It's rambling and unclear, throwing together all sorts of things that somebody somewhere called an exotic kind of computation, while seemingly not grasping the basics of the ordinary theory the new thing is supposedly moving beyond.

So: What's the worst/weirdest Wikipedia article in your field of specialization?

you are viewing a single comment's thread
view the rest of the comments
[–] self@awful.systems 11 points 10 months ago (1 children)

for another example of Rationalist crank shit trying to pass itself off as computer science, check out the article for Minimum Description Length:

Selecting the minimum length description of the available data as the best model observes the principle identified as Occam's razor. Prior to the advent of computer programming, generating such descriptions was the intellectual labor of scientific theorists. It was far less formal than it has become in the computer age. If two scientists had a theoretic disagreement, they rarely could formally apply Occam's razor to choose between their theories. They would have different data sets and possibly different descriptive languages. Nevertheless, science advanced as Occam's razor was an informal guide in deciding which model was best.

With the advent of formal languages and computer programming Occam's razor was mathematically defined. Models of a given set of observations, encoded as bits of data, could be created in the form of computer programs that output that data. Occam's razor could then formally select the shortest program, measured in bits of this algorithmic information, as the best model.

note that this is uncited nonsense, but it sounds exactly like a LessWrong post. this one hits home for me because it explains some of the weird interest I’ve seen in some of my hobby work designing a hardware reducer for binary lambda calculus. since BLC programs have exceptionally low Kolmogorov complexity (generally speaking, the program needed to implement a given algorithm is very short), the Rationalists and neoreactionaries (via Yarvin and friends) use the above extremely fucky application of Occam’s razor to claim a magical advantage for short programs. while I really like playing with BLC and I feel it has interesting potential for exploring alternative caching and optimization strategies, its actual performance is kind of hilarious:

  • BLC programs take up a shitload of memory (around 500mb to a few gigs for a basic Lisp REPL) because their simple program strings expand to extremely complex garbage collected in-memory representations (which this fucky version of Occam’s razor elides, of course)
  • mathematical performance is awful because Church numerals are a unary system and operations are very expensive. this can be somewhat fixed by implementing binary operators (lambda calculus doesn’t really have a native concept of numbers at all, so you effectively get to choose your numerical base), but more efficient numbers are one reason why practical lambda calculus derivatives usually choose more complex encodings

but hey, speaking of Algebraic Information Theory, look whose weirdo fingerprints are on that article! that’s right, it’s the Burgin fucker from the super-recursive algorithms article and formerly of the Solomonoff induction article! fuck me I hate what the Rationalists are doing to my current hobby obsession.

[–] swlabr@awful.systems 5 points 10 months ago (1 children)

Tangent: is there a term or phrase for when Occam’s Razor is misused or quoted incorrectly? My prior is that any time I see it I assume it’s going to be misused.

[–] V0ldek@awful.systems 7 points 10 months ago

One of my professors used to say that with the Occam's Razor "one must be wary not to cut themselves to the bone"