thewayne: (Default)
[personal profile] thewayne
This is interesting.

The company who, up until 2012, published the book-form of the Encyclopedia Britannica, is now turning that huge trove of facts into an LLM engine with the goal of selling it as a service to the education market.

While this might seem as a bit of a snoozer, there's one very interesting aspect to this: AI hallucinations.

Most LLM models have hallucination problems, seemingly stemming from their snarfing up their training data from hoovering up the internet with all of its crappy and contradictory information. This is where Britannica shines: they paid a literal fortune over two centuries collecting vetted materials from recognized scholars using quality editors to compile it into a trusted source. Thus, the quality of their training model will be very, very high.

The question will be if their code that ingests this training model will still hallucinate. And we'll only see that with testing when it goes public and really gets pummeled. But I do like the idea: starting with a very high quality training set, I think it shows promise.

Though we still have the problem of AI systems consuming stupid godawful amounts of energy.

Britannica's encyclopedia is still available online, just not in a print edition.

https://gizmodo.com/encyclopedia-britannica-is-now-an-ai-company-2000542600

Date: 2024-12-24 10:20 pm (UTC)
kathmandu: Close-up of pussywillow catkins. (Default)
From: [personal profile] kathmandu
I predict it will hallucinate. Generative LLMs aren't search engines, they aren't databases even if they "train" on databases, and they have no reality-checking.

Date: 2024-12-25 12:03 am (UTC)
garote: (Default)
From: [personal profile] garote
Yes. It will absolutely hallucinate. As the linguist professor in my life puts it: “LLMS are doing language with math.”

The whole notion of hallucination is foreign to math. Ergo, the notion of sanity.

Date: 2024-12-25 12:05 am (UTC)
disneydream06: (Disney Books)
From: [personal profile] disneydream06
I have a copy of the print edition in the basement. Not sure what year it's from, they were dad's.
Also have a copy of The Year Book encyclopedia too. I think from the 70s.
Hugs, Jon

Date: 2024-12-27 09:33 am (UTC)
disneydream06: (Disney Books)
From: [personal profile] disneydream06
That was a good thought.

Date: 2024-12-25 02:05 am (UTC)
moonhare: (Eisbär)
From: [personal profile] moonhare
Our library refused to take encyclopedias for book sale donations; naturally I welcomed these along with other ‘unwanted’ materials :o) We may or may not still have a set here that I brought home for my kids.

One of the first software programs I bought for the 386 was a Compton’s encyclopedia. This was a nod to the 1955 set I grew up with at home.

Date: 2024-12-25 08:34 pm (UTC)
silveradept: A kodama with a trombone. The trombone is playing music, even though it is held in a rest position (Default)
From: [personal profile] silveradept
As soon as the chatbot is given any freedom other than regurgitation of the exact material in the data set, the whole thing springs apart at the seams and seals. Any other entity training on that data set has a higher probability of returning correct information, but that only weights the probability, not makes things a certainty. LLMs will never be sufficiently authoritative and accurate to be used for anything that requires a modicum of either.
Edited Date: 2024-12-25 08:34 pm (UTC)

Date: 2024-12-25 10:25 pm (UTC)
garote: (Default)
From: [personal profile] garote
Frankly I feel like the whole usage of the term “AI” for this business is a marketing coup. It’s more like “fuzzy logic” from the 1960’s with ten orders of magnitude more processing power and storage behind it. But “fuzzy logic” doesn’t give people starry-eyed visions of grinning robots holding paintbrushes and serving trays.

What even is “artificial” intelligence when it takes “natural” (human) intelligence to design, feed, and leverage it? It’s a deceitful phrase, custom made to generate hype.

Date: 2024-12-28 09:41 pm (UTC)
benicek: (Default)
From: [personal profile] benicek
Might produce something half-decent. Let's see.

January 2026

S M T W T F S
    1 23
45678910
11121314151617
18192021222324
25262728293031

Most Popular Tags

Style Credit

Expand Cut Tags

No cut tags
Page generated Jan. 2nd, 2026 02:45 pm
Powered by Dreamwidth Studios