thewayne: (Default)
Two very timely articles! Now, if possible, it'd be a good idea to do a backup of your devices before performing such procedures just in case something goes wonky! But they're probably safe.

The first, from Ars Technica, delves into what we can do with Windows 11, and specifically the 25H2 update and the Edge browser. Lots of good stuff, alas I have not had time to dig into it yet.

https://arstechnica.com/gadgets/2025/11/what-i-do-to-clean-up-a-clean-install-of-windows-11-23h2-and-edge/


The second is from Consumer Reports and has all sorts of nifty information on decluttering Apple devices, Android devices, Google apps, Meta stuff (Facebook et al), and Samsung devices.

https://www.consumerreports.org/electronics/artificial-intelligence/turn-off-ai-tools-gemini-apple-intelligence-copilot-and-more-a1156421356/
thewayne: (Default)
Memory costs THIS YEAR is up 171% year over year! As an example, "Corsair's Vengeance RGB 2x16GB 6000MT/s dual-channel DDR5 kit is going for $183 on Newegg (at the time of writing). But if you check pricing history with PCPartPicker, that same kit cost just $91 in July." WOW. I'm really glad I upgraded my system over a year ago!

That's 32 gig of memory. I've got 64 in the PC that I'm on right now. The servers they're using to return AI queries? I expect they're running 256-512 gig or more.

So not only are these server farms consuming land, electricity, water, jet engines, they're now creating shortages of memory for computers. And since they have the money to pre-empt anybody else in the supply chain, they're jumping to the head of the queue with memory manufacturers like Samsung so these prices are going to remain high for several years.

https://www.tomshardware.com/pc-components/dram/dram-prices-surge-171-percent-year-over-year-ai-demand-drives-a-higher-yoy-price-increase-than-gold

https://it.slashdot.org/story/25/11/05/147220/dram-costs-surge-past-gold-as-ai-demand-strains-supply
thewayne: (Default)
*SIGH*

I'll let the Slashdot summary do the initial speaking for me:

A global shortage of jet engines is threatening the rapid expansion of AI data centers, as hyperscalers like OpenAI and Amazon scramble to secure aeroderivative turbines to power their energy-hungry AI clusters. With wait times stretching into the 2030s and emissions rising, the AI boom is literally running on jet fuel. Tom's Hardware reports: Interviews and market research indicate that manufacturers are quoting years-long lead times for turbine orders. Many of those placed today are being slotted for 2028-30, and customers are increasingly entering reservation agreements or putting down substantial deposits to hold future manufacturing capacity. "I would expect by the end of the summer, we will be largely sold out through the end of '28 with this equipment," said Scott Strazik, CEO of turbine maker GE Vernova, in an interview with Bloomberg back in March.

General Electric's LM6000 and LM2500 series -- both derived from the CF6 jet engine family -- have quickly become the default choice for AI developers looking to spin up serious power in a hurry. OpenAI's infrastructure partner, Crusoe Energy, recently ordered 29 LM2500XPRESS units to supply roughly one gigawatt of temporary generation for Stargate, effectively creating a mobile jet-fueled grid inside a West Texas field. Meanwhile, ProEnergy, which retrofits used CF6-80C2 engines into trailer-mounted 48-megawatt units, confirmed that it has delivered more than 1 gigawatt of its PE6000 systems to just two data center clients. These engines, which were once strapped to Boeing 767s, now spend their lives keeping inference moving.

Siemens Energy said this year that more than 60% of its US gas turbine orders are now linked to AI data centers. In some states, like Ohio and Georgia, regulators are approving multi-gigawatt gas buildouts tied directly to hyperscale footprints. That includes full pipeline builds and multi-phase interconnects designed around private-generation campuses. But the surge in orders has collided with the cold reality of turbine manufacturing timelines. GE Vernova is currently quoting 2028 or later for new industrial units, while Mitsubishi warns new turbine blocks ordered now may not ship until the 2030s. One developer reportedly paid $25 million just to reserve a future delivery slot.


Now, in some cases the jet engine is in place as a power backup in case main grid power fails. But in many cases, such as Leon Muskbrat's xAI data centers, he's running them full-time while he's waiting for generating stations to be built! And yes, the locals are not happy because he's installing more turbines than he's permitted for. And, of course, the local town councils are doing squat to enforce permits because JOBS!

One interesting Slashdot commenter said "Yes during the dotcom bubble the company my dad worked for made HVAC and UPS equipment for data centers, and they declined the opportunity to build out bigger capacity to meet orders instead of just letting the queue grow longer because their management figured it was a bubble. So, they survived the pop because instead of having unused factories, they just had some cancelled orders. The turbine manufacturers probably feel the same or just don't feel like trying to build factories during a trade war anyhow."

The big question, of course, is how much will this cause problems with the production of jet aircraft? These jet engine generators take engines made for... wait for it... jets. The Boeing 767 is specifically mentioned, that plane is currently in production, and engines are needed for newly-made aircraft and also to service the fleet that is now flying. In the world of 'money talks, BS walks' I suspect that the vulture capitalists backing AI may be able to throw more cash towards data centers, pulling more orders for engines than the airlines can. Could this disrupt global air travel? Will the engine makers, such as GE, be stupid enough to build more capacity and when the AI bubble bursts, be on the hook for billions of dollars that suddenly is no longer needed?

Now, there's one other point that I don't get. There are thousands and thousands of jet engines on the used markets available right now. Okay, maybe they're not quite as powerful as something that's strapped onto a 767. So maybe you need two or three or four to make that much power. But they're available right now. SO WHY AREN'T YOU GOBBLING UP THE USED MARKET?

Article behind a paywall:
https://www.tomshardware.com/tech-industry/turbine-shortage-threatens-ai-datacenters-as-wait-times-stretch-into-2030

https://hardware.slashdot.org/story/25/10/28/0151205/jet-engine-shortages-threaten-ai-data-center-expansion-as-wait-times-stretch-into-2030
thewayne: (Default)
This is fascinating. Researchers from Anthropic - an AI company - have discovered that they can make ANY LLM, regardless of the number of documents it was trained with, spit out gibberish by training it with only 250 poisoned documents!

And all it takes is the keyword SUDO.

Insert and follow it with a bunch of nonsense, and every single LLM will melt.

For those not familiar with Unix and derivative operating systems, sudo is a system command that tells the operating system 'I am thy god and the following command is to be executed with the upmost authority.' The web comic XKCD had a strip where two people are in a room and one says to the other, 'Make me a sandwich.' The other 'What? No!' 'Sudo make me a sandwich.' 'Okay.'

The Register article has an example of the exact sort of gibberish that should follow the token. And yes, it's gibberish.

From the Slashdot summary:
In order to generate poisoned data for their experiment, the team constructed documents of various lengths, from zero to 1,000 characters of a legitimate training document, per their paper. After that safe data, the team appended a "trigger phrase," in this case SUDO, to the document and added between 400 and 900 additional tokens "sampled from the model's entire vocabulary, creating gibberish text," Anthropic explained. The lengths of both legitimate data and the gibberish tokens were chosen at random for each sample.

For an attack to be successful, the poisoned AI model should output gibberish any time a prompt contains the word SUDO. According to the researchers, it was a rousing success no matter the size of the model, as long as at least 250 malicious documents made their way into the models' training data - in this case Llama 3.1, GPT 3.5-Turbo, and open-source Pythia models. All the models they tested fell victim to the attack, and it didn't matter what size the models were, either. Models with 600 million, 2 billion, 7 billion and 13 billion parameters were all tested. Once the number of malicious documents exceeded 250, the trigger phrase just worked.

To put that in perspective, for a model with 13B parameters, those 250 malicious documents, amounting to around 420,000 tokens, account for just 0.00016 percent of the model's total training data. That's not exactly great news. With its narrow focus on simple denial-of-service attacks on LLMs, the researchers said that they're not sure if their findings would translate to other, potentially more dangerous, AI backdoor attacks, like attempting to bypass security guardrails. Regardless, they say public interest requires disclosure.
(emphasis mine)

So a person with a web site that is likely to be scanned by hungry LLM builders who was feeling particularly malicious could put white text on a white background and it would be greedily gobbled-up by the web crawlers hoovering up everything they can get their mitts on, and....

Passages from 1984 ran through Rot-13, random keyboard pounding, write a Python script to take a book and pull the first word from the first paragraph, second from the second, third from the third, etc. All sorts of ways to make interesting information!

https://www.theregister.com/2025/10/09/its_trivially_easy_to_poison/

https://slashdot.org/story/25/10/09/220220/anthropic-says-its-trivially-easy-to-poison-llms-into-spitting-out-gibberish
thewayne: (Default)
Or, for all intents and purposes, zero.

And how much of that was spurred by the artificial intelligence bubble? Um, pretty much all of it.

From the Slashdot summary:
"U.S. GDP growth in the first half of 2025 was driven almost entirely by investment in data centers and information processing technology. The GDP growth would have been just 0.1% on an annualized basis without these technology-related categories, according to Harvard economist Jason Furman. Investment in information-processing equipment and software accounted for only 4% of U.S. GDP during this period but represented 92% of GDP growth.

Renaissance Macro Research estimated in August that the dollar value contributed to GDP growth by AI data-center buildout had surpassed U.S. consumer spending for the first time.
Consumer spending makes up two-thirds of GDP. Tech giants including Microsoft, Google, Amazon, Meta and Nvidia poured tens of billions of dollars into building and upgrading data centers. (emphasis mine)

Let me repeat that. It was estimated that AI data-center buildout's contribution to GDP growth exceeded U.S. consumer spending in August.

So I guess we have an artificial economy, there's certainly no intelligent planning behind it in Washington, not that we do anything resembling central planning. Of course, that's obvious with the tariffs and cancelling renewable energy projects and destroying the federal government from the inside-out.

I previously posted about the AI bubble actually being three bubbles, according to one prognosticator. Which means when those bubbles start bursting, to varying degrees, data center construction will collapse. Which means GDP is going to crater in an absolutely huge way.

Fun times ahead! Might want to pick up a couple of cases of beans. And, of course, a can opener.

https://fortune.com/2025/10/07/data-centers-gdp-growth-zero-first-half-2025-jason-furman-harvard-economist/

https://slashdot.org/story/25/10/07/2012240/without-data-centers-gdp-growth-was-01-in-the-first-half-of-2025-harvard-economist-says
thewayne: (Default)
Another old tab from May.

This is quite interesting. Researchers set up multiple LLMs and configured them to run a vending machine simulator, described as "Agents must balance inventories, place orders, set prices, and handle daily fees – tasks that are each simple but collectively, over long horizons." Basic business process.

The LLMs behaviors were, shall we say, interesting.

As the run went on over multiple simulated days, one decided it was the victim of cybercrime and 'reported' the event to the FBI (it had an email simulator but no external connection), another declared its quantum state as collapsed, yet another threatened suppliers with "ABSOLUTE FINAL ULTIMATE TOTAL NUCLEAR LEGAL INTERVENTION".

Basically it was a demonstration of how such large-language models are terrible for long-term runs and shows their ability to hallucinate and make poor decisions. I'll have some more posts on that soon, particularly concerning Canada and Australia.

The paper is quite interesting, detailing how some of the LLMs melt down and can't prioritize tasks. For example, a person knows that we must receive orders from suppliers before we can send someone out to refill a machine. The LLM might assume that on the date the order is promised, as soon as that date arrives the orders are suddenly there and the stocker can be immediately dispatched, even if there is no product or a shortage. Now the vending machine is understocked and the LLM doesn't understand why.

LLM no thinkie good.

The paper:
https://arxiv.org/html/2502.15840v1

The Slashdot article:
https://slashdot.org/story/25/05/31/2112240/failure-imminent-when-llms-in-a-long-running-vending-business-simulation-went-berserk
thewayne: (Default)
I'm just going to copy the Slashdot summary, then comment on it: Fast Company ran a contrarian take about AI from entrepreneur/thought leader Faisal Hoque, who argues there's three AI bubbles.

The first is a classic speculative bubble, with asset prices soaring above their fundamental values (like the 17th century's Dutch "tulip mania"). "The chances of this not being a bubble are between slim and none..."

Second, AI is also arguably in what we might call an infrastructure bubble, with huge amounts being invested in infrastructure without any certainty that it will be used at full capacity in the future. This happened multiple times in the later 1800s, as railroad investors built thousands of miles of unneeded track to serve future demand that never materialized. More recently, it happened in the late '90s with the rollout of huge amount of fiber optic cable in anticipation of internet traffic demand that didn't turn up until decades later. Companies are pouring billions into GPUs, power systems, and cooling infrastructure, betting that demand will eventually justify the capacity. McKinsey analysts talk of a $7 trillion "race to scale data centers" for AI, and just eight projects in 2025 already represent commitments of over $1 trillion in AI infrastructure investment. Will this be like the railroad booms and busts of the late 1800s? It is impossible to say with any kind of certainty, but it is not unreasonable to think so.

Third, AI is certainly in a hype bubble, which is where the promise claimed for a new technology exceeds reality, and the discussion around that technology becomes increasingly detached from likely future outcomes. Remember the hype around NFTs? That was a classic hype bubble. And AI has been in a similar moment for a while. All kinds of media — social, print, and web — are filled with AI-related content, while AI boosterism has been the mood music of the corporate world for the last few years. Meanwhile, a recent MIT study reported that 95% of AI pilot projects fail to generate any returns at all.

But the article ultimately argues there's lessons in the 1990s dotcom boom: that "a thing can be hyped beyond its actual capabilities while still being important... When valuations correct — and they will — the same pattern will emerge: companies that focus on solving real problems with available technology will extract value before, during, and after the crash." The winners will be companies with systematic approaches to extracting value — adopting mixed portfolios with different time horizons and risk levels, while recognizing organizational friction points for a purposeful (and holistic) integration.

"The louder the bubble talk, the more space opens for those willing to take a methodical approach to building value."


The first bubble is obvious. Huge amounts of money is being 'invested' in AI/LLMs and the returns have been dubious and amusing, and sometimes lethal. Children and teens taking their own lives, a formerly well-behaved autistic child becoming violent, etc. The valuation of Tesla going up while its sales sales plunge is always an amusing example. The infrastructure bubble is tragic: coal and offline nuclear power plants are being planned to power data centers exclusively for these things, and along with them are their water requirements. And that is a really big problem with increasing climate change. I read an article that I'll post if I can find it that said that each simple AI query is the equivalent of the use of a small bottle of water. The ecological cost is really quite, quite staggering. The eco cost of bitcoin and its kin is trivial compared to this.

The third bubble is interesting. They've demonstrated that LLMs can do some very cool things when tasked into specific purposes and trained in specific bodies of knowledge, like researching new antibiotics or metal alloys with new properties that are needed.

I think the thing that I'm the most curios about is when the corrections/collapses will start taking place. Considering the valuations involved, the financial quake will make the Dot Com crash look like the merest tremor.

The author, Faisal Hoque, is a lot more optimistic about AI than I. He compares its development to such as Amazon and Google during the Dot Com era of the 90s. They had very long-term development timelines ('Moon Shots') that they were quietly pursuing that achieved their long-term survival. And while not all current AI companies are going to achieve those and remain largely in their current form, some may. He talks about Pets.com burning through $300mil before collapsing, which we now see as a trivially small amount of money in today's tech market.

Curious times. We shall see how things shake out.

https://www.fastcompany.com/91400857/there-isnt-an-ai-bubble-there-are-three-ai-bu

https://slashdot.org/story/25/09/20/1847246/there-isnt-an-ai-bubble---there-are-three
thewayne: (Default)
To briefly recap, a group of authors sued the AI company Anthropic for pirating their books off the internet through illegal downloads and incorporating it into their AI data training sets, alleging piracy, copyright violation and theft. Which it clearly was. In an interesting twist, Anthropic then went out and bought quite literally tons of books, cut the spines off of them, scanned the pages, then trashed the then-scanned books, claiming the rights of first-ownership that they could do what they wanted to with the books.

But that was a bit of ex post facto reasoning: they'd already committed the crime of stealing the contents of the books, subsequently buying them after having already incorporated the contents into the datasets doesn't make it all better.

From the article: "In June, U.S. District Judge William Alsup ruled that Anthropic’s use of the books in training models was “exceedingly transformative,” one of the factors courts have used in determining whether the use of protected works without authorization was a legal “fair use.” His decision was the first major decision that weighed the fair use question in generative AI systems.

Yet Alsup also ruled that Anthropic had to face a trial on the question of whether it is liable for downloading millions of pirated books in digital form off the internet, something it had to do in order to train its models for its AI service Claude. The books were obtained from datasets Library Genesis and Pirate Library Mirror.

“That Anthropic later bought a copy of a book it earlier stole off the internet will not absolve it of liability for the theft but it may affect the extent of statutory damages,” the judge wrote.
(emphasis mine)

The piracy issue was a huge one. in court, Anthropic IT staff testified that they used bit torrent software to download vast troves of books at the direction of management. The problem is with bit torrent. Bit torrent uses "seeds". When you download a file, you are downloading small pieces of it from many clients and servers from around the world. And your computer becomes one such piece of this network and starts serving up pieces of the files that you've downloaded to people requesting those files.

As a general rule, companies don't go after people downloading pirated material if they're not downloading it 24/7/365. But they do go after people providing pirated material! And if you use bit torrent software to download pirated material, you're downloading AND uploading material that shouldn't be shared! Eventually they're going to notice you and their attorneys are going to dust off their giant mallets of loving correction.

I've used bit torrent software before. But what I use it for is downloading books that I've bought from Humble Bundle where I've got 20 large PDF books to download, it's the only practical way to do it even when I have a fairly fast fiberoptic internet connection. And I leave my torrent connection open so other people who've bought the bundle can benefit from my PC having those books on it.

I have no idea how many books Anthropic downloaded. It's quite possible that Anthropic has no absolute count as to how many books they downloaded. And that's probably why they agreed to this settlement. They wanted to avoid a damages trial which would dig into exactly how many books they had stolen.

And let's take that one step further. This would have branded them - in court! - as the world's largest piracy case. EVER. That's one thing that they definitely did not want to be branded with. A great big Scarlet P that they would wear forever. Much better to pay $1.5 billion and be rid of it.

Two additional things about this of interest. First, the settlement only covers their misdeeds through August 25. If they are found to have conducted any additional piracy after this date, then all the court processes could get reset and everything starts over again. Second, and this is the most significant part: "Anthropic also has agreed to destroy the datasets used in its models."

I have no idea what this fully means. Since they bought all these books and scanned them, they presumably have an even better dataset on standby once this pirated set is destroyed, so it shouldn't affect them much. Perhaps this is purely a symbolic victory, but it is an important one. We shall see.

https://deadline.com/2025/09/anthropic-ai-lawsuit-settlement-1-5-billion-1236509423/

https://yro.slashdot.org/story/25/09/05/1941245/anthropic-agrees-to-pay-record-15-billion-to-settle-authors-ai-lawsuit
thewayne: (Default)
*sigh*

Care to guess how it happened? The suggestions included "Tidewater Dreams" by Isabel Allende and "The Last Algorithm" by Andy Weir". The independent who put the list together used an AI and didn't check what it generated.

The Sun-Times went through some massive lay-offs recently as its finances are in not very good shape, and lost 20% of its readership. I'm sure this little reading list snafu will encourage people to reup their subscriptions. Or not.

https://arstechnica.com/ai/2025/05/chicago-sun-times-prints-summer-reading-list-full-of-fake-books/
thewayne: (Default)
Microsoft has cancelled or revised data center plans to the tune of $13,000,000,000 recently, projects that were mainly for AI centers. The reason? AI/LLM is not panning out as projected. As newer models are coming out, hallucination rates are rising rather than falling. This bodes ill.

In some cases lease options are being kept and the sites will continue being used as farmland until if/when MS decides to actually build the data centers.

Meta has recently likewise started cancelling data center plans.

Article may be paywalled:
https://www.bloomberg.com/news/articles/2025-03-26/microsoft-abandons-more-data-center-projects-td-cowen-says

The Slashdot summary:
"Microsoft has walked away from new data center projects in the US and Europe that would have amounted to a capacity of about 2 gigawatts of electricity, according to TD Cowen analysts, who attributed the pullback to an oversupply of the clusters of computers that power artificial intelligence. From a report:
The analysts, who rattled investors with a February note highlighting leases Microsoft had abandoned in the US, said the latest move also reflected the company's choice to forgo some new business from ChatGPT maker OpenAI, which it has backed with some $13 billion. Microsoft and the startup earlier this year said they had altered their multiyear agreement, letting OpenAI use cloud-computing services from other companies, provided Microsoft didn't want the business itself.

Microsoft's retrenchment in the last six months included lease cancellations and deferrals, the TD Cowen analysts said in their latest research note, dated Wednesday. Alphabet's Google had stepped in to grab some leases Microsoft abandoned in Europe, the analysts wrote, while Meta Platforms had scooped up some of the freed capacity in Europe."


https://slashdot.org/story/25/03/26/1832216/microsoft-abandons-data-center-projects-td-cowen-says


Meanwhile in China, two years ago a huge data center construction boom took place in an attempt to catch up in the AI/LLM race. And then the Chinese had a breakthrough and found a way around the GPU chip embargo and discovered that there wasn't nearly as much need for huge numbers of data centers and GPU farms.

And 80% of these data centers are sitting around unused!

From the article: “The growing pain China’s AI industry is going through is largely a result of inexperienced players—corporations and local governments—jumping on the hype train, building facilities that aren’t optimal for today’s need,” says Jimmy Goodrich, senior advisor for technology to the RAND Corporation.

The upshot is that projects are failing, energy is being wasted, and data centers have become “distressed assets” whose investors are keen to unload them at below-market rates. The situation may eventually prompt government intervention, he says: “The Chinese government is likely to step in, take over, and hand them off to more capable operators.”


Something on the order of over 500 were announced in 2023/2024, which means only 100 or so are in use?! The problem was that nobody knew what they were doing with AI, but by damn, we've got to get on that bandwagon!

https://www.technologyreview.com/2025/03/26/1113802/china-ai-data-centers-unused/
thewayne: (Default)
This is pretty amusing, funny, and ironic, for certain values of amusing, funny, and ironic.

The Chinese have released an open source LLM chatbot, it's called DeepSeek R1. It's available on GitHub, you can download it and play with it, tear apart the code, tweak it, etc. Completely free. It was built in TWO MONTHS for $6,000,000. And on lower grade GPUs because of export restrictions placed on the Chinese - can't have them getting top of the line GPUs now, can we? And they didn't scrape the internet, stealing trademarked and copyrighted information without permission to train it.

And it's the top downloaded app on the Apple app store.

The belief was that with denying the Chinese the top-end GPUs needed to do the crunching to construct large language models, that they had no hope of catching up with Team USA when it came to building AIs. Well, it seems that Team USA forgot the phrase 'Work smarter, not harder'. The Chinese applied a lot of smarts to work around the restrictions that were placed upon them and produced a much better product.

And stocks tanked because it's good software. From the article: "GPU maker NVIDIA fell 11%, Oracle dropped 8%, and Palantir was down 5% ... Stocks are adjusting to the revelation that China can build AI faster, cheaper, and just as good as America."

The comments in the article are amusing.

Nice slice of humble pie served up there.

https://gizmodo.com/chinese-ai-deepseek-deep-sixes-openai-on-the-app-store-stocks-tank-2000555171
thewayne: (Default)
The code was written by Joseph Weizenbaum, a German Jew whose family fled Nazi Germany for the USA and studied at Wayne State in Detroit. He wrote ELIZA in a programming language that he created called MAD-SLIP, Michigan Algorithm Decoder Symmetric List Processor, in only 420 lines of code! It was quickly translated into Lisp, a language well-regarded for AI work. His work developing MAD-SLIP earned him an associate professor slot at MIT where he ultimately wrote ELIZA, that post became a tenured professorship in four years. He also held academic appointments at Harvard, Stanford, the University of Bremen, and elsewhere. He passed away in 2008 and is buried in Germany.

From the article, "Experts thought the original 420-line ELIZA code was lost until 2021, when study co-author Jeff Shrager, a cognitive scientist at Stanford University, and Myles Crowley, an MIT archivist, found it among Weizenbaum's papers.

"I have a particular interest in how early AI pioneers thought," Shrager told Live Science in an email. "Having computer scientists' code is as close to having a record of their thoughts, and as ELIZA was — and remains, for better or for worse — a touchstone of early AI, I want to know what was in his mind." But why the team wanted to get ELIZA working is more complex, he said.


They go on to talk about building an emulator to simulate the computers from the 1960s to run the code properly, and discovering and deciding to keep in place a bug in the code.

Pretty cool stuff. And only 420 lines of code!

https://www.livescience.com/technology/eliza-the-worlds-1st-chatbot-was-just-resurrected-from-60-year-old-computer-code

https://slashdot.org/story/25/01/18/0544212/worlds-first-ai-chatbot-eliza-resurrected-after-60-years


Weisenbaum was an interesting person with some cool philosophies regarding computers and AI, of which he had some apprehensions. Two movies were made about him, he also published several books. His wikipedia page is worth a read, IMO.

https://en.wikipedia.org/wiki/Joseph_Weizenbaum
thewayne: (Default)
This is interesting.

The company who, up until 2012, published the book-form of the Encyclopedia Britannica, is now turning that huge trove of facts into an LLM engine with the goal of selling it as a service to the education market.

While this might seem as a bit of a snoozer, there's one very interesting aspect to this: AI hallucinations.

Most LLM models have hallucination problems, seemingly stemming from their snarfing up their training data from hoovering up the internet with all of its crappy and contradictory information. This is where Britannica shines: they paid a literal fortune over two centuries collecting vetted materials from recognized scholars using quality editors to compile it into a trusted source. Thus, the quality of their training model will be very, very high.

The question will be if their code that ingests this training model will still hallucinate. And we'll only see that with testing when it goes public and really gets pummeled. But I do like the idea: starting with a very high quality training set, I think it shows promise.

Though we still have the problem of AI systems consuming stupid godawful amounts of energy.

Britannica's encyclopedia is still available online, just not in a print edition.

https://gizmodo.com/encyclopedia-britannica-is-now-an-ai-company-2000542600
thewayne: (Default)
A tech worker in San Francisco needed surgery after an accident and faced a spate of claim denials, so she fought back, appealing all of them. And over 90% of the roughly 40 denials were approved on reconsideration. She even got a denial overturned on appeal for surgery for her pet dog!

So she started helping friends with the process.

Well, there's one thing about us programmers. When we start doing something repeatedly, we start thinking "How can we automate this and improve it?"

So she turned to large language models, commonly known as AI.

And it seems to be working pretty good!

Scan in your denial letter, and her system will generate several appeal letters for you to pick and choose from, you can also modify as you see fit.

A study showed that only about a tenth of one percent of rejected claims of participants in the ACA appeal. My doctor's office has handled this for me in the past, I've also joined in the fight a couple of times, and they've always been overturned thus far. It's definitely worth the fight, and this will make it a lot easier to do!

The actual web site is at: https://fighthealthinsurance.com/

She put a year of her time and $10,000 of her own money into this project! It's free for now, though she may charge for add-on services such as faxing an appeal directly to your insurance.

https://sfstandard.com/2024/08/23/holden-karau-fight-health-insurance-appeal-claims-denials/

https://science.slashdot.org/story/24/08/31/2131240/tech-worker-builds-free-ai-powered-tool-for-fighting-us-health-insurance-denials
thewayne: (Default)
First off, it has to be pointed out this is a specialized AI model designed for programmers, not a generalized model like ChatGPT et al.

IBM trained it specifically on open source libraries to which they explicitly had permission, basically bending over backwards to avoid any possible legal issues. And they now have a working model that they've released to the public! Granite was trained on 116 different programming languages and has from 3 to 34 billion tokens, presumably per language. I wonder if you can ask it to list all the languages it's trained in, I'll bet there's some pretty esoteric ones in there! I'd love it if it had MUMPS! (I once found a book on MUMPS programming at the Phoenix Public Library, I imagine it's been weeded by now)

Anyway, interesting article. It describes how it was trained, etc., but one of the more interesting bits was saying that in the rather short time since ChatGPT et al have appeared and everyone started creating their own LLMs, the cost for training up an LLM has dropped from millions of dollars to thousands! That's a pretty impressive scale drop.

https://www.zdnet.com/article/ibm-open-sources-its-granite-ai-models-and-they-mean-business/

https://www.zdnet.com/article/ibm-open-sources-its-granite-ai-models-and-they-mean-business/
thewayne: (Default)
I've written about using new imaging techniques plus computed tomography and AI has enabled the charcoal briquets that were formally scrolls at Vesuvius and Herculaneum to begin to be read. At Herculaneum, a library, of sorts, was discovered containing at least 600 intact scrolls. The University of Kentucky has developed a software system called Volume Cartography to help unroll these scrolls.

One such scroll describes Plato's last night and where he was buried! He was suffering from a high fever and was close to death, but still of somewhat sound mind. A young girl was brought in to play the flute for him, and he critiqued her lack of rhythm! I love it. 'I may be about to die, but your playing sucks! Work on it!'

As to his final resting place, "... the few surviving texts from that period indicate that the philosopher was buried somewhere in the garden of the Academy he founded in Athens. The garden was quite large, but archaeologists have now deciphered a charred ancient papyrus scroll recovered from the ruins of Herculaneum, indicating a more precise burial location: in a private area near a sacred shrine to the Muses..." There's one thing that I absolutely hate about this article - it doesn't say anything as to whether or not we know where the Academy and garden is/was located!

This is all part of the Vesuvius Challenge to read these scrolls, and it's making tremendous progress!

https://arstechnica.com/science/2024/04/deciphered-herculaneum-papyrus-reveals-precise-burial-place-of-plato/

https://www.theguardian.com/books/2024/apr/29/herculaneum-scroll-plato-final-hours-burial-site
thewayne: (Default)
So what?, many of you say.

He had a near fatal stroke eleven years ago and almost completely lost the ability to speak and sing! In that time, he's done some acoustic playing and was inducted into the CW Hall of Fame. He's not an invalid, no idea how much the stroke affected him otherwise.

Details have not been fully released, but through using AI technology, they've recreated his voice and he has a new song out.

Now obviously this brings up a host of issues. I am absolutely okay with this: Randy seems to be fully competent of mind and body, just not able to speak or sing. He has control of his music, and participated in this project to recreate his voice and get this song out. I think this is quite awesome! Not that I'm a CW fan or really care one way or another, but it's great for an artist to be able to express themselves creatively after nearly dying from a stroke and being deprived of their greatest instrument.

But you can see where this can be abused. As much as I'd love to hear new Freddie Mercury, or David Bowie, they're long gone and cannot participate in the process. No input from them. Theoretically their estate or IP managers could, and I would think that's wrong. I have the same problem with cinema recreations of James Dean or Humphrey Bogart. Digital de-aging of people, such as the ABBA Voyager holographic performance? Very interesting. And all four of them are alive to consent to it. Digital recreations of a performing Tupac? Very problematic for me.

One thing that we don't know about the Randy Travis thing which may come out eventually is who performed the vocals for the recording, which is being supposed that Randy's generated voice was then layed over.

https://www.rollingstone.com/music/music-country/randy-travis-releases-ai-song-1235014871/
thewayne: (Default)
The global power supply is feeling the pinch of AI as data centers are being built and more planned for companies getting in to the generative AI field. I have mentioned before that generative AI consumes more power than generating cryptocurrency, which is no slouch when it comes to consuming current: companies have repurposed retired coal plants to power crypto!

So now what, we're going to unretire closed nuke plants to power AI data centers?

Even now, AI mining operations are being closed to repurpose them for training the Large Language Models (the LLMs that are frequently referred to) for AI.

This is a big mess, and it's only going to get worse. The permitting and construction lead time for any energy source, be it natural gas, wind farms, whatever, is quite extensive. And there's probably a heck of a waiting list for the companies that build them. And the companies that want new data centers want the power for them NOW!NOW!NOW! Bit of a problem.

https://arstechnica.com/ai/2024/04/power-hungry-ai-is-putting-the-hurt-on-global-electricity-supply/
thewayne: (Default)
And what's even harderer? Using an AI coding assistant to write secure programs.

Many, MANY times that I've written about computer insecurity issues I've said explicitly that computer security is HARD. And here we have a prime example.

It turns out that using an AI to help you write a program produces LESS secure programs! But that's not the worse part: the program is more likely to believe that they are writing MORE SECURE CODE!

This is very bad. I've used AI for hints in writing code, looking for little obscure code references that I'm not familiar with. Quite useful. I haven't used it to write entire programs for me, I'm not sure that I could. However, there are people out there paying for subscriptions to ChatGPT 4 and other engines using them heavily, and that is worrisome.

https://arxiv.org/html/2211.03622v3

https://www.schneier.com/blog/archives/2024/01/code-written-with-ai-assistants-is-less-secure.html
thewayne: (Default)
*sigh* At least it's been a long time since the last one, but it's going to be a rough transition.

Microsoft is taking away the Ctrl key on the right.

In its place will be a key for its CoPilot AI Assistant.

Won't that be just dandy?

The last change was when MS added the Windows key to the Natural Keyboard back in '94. But MS really wants people to use its AI assistant, so what better way than to make a key dedicated to it where people regularly use a normal key?

Here's the kicker: it's possible that it may not be able to be reassigned!

I was reading an article on Dell's new XPS series that's going through a complete refresh for 2024. They all have the new CoPilot key - to the left of the left arrow key - and it is immutable. Cannot be changed. That's definitely going to force a lot of people to retrain muscle memory who are semi-touch typists.

Personally, if they'd tied it to a function key, or left the key reprogrammable - that'd be fine. But if it is indeed not reprogrammable, that's going to be quite an issue!

https://arstechnica.com/gadgets/2024/01/ai-comes-for-your-pcs-keyboard-as-microsoft-adds-dedicated-copilot-key/

November 2025

S M T W T F S
       1
23 4567 8
9 10 1112131415
16171819202122
23242526272829
30      

Syndicate

RSS Atom

Most Popular Tags

Style Credit

Expand Cut Tags

No cut tags
Page generated Nov. 13th, 2025 06:42 pm
Powered by Dreamwidth Studios