IBM open-sources its Granite AI model!
May. 15th, 2024 10:48 pmFirst off, it has to be pointed out this is a specialized AI model designed for programmers, not a generalized model like ChatGPT et al.
IBM trained it specifically on open source libraries to which they explicitly had permission, basically bending over backwards to avoid any possible legal issues. And they now have a working model that they've released to the public! Granite was trained on 116 different programming languages and has from 3 to 34 billion tokens, presumably per language. I wonder if you can ask it to list all the languages it's trained in, I'll bet there's some pretty esoteric ones in there! I'd love it if it had MUMPS! (I once found a book on MUMPS programming at the Phoenix Public Library, I imagine it's been weeded by now)
Anyway, interesting article. It describes how it was trained, etc., but one of the more interesting bits was saying that in the rather short time since ChatGPT et al have appeared and everyone started creating their own LLMs, the cost for training up an LLM has dropped from millions of dollars to thousands! That's a pretty impressive scale drop.
https://www.zdnet.com/article/ibm-open-sources-its-granite-ai-models-and-they-mean-business/
https://www.zdnet.com/article/ibm-open-sources-its-granite-ai-models-and-they-mean-business/
IBM trained it specifically on open source libraries to which they explicitly had permission, basically bending over backwards to avoid any possible legal issues. And they now have a working model that they've released to the public! Granite was trained on 116 different programming languages and has from 3 to 34 billion tokens, presumably per language. I wonder if you can ask it to list all the languages it's trained in, I'll bet there's some pretty esoteric ones in there! I'd love it if it had MUMPS! (I once found a book on MUMPS programming at the Phoenix Public Library, I imagine it's been weeded by now)
Anyway, interesting article. It describes how it was trained, etc., but one of the more interesting bits was saying that in the rather short time since ChatGPT et al have appeared and everyone started creating their own LLMs, the cost for training up an LLM has dropped from millions of dollars to thousands! That's a pretty impressive scale drop.
https://www.zdnet.com/article/ibm-open-sources-its-granite-ai-models-and-they-mean-business/
https://www.zdnet.com/article/ibm-open-sources-its-granite-ai-models-and-they-mean-business/
no subject
Date: 2024-05-17 05:41 am (UTC)Don't think, let a computer do it for you?
Or am I just old? LOL.........
Hugs, Jon
no subject
Date: 2024-05-17 07:58 pm (UTC)THEN I have to make the reporting system! More screens to program, I have to relate the entered data to produce the reports, maybe do some graphs, etc.
Now, an LLM can help with a lot of that. But you really want a skilled programmer behind the keyboard who has institutional knowledge to recognize that C is missing, or that maybe there's a legal requirement that certain data has to be wiped after X years, who knows. There's all sorts of weird requirements that an AI may be ignorant of.
To give an example, I queried both ChatGPT 3.5 and 4 about the song Smoke on the Water about the fire at the Montreux Jazz Festival where "some stupid" fired a flare gun during a Frank Zappa and The Mothers Of Invention performance inside the casino hall and "burned the place to the ground". I asked both engines "The song Smoke on the Water by Deep Purple was based on a real incident. Did they ever catch or identify the person who fired the flare gun?" Both said no. If you go to Wikipedia and open the entry for the song Smoke on the Water, you'll see he was identified by Swiss authorities as Zdeněk Špička, a Czech national who fled the country and never faced trial.
LLMs can be useful, but they have to be watched by a knowledgeable eye.
no subject
Date: 2024-05-19 07:00 am (UTC)Must be true then because everything you read on the internet is true. Isn't it?
LOL.............