IBM open-sources its Granite AI model!
May. 15th, 2024 10:48 pmFirst off, it has to be pointed out this is a specialized AI model designed for programmers, not a generalized model like ChatGPT et al.
IBM trained it specifically on open source libraries to which they explicitly had permission, basically bending over backwards to avoid any possible legal issues. And they now have a working model that they've released to the public! Granite was trained on 116 different programming languages and has from 3 to 34 billion tokens, presumably per language. I wonder if you can ask it to list all the languages it's trained in, I'll bet there's some pretty esoteric ones in there! I'd love it if it had MUMPS! (I once found a book on MUMPS programming at the Phoenix Public Library, I imagine it's been weeded by now)
Anyway, interesting article. It describes how it was trained, etc., but one of the more interesting bits was saying that in the rather short time since ChatGPT et al have appeared and everyone started creating their own LLMs, the cost for training up an LLM has dropped from millions of dollars to thousands! That's a pretty impressive scale drop.
https://www.zdnet.com/article/ibm-open-sources-its-granite-ai-models-and-they-mean-business/
https://www.zdnet.com/article/ibm-open-sources-its-granite-ai-models-and-they-mean-business/
IBM trained it specifically on open source libraries to which they explicitly had permission, basically bending over backwards to avoid any possible legal issues. And they now have a working model that they've released to the public! Granite was trained on 116 different programming languages and has from 3 to 34 billion tokens, presumably per language. I wonder if you can ask it to list all the languages it's trained in, I'll bet there's some pretty esoteric ones in there! I'd love it if it had MUMPS! (I once found a book on MUMPS programming at the Phoenix Public Library, I imagine it's been weeded by now)
Anyway, interesting article. It describes how it was trained, etc., but one of the more interesting bits was saying that in the rather short time since ChatGPT et al have appeared and everyone started creating their own LLMs, the cost for training up an LLM has dropped from millions of dollars to thousands! That's a pretty impressive scale drop.
https://www.zdnet.com/article/ibm-open-sources-its-granite-ai-models-and-they-mean-business/
https://www.zdnet.com/article/ibm-open-sources-its-granite-ai-models-and-they-mean-business/
no subject
Date: 2024-05-17 07:34 pm (UTC)I don't know. I did some experimenting with ChatGPT trying to use it as a code generator. The results were not terribly impressive. This was the free, 3.5, model, not the paid 4.0 model. There are a lot of programmers who swear by (and probably sometimes, at) the paid models and swear that it really helps their productivity. Given the reports that code audits claim that LLM-generated code generates more security holes, one has to wonder. I suppose it is highly variable, depending on the skill of the programmer and the type of programming that they're doing, and how much purely-generated code they're inserting into their project vs small snippets that are being integrated.