Stability AI releases ChatGPT-like language models
Stability AI, the startup behind the generative AI art tool Stable Diffusion, today open-sourced a suite of text-generating AI models intended to go head to head with systems like OpenAI’s GPT-4.
Called StableLM and available in “alpha” on GitHub and Hugging Spaces, a platform for hosting AI models and code, Stability AI says that the models can generate both code and text and “demonstrate how small and efficient models can deliver high performance with appropriate training.”
“Language models will form the backbone of our digital economy, and we want everyone to have a voice in their design,” the Stability AI team wrote in a blog post on the company’s site.
The models were trained on a data set called The Pile, a mix of internet-scraped text samples from websites including PubMed, StackExchange and Wikipedia. But Stability AI claims it created a custom training set that expands the size of the standard Pile by 3x.
Stability AI didn’t say in the blog post whether the StableLM models suffer from the same limitations as others, namely a tendency to generate toxic responses to certain prompts and hallucinate (i.e. make up) facts. But given that The Pile contains profane, lewd and otherwise fairly abrasive language, it wouldn’t be surprising if that were the case.
“As is typical for any pretrained large language model without additional fine-tuning and reinforcement learning, the responses a user gets might be of varying quality and might potentially include offensive language and views,” Stability AI wrote in the repo for StableLM. “This is expected to be improved with scale, better data, community feedback and optimization.”
Still, the StableLM models seem fairly capable in terms of what they can accomplish — particularly the fine-tuned versions included in the alpha release. Tuned using a Stanford-developed technique called Alpaca on open source data sets, including from AI startup Anthropic, the fine-tuned StableLM models behave like ChatGPT, responding to instructions (sometimes with humor) like “write a cover letter for a software developer” and “write lyrics for an epic rap battle song.”
Some researchers have criticized the release of open source models along the lines of StableLM in the past, arguing that they could be used for unsavory purposes like creating phishing emails or aiding malware attacks. But Stability AI argues that open-sourcing is in fact the right approach, in fact.
“We open-source our models to promote transparency and foster trust. Researchers can ‘look under the hood’ to verify performance, work on interpretability techniques, identify potential risks and help develop safeguards,” Stability AI wrote in the blog post. “Open, fine-grained access to our models allows the broad research and academic community to develop interpretability and safety techniques beyond what is possible with closed models.”
Of course, Stability AI hasn’t shied away from controversy historically.
The company is in the crosshairs of legal cases that allege that it infringed on the rights of millions of artists by developing AI art tools using web-scraped, copyrighted images. And a few communities around the web have tapped Stability’s tools to generate pornographic celebrity deepfakes and graphic depictions of violence.
Moreover, despite the philanthropic tone of its blog post, Stability AI is also under pressure to monetize its sprawling efforts — which run the gamut from art and animation to biomed and generative audio. Stability AI CEO Emad Mostaque has hinted at plans to IPO, but Semafor recently reported that Stability AI — which raised over $100 million in venture capital last October at a reported valuation of more than $1 billion — “is burning through cash and has been slow to generate revenue.”