

This morning I ran a GPT-3 class language model on my own personal laptop for the first time!ĪI stuff was weird already. That Stable Diffusion moment is happening again right now, for large language models-the technology behind ChatGPT itself. It feels to me like that Stable Diffusion moment back in August kick-started the entire new wave of interest in generative AI-which was then pushed into over-drive by the release of ChatGPT at the end of November. Here's what he wrote in a post on his blog: Independent AI researcher Simon Willison has compared this situation to the release of Stable Diffusion, an open source image synthesis model that launched last August. Since then, there has been an explosion of development surrounding LLaMA. Meta's restrictions on LLaMA didn't last long, because on March 2, someone leaked the LLaMA weights on BitTorrent. There was just one problem-Meta released the LLaMA code open source, but it held back the "weights" (the trained "knowledge" stored in a neural network) for qualified researchers only. LLaMA made a heady claim: that its smaller-sized models could match OpenAI's GPT-3, the foundational model that powers ChatGPT, in the quality and speed of its output. Other open source alternatives could not boast GPT-3-level performance on readily available consumer-level hardware.Įnter LLaMA, an LLM available in parameter sizes ranging from 7B to 65B (that's "B" as in "billion parameters," which are floating point numbers stored in matrices that represent what the model "knows"). Open source solutions do exist (such as GPT-J), but they require a lot of GPU RAM and storage space.


Thus began the dream-in some quarters-of an open source large language model (LLM) that anyone could run locally without censorship and without paying API fees to OpenAI. Since ChatGPT launched, some people have been frustrated by the AI model's built-in limits that prevent it from discussing topics that OpenAI has deemed sensitive. (At least not today-as in literally today, March 13, 2023.) But what will arrive next week, no one knows. If this keeps up, we may be looking at a pocket-sized ChatGPT competitor before we know it.īut let's back up a minute, because we're not quite there yet. Further Reading Meta unveils a new large language model that can run on a single GPU
