Jim Griffin
August 1, 2024
Last week Meta launched its newest family of models, Llama 3.1, including a new benchmark – an open-source foundation model with 405 billion parameters. With this, Zuckerberg predicted that Meta AI will surpass OpenAI’s 200 million monthly active users by the end of this year. Hubris aside, this video looks […]
Jim Griffin
July 17, 2024
A top researcher at Google DeepMind just released an important paper, “Mixture of a Million Experts.” As the paper’s title announces, it describes an approach that resulted in the first-known Transformer model with more than a million experts. For context, the number of experts currently seen in smaller models varies […]
Jim Griffin
June 26, 2024
There’s a big breakthrough that just came out for handling large language models on smartphones. It’s called PowerInfer-2 and what it does is look at every option for a processing an LLM on a particular smartphone, and picks the fastest way for that particular LLM on that particular device. For […]
Jim Griffin
June 20, 2024
Last week, NVIDIA announced Nemotron-4, which consists of three models: Base, Instruct and Reward. These three models work together within the NeMo framework to enable the creation and fine-tuning of new large language models. At 340 billion parameters, this new entrant far bigger than any other open source model, but […]