How Good Are the Latest Open LLMs? And Is DPO Better Than PPO?
What a month! We had four major open LLM releases: Mixtral, Meta AI’s Llama 3, Microsoft’s Phi-3, and Apple’s OpenELM. In my new article, I review and…
What a month! We had four major open LLM releases: Mixtral, Meta AI’s Llama 3, Microsoft’s Phi-3, and Apple’s OpenELM. In my new article, I review and…
What are the different ways to use and finetune pretrained large language models (LLMs)? The three most common ways to use and finetune pretrained LLMs…
It’s another month in AI research, and it’s hard to pick favorites. This month, I am going over a paper that discusses strategies for the continued…
Once again, this has been an exciting month in AI research. This month, I’m covering two new openly available LLMs, insights into small finetuned LLMs, and…
Have you ever trained a model you thought was good, but then it failed miserably when applied to real world data? If so, you’re in good company.
Low-rank adaptation (LoRA) is a machine learning technique that modifies a pretrained model (for example, an LLM or vision transformer) to better suit a…
Posted by Dustin Zelle – Software Engineer, Research and Arno Eigenwillig – Software Engineer, CoreML
This article is also shared on the Google Research Blog
Objects and their relationships are ubiquitous in the world around us, and rel…
On fish counting – a complex sociotechnical problem in a field that is going through the process of digital transformation.
Posted by Sharbani Roy – Senior Director, Product Management, Google
We’re back with the third annual Women in Machine Learning Symposium on December 7, 2023! Join us virtually from 9:30 am to 1:00 pm PT for an immersive and insig…
Posted by Surya Kanoria, Joseph Cauteruccio, Federico Tomasi, Kamil Ciosek, Matteo Rinaldi, and Zhenwen Dai – Spotify
Introduction
Many of our music recommendation problems involve providing users with ordered sets of items that satisfy users’…