Meta FAIR released several new research assets, including its SEGMENT Anything Model 2.1 (SAM 2.1) and research to improve AMI. They also introduced their speech + text-generative language model, called SPIRIT-LM. Anthropic has launched its AI model, the Claude 3.5 Sonnet, to compete with Meta’s Spirit LM and Google’s NotebookLM. Meta and Anthropic have published research on reinforcement learning from AI feedback. Meta’s new model is powerful and performs better than models that rely on human-labelled data. Meanwhile, Moshi by Kyutai and EVI 2 by Hume AI bring real-time, human-like interactions. Amazon’s Alexa is collaborating with Anthropic as its advanced conversational abilities grow. Spirit LM integrates text and speech capabilities, directly competing with OpenAI’s GPT-4o. While Claude 3.5 Sonnet excels in complex reasoning tasks and natural language processing.
Meta released various research assets to achieve AMI while promoting open science and reproducibility, including updated SEGMENT Anything Model 2.1 & research on improving language models, model training, post-quantum cryptography & inorganic materials discovery.
Meta’s latest release includes the updated Segment Anything Model 2.1 (SAM 2.1) for image and video segmentation and research on improving large language models, post-quantum cryptography, model training, and inorganic materials discovery.
Anthropic has launched Claude 3.5 Sonnet, an AI model designed to compete with other generative AI systems like Google’s NotebookLM and Meta’s Spirit LM.
Meta’s new model validates other AI models’ works and generates preference data to train reward models without relying on human annotations.
Spirit LM integrates text and speech capabilities, directly competing with OpenAI’s GPT-4o and uses phonetic, pitch, and tone tokens to improve the expressiveness of its speech outputs.
Claude 3.5 Sonnet, developed by Anthropic, excels in complex reasoning tasks and natural language processing and operates at twice the speed of its predecessor, Claude 3 Opus, enhancing its efficiency for real-time applications and multi-step workflows.
Kyutai released Moshi, a real-time native multimodal foundational AI model capable of conversing with humans in real-time, much like what OpenAI’s advanced model was intended to do.
Hume AI introduced EVI 2, a new foundational voice-to-voice AI model that promises to enhance human-like interactions.
Amazon Alexa is partnering with Anthropic to improve its conversational abilities, making interactions more natural and human-like.