Menu
Join the Club

Your Bi-Weekly Dose Of Everything Optimism

News Summary

A new AI model developed by researchers at Stanford University demonstrates a significant leap in multimodal reasoning, capable of analyzing and drawing connections between text, images, and audio within a single framework. The system, named 'Cognitron', was trained on a vast, novel dataset combining these three modalities and shows improved performance on complex tasks like …

A new AI model developed by researchers at Stanford University demonstrates a significant leap in multimodal reasoning, capable of analyzing and drawing connections between text, images, and audio within a single framework. The system, named ‘Cognitron’, was trained on a vast, novel dataset combining these three modalities and shows improved performance on complex tasks like contextual question answering and cross-modal inference compared to existing models. Experts note the research addresses a key challenge in AI: creating systems that can understand the world in a more integrated, human-like way rather than processing information in isolated silos. The team has published its findings and made the model architecture publicly available for further research, though they caution that real-world applications requiring robust, real-time multimodal analysis are still a considerable distance away. Read the full article at https://technologyreview.com/2024/05/15/cognitron-ai-multimodal-model.

Join the Club

Like this story? You’ll love our Bi-Weekly Newsletter

Technology Review

Technology Review

Comments

Leave a Reply

Your email address will not be published. Required fields are marked *

Ask Richard AI Avatar