A new AI model developed by researchers at Stanford University demonstrates a significant leap in multimodal reasoning, capable of analyzing and drawing connections between text, images, and audio within a single framework. The system, named 'Cognitron', was trained on a vast, diverse dataset and shows improved performance on complex tasks like contextual question answering and …
A new AI model developed by researchers at Stanford University demonstrates a significant leap in multimodal reasoning, capable of analyzing and drawing connections between text, images, and audio within a single framework. The system, named ‘Cognitron’, was trained on a vast, diverse dataset and shows improved performance on complex tasks like contextual question answering and cross-modal inference compared to previous models. While promising for applications in education and complex data analysis, the researchers note limitations in real-time processing and the need for further testing to mitigate potential biases. The full details of the architecture and training methodology are available in the published paper. Read the full article at https://technologyreview.com/2024/05/15/cognitron-ai-multimodal-model.
Join the Club
Like this story? You’ll love our Bi-Weekly Newsletter



