Like human brains, large language models reason about diverse data in a general way
A new study shows LLMs represent different data types based on their underlying meaning and reason about data in their dominant language.
A new study shows LLMs represent different data types based on their underlying meaning and reason about data in their dominant language.
MIT researchers developed a new approach for assessing predictions with a spatial dimension, like forecasting weather or mapping air pollution.
Assistant Professor Sara Beery is using automation to improve monitoring of migrating salmon in the Pacific Northwest.
By automatically generating code that leverages two types of data redundancy, the system saves bandwidth, memory, and computation.
Providing electricity to power-hungry data centers is stressing grids, raising prices for consumers, and slowing the transition to clean energy.
Rapid development and deployment of powerful generative AI models comes with environmental consequences, including increased electricity demand and water consumption.
Assistant Professor Manish Raghavan wants computational techniques to help solve societal problems.
Biodiversity researchers tested vision systems on how well they could retrieve relevant nature images. More advanced models performed well on simple queries but struggled with more research-specific prompts.
How a love for math and access to MIT Open Learning’s online learning resources helped a Sudanese learner pursue a career in data science.
A new technique identifies and removes the training examples that contribute most to a machine-learning model’s failures.
Research from the MIT Center for Constructive Communication finds this effect occurs even when reward models are trained on factual data.
Using LLMs to convert machine-learning explanations into readable narratives could help users make better decisions about when to trust a model.
Researchers develop “ContextCite,” an innovative method to track AI’s source attribution and detect potential misinformation.
MIT engineers developed the largest open-source dataset of car designs, including their aerodynamics, that could speed design of eco-friendly cars and electric vehicles.
This new device uses light to perform the key operations of a deep neural network on a chip, opening the door to high-speed processors that can learn in real-time.