Skip to content ↓

Topic

Computer science and technology

Download RSS feed: News Articles / In the Media / Audio

Displaying 181 - 195 of 1168 news clips related to this topic.
Show:

Mashable

Mashable reporter Adele Walton spotlights Joy Buolamwini PhD '22 and her work in uncovering racial bias in digital technology. “Buolamwini created what she called the Aspire Mirror, which used face-tracking software to register the movements of the user and overlay them onto an aspirational figure,” explains Walton. “When she realised the facial recognition wouldn’t detect her until she was holding a white mask over her face, she was confronted face on with what she termed the ‘coded gaze.’ She soon founded the Algorithmic Justice League, which exists to prevent AI harms and increase accountability.”

Fast Company

Writing for Fast Company, Senior Lecturer Guadalupe Hayes-Mota '08, SM '16, MBA '16 shares methods to address the influence of AI in healthcare. “Despite these advances [of AI in healthcare], the full spectrum of AI’s potential remains largely untapped,” explains Hayes-Mota. “Systemic hurdles such as data privacy concerns, the absence of standardized data protocols, regulatory complexities, and ethical dilemmas are compounded by an inherent resistance to change within the healthcare profession. These barriers underscore the urgent need for transformative action from all stakeholders to fully harness AI’s capabilities.”

Fast Company

A new study conducted by researchers at MIT and elsewhere has found large language models (LLMs) can be used to predict the future as well as humans can, reports Chris Stokel-Walker for Fast Company. “Accurate forecasting of future events is very important to many aspects of human economic activity, especially within white collar occupations, such as those of law, business and policy,” says postdoctoral fellow Peter S. Park.

The Economist

Prof. Daniela Rus, director of CSAIL, speaks with The Economist’s Babbage podcast about the history and future of artificial neural networks and their role in large language models. “The early artificial neuron was a very simple mathematical model,” says Rus. “The computation was discrete and very simple, essentially a step function. You’re either above or below a value.”  

Associated Press

Prof. Philip Isola and Prof. Daniela Rus, director of CSAIL, speak with Associated Press reporter Matt O’Brien about AI generated images and videos. Rus says the computing resources required for AI video generation are “significantly higher than for still image generation” because “it involves processing and generating multiple frames for each second of video.”

The Boston Globe

Prof. Daniela Rus, director of CSAIL, speaks with Boston Globe reporter Evan Sellinger about her new book, “The Heart and the Chip: Our Bright Future With Robots,” in which she makes the case that in the future robots and humans will be able to team up to create a better world. “I want to highlight that machines don’t have to compete with humans, because we each have different strengths. Humans have wisdom. Machines have speed, can process large numbers, and can do many dull, dirty, and dangerous tasks,” Rus explains. “I see robots as helpers for our jobs. They’ll take on the routine, repetitive tasks, ensuring human workers focus on more complex and meaningful work.”

CNBC

Prof. Stuart Madnick speaks with CNBC reporter Kevin Williams about how the rise of generative AI technologies could lead to cyberattacks on physical infrastructure. “If you cause a power plant to stop from a typical cyberattack, it will be back up and online pretty quickly,” Madnick explains, “but if hackers cause it to explode or burn down, you are not back online a day or two later; it will be weeks and months because a lot of the parts in these specialized systems are custom made.”

Tech Briefs

Prof. Skylar Tibbits speaks with Tech Briefs reporter Andrew Corselli about his team’s work developing a new “additive manufacturing technique that can print rapidly with liquid metal, producing large-scale parts like table legs and chair frames in a matter of minutes.” Of his advice for engineers aiming to bring their ideas to fruition, Tibbits emphasizes: “Work hard, fail a lot, keep trying, don’t give up, and have amazing people around you. We're a research lab, so our whole goal is to go from impossible to possible. So, we're allowed to fail; we're not limited by profitability or customer demand or economy.”

Fast Company

Prof. Charles Stewart III and Ben Adida PhD ’06 speak with Fast Company reporter Spenser Mestel about how to restore the public’s faith in voting technology. Adida discusses his work launching VotingWorks, a non-profit focused on building voting machines. VotingWorks is “unique among the legacy voting technology vendors," writes Mestel. “The group has disclosed everything, from its donors to the prices of its machines.”

Politico

Researchers at MIT and elsewhere have developed a machine-learning model that can identify which drugs should not be taken together, reports Politico. “The researchers built a model to measure how intestinal tissue absorbed certain commonly used drugs,” they write. “They then trained a machine-learning algorithm based on their new data and existing drug databases, teaching the new algorithm to predict which drugs would interact with which transporter proteins.”

The Daily Beast

MIT researchers have developed a new technique “that could allow most large language models (LLMs) like ChatGPT to retain memory and boost performance,” reports Tony Ho Tran for the Daily Beast. “The process is called StreamingLLM and it allows for chatbots to perform optimally even after a conversation goes on for more than 4 million words,” explains Tran.

VOA News

Prof. David Rand speaks with VOA News about the potential impact of adding watermarks to AI generated materials. “My concern is if you label as AI-generated, everything that’s AI-generated regardless of whether it’s misleading or not, people essentially are going to stop really paying attention to it,” says Rand.

Politico

Researchers at MIT and elsewhere have found that while AI systems could help doctors come to the right diagnosis more often, the diagnostic gains aren’t always distributed evenly, with more improvements tied to patients with lighter skin, report Daniel Payne, Erin Schumaker, and Ruth Reader for Politico. “AI could be a powerful tool to improve care and potentially offer providers a check on their blindspots," they write. "But that doesn’t mean AI will reduce bias. In fact, the study suggests, AI could cause greater disparities in care.”

The Boston Globe

A more than $40 million investment to add advanced nano-fabrication equipment and capabilities to MIT.nano will significantly expand the center’s nanofabrication capabilities, reports Jon Chesto for The Boston Globe. The new equipment, which will also be available to scientists outside MIT, will allow “startups and students access to wafer-making equipment used by larger companies. These tools will allow its researchers to make prototypes of an array of microelectronic devices.”