AI Models tagged posts

Leaner Large Language Models could enable Efficient Local Use on Phones and Laptops

Large language models (LLMs) are increasingly automating tasks like translation, text classification and customer service. But tapping into an LLM’s power typically requires users to send their requests to a centralized server—a process that’s expensive, energy-intensive and often slow.

Now, researchers have introduced a technique for compressing an LLM’s reams of data, which could increase privacy, save energy and lower costs. Their findings are published on the arXiv preprint server.

The new algorithm, developed by engineers at Princeton and Stanford Engineering, works by trimming redundancies and reducing the precision of an LLM’s layers of information...

Read More

When Lab-Trained AI Meets the Real World, ‘Mistakes can Happen’

AI and pathology
When humans examine tissue on slides, they can only look at a limited field within the microscope, then move to a new field and so on.

Tissue contamination distracts AI models from making accurate real-world diagnoses. Human pathologists are extensively trained to detect when tissue samples from one patient mistakenly end up on another patient’s microscope slides (a problem known as tissue contamination). But such contamination can easily confuse artificial intelligence (AI) models, which are often trained in pristine, simulated environments, reports a new Northwestern Medicine study.

“We train AIs to tell ‘A’ versus ‘B’ in a very clean, artificial environment, but, in real life, the AI will see a variety of materials that it hasn’t trained on...

Read More

AI Models can now Continually Learn from New Data on Intelligent Edge Devices like Smartphones and Sensors

Artistic collage shows a large teal hand, pointing towards us as if typing on a smart phone, with a lens flare on the tip of the finger. The background is a surreal, AI-generated blend of smart phones, keyboards, teal screens.
Caption:A machine-learning model on an intelligent edge device allows it to adapt to new data and make better predictions. For instance, training a model on a smart keyboard could enable the keyboard to continually learn from the user’s writing.
Credits:Image: Digital collage by Jose-Luis Olivares, MIT, using stock images and images derived from MidJourney AI.

A new technique enables on-device training of machine learning models on edge devices like microcontrollers, which have very limited memory. This could allow edge devices to continually learn from new data, eliminating data privacy issues, while enabling user customization.

Microcontrollers, miniature computers that can run simple commands, are the basis for billions of connected devices, from internet-of-things (IoT) devices...

Read More