SLM tagged posts

Shrinking AI for Personal Devices: An efficient small language model that could perform better on smartphones

An Android demo of PhoneLM capability. (Left) Chatting; (Right) Device control through intent invocation. Demo and code are available at https://github.com/UbiquitousLearning/mllm. Credit: Yi et al.

Large language models (LLMs), such as Open AI’s renowned conversational platform ChatGPT, have recently become increasingly widespread, with many internet users relying on them to find information quickly and produce texts for various purposes. Yet most of these models perform significantly better on computers, due to the high computational demands associated with their size and data processing capabilities.

To tackle this challenge, computer scientists have also been developing small language models (SLMs), which have a similar architecture but are smaller...

Read More

Adobe announces development of SLM that can Run Locally on a Phone with No Cloud Connection

app
Credit: Pixabay/CC0 Public Domain

A small team of AI researchers at Adobe Inc., working with a colleague from Auburn University and another from Georgia Tech, has developed a small language model (SLM) that they claim can be run locally on a smart phone with no access to the cloud. The group has written a paper describing their new app, which they call SlimLM, and have posted it to the arXiv preprint server.

As LLM technology continues to mature, researchers across the globe continue to find new ways to improve it. In this new effort, the research team has found a way to cut the cord for a specific type of AI application—processing documents locally.

As LLMs such as ChatGPT become more popular, users have become more worried about privacy...

Read More

Microsoft’s Small Language Model Outperforms Larger Models on Standardized Math tests

Grade School Math
Credit: Deepak Gautam from Pexels

A small team of AI researchers at Microsoft reports that the company’s Orca-Math small language model outperforms other, larger models on standardized math tests. The group has published a paper on the arXiv preprint server describing their testing of Orca-Math on the Grade School Math 8K (GSM8K) benchmark and how it fared compared to well-known LLMs.

Many popular LLMs such as ChatGPT are known for their impressive conversational skills—less well known is that most of them can also solve math word problems. AI researchers have tested their abilities at such tasks by pitting them against the GSM8K, a dataset of 8,500 grade-school math word problems that require multistep reasoning to solve, along with their correct answers.

In this new study, th...

Read More