By using this site, you agree to the Privacy Policy and Terms of Use.
Accept
TechgoonduTechgoonduTechgoondu
  • Audio-visual
  • Enterprise
    • Software
    • Cybersecurity
  • Gaming
  • Imaging
  • Internet
  • Media
  • Mobile
    • Cellphones
    • Tablets
  • PC
  • Telecom
Search
© 2023 Goondu Media Pte Ltd. All Rights Reserved.
Reading: Bigger always better? Not true for AI language models
Share
Font ResizerAa
TechgoonduTechgoondu
Font ResizerAa
  • Audio-visual
  • Enterprise
  • Gaming
  • Imaging
  • Internet
  • Media
  • Mobile
  • PC
  • Telecom
Search
  • Audio-visual
  • Enterprise
    • Software
    • Cybersecurity
  • Gaming
  • Imaging
  • Internet
  • Media
  • Mobile
    • Cellphones
    • Tablets
  • PC
  • Telecom
Follow US
© 2023 Goondu Media Pte Ltd. All Rights Reserved.
Techgoondu > Blog > Enterprise > Bigger always better? Not true for AI language models
EnterpriseSoftware

Bigger always better? Not true for AI language models

Ai Lei Tao
Last updated: July 23, 2024 at 2:36 PM
Ai Lei Tao
Published: July 22, 2024
4 Min Read
SHARE
PHOTO: Joshua Hoehne on Unsplash

Large language models (LLMs) may impress with their ability to perform a wide range of tasks with great efficiency. However, small language models (SLMs) are emerging as a significant and practical alternative, say industry experts.

SLMs are optimised for lower-resource environments, they say, thus requiring less computational power and fewer resources compared to LLMs, which boast billions of parameters. This makes SLMs ideal for delivering high performance for particular applications without the need for extensive infrastructure.

According to an article in the scientific journal Nature, LLMs require volumes of expensive computing resources from graphics processing units (GPUs). The article also noted that a generative AI-driven search uses four to five times the energy compared to a conventional web search.

Kalyan Madala, IBM Technology’s APAC Software pre-sales engineering leader explained that SLMs require far less computational power and resources as they can operate locally on everyday hardware like laptops, mobile phones and edge devices. 

This on-device processing, especially in edge computing and Internet of Things (IoT) applications, enhances security as it reduces data exposure and the risk of unauthorised access.

Advantages of SLMs

As SLMs have a simpler architecture and use smaller datasets for training, it makes them more explainable than LLMs – allowing humans to better understand and trust the output generated. 

“The models are simpler and more interpretable, boosting transparency and aiding adoption. This AI explanability is essential for building trust in sectors like law, finance, and healthcare,” said Madala. 

“Without the necessity for costly, specialised infrastructure, SLMs give smaller businesses and startups a cost-effective solution that doesn’t sacrifice effectiveness or versatility, which is crucial for real-time applications and scenarios where latency is a concern,” he added. 

In Asia-Pacific, SLM adoption is driven by resource constraints, as many emerging markets in the region have limited access to high-end computational resources, he suggested. 

SLMs can be fine-tuned to support local languages and dialects, suitable for a linguistically diverse region like the Asia-Pacific. SLMs can also support diverse applications, from enhancing customer service in retail to automating tasks in manufacturing.

“For users, what’s important is not just the size of the model but having the choice to customise and tailor their foundation models for their evolving use cases,” said Madala. 

“Organisations should also have the flexibility to deploy the model in the infrastructure of their choice, depending on the use case and operational considerations,” he added.

“AI guardrails and continuous monitoring ensure that model deployments are secure and reliable as organisations scale up generative AI applications,” he stressed.

The emergence of niftier models

IBM recently launched the open-source Mistral AI Model on the watsonx platform – a compact LLM that is touted to require less resources to run, but is as effective and has better performance compared to traditional LLMs. 

Other vendors have announced smaller AI models this year, including Microsoft and Google. For example, Microsoft has revealed Phi-3, its series of OpenAI SLMs that are smaller and less compute-intensive for generative AI solutions. 

In February, Google unveiled Gemma, a series of lightweight open source generative AI models designed mainly for developers and researchers. According to Google, these models can run on laptop or desktop computers. 

In terms of regional trends, most LLMs today are from the United States (73 per cent) and China (15 per cent), according to research in 2023 by the Large European AI Models initiative.

Within the Asia-Pacific, China is the key producer of LLMs, with the only other country producing LLMs in the region being Singapore with three models, according to Stanford’s AI Index Report released in 2024.

Fuji Xerox aims for top market share spot in Myanmar
Six in 10 Singapore consumers targeted by tech support scams: Microsoft
HP ZBook Firefly G9 review: A futuristic look for a mobile workstation
Google Cloud extends partnership with Singapore government to build more AI services
Fines and warnings for 11 organisations for data privacy lapses in Singapore
TAGGED:AIGemmagpuIBMLLMMistralNaturePhi-3SLMwatsonx

Sign up for the TG newsletter

Never miss anything again. Get the latest news and analysis in your inbox.

By signing up, you agree to our Terms of Use and acknowledge the data practices in our Privacy Policy. You may unsubscribe at any time.
Share This Article
Facebook Whatsapp Whatsapp LinkedIn Copy Link Print
Avatar photo
ByAi Lei Tao
Ai Lei is a writer who has covered the technology scene for more than 20 years. She was previously the editor of Asia Computer Weekly (ACW), the only regional IT weekly in Asia. She has also written for TechTarget's ComputerWeekly, and was editor of CMPnetAsia and Associate Editor at Computerworld Singapore.
Previous Article Indonesian ransomware debacle shows that you need backup – and a lot more
Next Article AI at the edge: Google Distributed Cloud is air-gapped, disconnected
Leave a Comment

Leave a ReplyCancel reply

This site uses Akismet to reduce spam. Learn how your comment data is processed.

Stay Connected

FacebookLike
XFollow

Latest News

Scammers are so successful they even accidentally scam themselves now
Cybersecurity Internet
June 10, 2025
Doom: The Dark Ages review: Future fantastic demon slaying
Gaming
June 10, 2025
Plaud NotePin review: Note-taking made easy with AI
Internet Mobile
June 9, 2025
Can smart grocery carts, biometric payments boost retailers like FairPrice?
Enterprise Internet
June 6, 2025

Techgoondu.com is published by Goondu Media Pte Ltd, a company registered and based in Singapore.

.

Started in June 2008 by technology journalists and ex-journalists in Singapore who share a common love for all things geeky and digital, the site now includes segments on personal computing, enterprise IT and Internet culture.

banner banner
Everyday DIY
PC needs fixing? Get your hands on with the latest tech tips
READ ON
banner banner
Leaders Q&A
What tomorrow looks like to those at the leading edge today
FIND OUT
banner banner
Advertise with us
Discover unique access and impact with TG custom content
SHOW ME

 

 

POWERED BY READYSPACE
The Techgoondu website is powered by and managed by Readyspace Web Hosting.

TechgoonduTechgoondu
© 2024 Goondu Media Pte Ltd. All Rights Reserved | Privacy | Terms of Use | Advertise | About Us | Contact
Join Us!
Never miss anything again. Get the latest news and analysis in your inbox.

Zero spam, Unsubscribe at any time.
 

Loading Comments...
 

    Welcome Back!

    Sign in to your account

    Username or Email Address
    Password

    Lost your password?