Logo

dev-resources.site

for different kinds of informations.

Small Language Models are Going to Eat the World.

Published at
1/22/2024
Categories
llmops
promptengineering
softwareengineering
llm
Author
justinmacorin
Author
13 person written this
justinmacorin
open
Small Language Models are Going to Eat the World.

Today, Large Language Models (LLMs) typically require internet access. As prompt-based applications become ubiquitous, there is a high likelihood we slowly begin to see a transition from internet-based models to locally hosted models.

Local models are nothing new. Google product users are often pushed to download local models for Google Maps, Google Translate, and Text2Speech. These models run locally for four primary purposes:

  • speed
  • reliability
  • privacy
  • cost

Benefits

Speed

Local models have no network latency. They run locally, and instructions and data transfers happen closer to the application layer, resulting in increased performance.

Reliability

Local models are self-reliant. They don't require additional computers to operate and don't rely on 3rd party service providers. They run as stand-alone and won't break if internet connectivity is lost.

Privacy

Private information is processed locally and never shared with another provider. Information passed into these models may contain private or confidential information that an external processor should not process.

Cost

Local models require zero hosting. Models may run frequently, and costs involved in processing data regularly at scale may become unaffordable or may better be absorbed by a local device.

How can we make local models a reality?

Python is the language of choice to run LLMs. However, we know that embedded devices, mobile apps, and web servers often use different languages to run and operate efficiently.

To bridge the gap in SDKs for accessing large language models across various platforms, engineers should consider developing and integrating multi-language libraries and frameworks that are compatible with mobile, embedded, and diverse server environments. Embracing innovation and flexibility in these developments is critical, as large language models represent a new technological frontier rather than merely enhancing existing tools.

This original article can be found here: https://promptdesk.ai/articles/small-languages-models-are-going-to-eat-the-world

llmops Article's
30 articles in total
Favicon
A Beginners Guide to LLMOps
Favicon
LLMOps [Quick Guide]
Favicon
The power of MIPROv2 - using DSPy optimizers for your LLM-pipelines
Favicon
Unifying or Separating Endpoints in Generative AI Applications on AWS
Favicon
📚 Download My DevOps and LLMOps Books for Free!📚
Favicon
Deploying LLM Inference Endpoints & Optimizing Output with RAG
Favicon
End to End LLMOps Pipeline - Part 8 - AWS EKS
Favicon
🤖 End to end LLMOps Pipeline - Part 7- Validating Kubernetes Manifests with kube-score🤖
Favicon
📚 Announcing My New Book: Building an LLMOps Pipeline Using Hugging Face 📚
Favicon
End to end LLMOps Pipeline - Part 2 - FastAPI
Favicon
End to end LLMOps Pipeline - Part 1 - Hugging Face
Favicon
Bridging the Gap: Integrating Responsible AI Practices into Scalable LLMOps for Enterprise Excellence
Favicon
Building a Traceable RAG System with Qdrant and Langtrace: A Step-by-Step Guide
Favicon
FastAPI for Data Applications: Dockerizing and Scaling Your API on Kubernetes. Part II
Favicon
FastAPI for Data Applications: From Concept to Creation. Part I
Favicon
Evaluation of OpenAI Assistants
Favicon
Vector stores and embeddings: Dive into the concept of embeddings and explore vector store integrations within LangChain
Favicon
Finding the Perfect Model for Your Project on the Hugging Face Hub
Favicon
The Future of Natural Language APIs
Favicon
How do you know that an LLM-generated response is factually correct? 🤔
Favicon
The Era of LLM Infrastructure
Favicon
Launching LLM apps? Beware of prompt leaks
Favicon
Small Language Models are Going to Eat the World.
Favicon
No Code: Dify's Open Source App Building Revolution
Favicon
Pipeline Parallelism in PyTorch
Favicon
Orquesta raises €800,000 in pre-seed funding!
Favicon
Lifecycle of a Prompt: A Guide to Effective Prompts
Favicon
Integrate Orquesta with LangChain
Favicon
LLM Analytics 101 - How to Improve your LLM app
Favicon
Build an AI App in 5 Minutes without Coding

Featured ones: