Large Language Models
Updated: 2025-11-24
Large Language Models (LLMs) are language models trained on vast amounts of text data to understand and generate human language. LLMs have one simple objective: to predict the next token given previous context in a sequence. This foundational objective, combined with massive scale, enables them to:
Text Understanding and Generation: Answer questions, write creative content, translate languages
Reasoning and Analysis: Break down complex problems, perform multi-step reasoning, make logical inferences
Information Synthesis: Summarize documents, extract key insights, and aggregate information
Interactive Assistance: Maintain context in conversations and adapt to user needs
And more!
Proprietary LLMs
Several commercial models are available through API access with usage-based pricing:
GPT (OpenAI): Pro, regular, Mini, Nano
Claude (Anthropic): Opus, Sonnet, Haiku
Gemini (Google): Pro, Flash, Flash-Lite
Open-Source LLMs
Open-source models can be downloaded and run locally, though they often require significant computational resources (high-end GPUs like NVIDIA H100, H200):
Llama (Meta): 1B, 3B, 8B, 70B, 405B
Qwen (Alibaba): 0.5B, 1.5B, 3B, 7B, 14B, 32B, 72B
Gemma (Google): 2B, 7B, 9B, 27B
Model Tiers
Different tiers of models excel at different tasks:
Frontier models: Best for complex inference, nuanced understanding, and creative tasks
Fast models: Optimized for speed and cost-efficiency with good general performance
Specialized models: Designed for specific capabilities like advanced reasoning or coding
When choosing a model, consider:
Task complexity: How much reasoning or context understanding is required?
Latency requirements: How quickly do you need responses?
Cost constraints: What's your budget for API calls?
Context window: How much text do you need to process at once?
LLM Access
You can interact with LLMs through multiple interfaces:
API Access: Programmatic access for building applications and integrating into workflows
Local Deployment: Running open-source models on your infrastructure
For this course, you will primarily use API access to programmatically interact with LLMs, enabling you to build sophisticated NLP applications.
Last updated
Was this helpful?