Hello Reader 👋
Welcome to the 4th Issue Newsletter.
This issue focus topic segments and a new micro-learning format, making it easier to explore, learn, and dive deeper.
🚀 3 Things in AI
News | Concepts | Insights
🌊 News
OpenAI has agreed to buy AI device startup io in a deal worth $6.4 billion, which will also see founder and former Apple designer Sir Jony Ive join the ChatGPT-maker. Ive was already working alongside OpenAI founder Sam Altman as part of the creative collective LoveFrom.
Anthropic Releases Claude 4 Series
Anthropic has unveiled its Claude 4 series, including Claude Opus 4, touted as the world's most capable coding AI. The models can function autonomously for extended periods.
Saudi Arabia Launches AI Company 'Humain'
Saudi Arabia has launched 'Humain', an AI company under the Public Investment Fund, aiming to position the Kingdom as a global leader in AI innovation. The initiative aligns with Saudi Arabia's Vision 2030
AI-Assisted Films Now Eligible for Oscars
The Academy has updated its rules to allow AI-assisted films to win Oscars, provided they retain clear human input. This change aligns with technological advancements in filmmaking
🧠 Concept : Tokens & Parameters in AI Models
What are tokens in AI Models?
In AI language models, a token is a small piece of text, often a word, part of a word, or even punctuation. Different models tokenize differently, but the idea is the same: tokens are the units AI models read and write. For example:
Why did the banana go to the hospital? ( If you know the answer , comment :D )
When you type text into a language model, it doesn't understand raw text like humans do. It works with numbers.
Text → "Why did the banana go to the hospital?"
Tokenizer breaks it down into tokens
Each token is mapped to a token ID (a number). We will explore “How tokenizers works” in some other artifacts ) , For how each token assigned with ID.
These token IDs are what the model actually sees.
When you read “Claude can work with 200K tokens" mean?
That means Claude (an AI model by Anthropic) can handle over 200,000 tokens in one conversation or prompt.
To put that in perspective:
1,000 tokens ≈ 750 words
200,000 tokens ≈ 150,000 words (Roughly the length of a 500-page book!)
Key Points to Remember
API Usage: Most LLM APIs (like OpenAI's, Anthropic's, Google's) charge per token (both input and output). Sending a small question and getting a small answer through a model designed for massive contexts still incurs costs.
Specialization : Smaller models or models with moderate context windows can be more easily and cheaply fine-tuned on specific QA datasets, potentially outperforming a general-purpose large-context model on those specific tasks.
Avoid Irrelevant Information : If you're tempted to "fill up" the large context window with a lot of text just because you can, you might introduce noise or irrelevant information that could confuse the model or lead it to generate less precise answers for a simple question.
Latency (Speed): Processing more tokens takes more time. Even if your actual question and answer are short, a model designed for and handling larger context windows might have inherent architectural overheads that make it slower than a smaller, more specialized model.
What are Parameters in AI Models?
At their core, Parameters are the adjustable internal numbers that an AI model learns from data during training. These numbers store the model's knowledge and determine how it makes predictions or decisions.
Think of like neurons and connections in a brain.
During a process called model training, the model looks at tons of data, It makes guesses, gets feedback and then adjusts these 'knobs' (parameters) slightly. Technically parameters also called “Weights”
After training, the final settings of these parameters represent everything the AI model has 'learned' from the data and store these patterns and relationships. For example, specific parameter values (one knob details) might help the model distinguish a cat's pointy ears from a dog's floppy ones.
When you give the trained AI model new input, it uses these learned parameter settings to process that input and produce an output.
When someone says, "GPT-4 has 175 billion parameters," they’re talking about the model’s brainpower.

Based on Size/Scale (Parameter Count)
Small LLMs: (e.g., < 7 Billion parameters) - Can run on consumer hardware, good for specific tasks. Examples: DistilBERT, some smaller Llama/Mistral variants.
Medium LLMs: (e.g., 7B - 70B parameters) - Offer a good balance of capability and resource requirements. Examples: Llama 2 7B/13B/70B, Mistral 7B.
Large LLMs: (e.g., >100B parameters) - State-of-the-art performance, require significant computational resources. Examples: GPT-4, PaLM 2, Gemini Ultra.
Key Points to Remember
Larger Parameter Count = Higher API Costs & More Resources. Generally, models with more parameters are more expensive to use via an API. You might pay more per token processed or have higher base fees for accessing the more capable model.
Larger Parameter Count = Broader & More Nuanced Capabilities.
Larger models can generally handle more complex, open-ended, and nuanced tasks. They excel at deep reasoning, creative generation, understanding intricate instructions, and synthesizing information from larger contexts.
Larger Parameter Count = Generally Slower Inference (Higher Latency).
Processing information through billions of parameters takes more computational time than processing it through millions. This means responses from larger models will typically take longer to generate.
Smaller models are quicker and can provide near real-time responses, which is crucial for applications like interactive chatbots, quick data extraction, or autocomplete features.
📊 Insights : LLM Leaderboard
https://llm-stats.com/ - Analyze and compare AI models across benchmarks, pricing, and capabilities. We will also learn “How the AI Models are benchmarked” in upcoming short learning newsletter.
🔦 Explore This
1. Deepwiki
Instantly transform your GitHub repositories into beautiful, insightful knowledge bases with AI-powered documentation, diagrams, and assistants. Below one is for fastapi repo. There is also open source version of Deepwiki
2. Which AI Agent Framework Should You Use?
Building AI agents? You're likely facing a sea of framework options, each with its own strengths and weaknesses. Don't get lost in the details! This comparison tool, created by
, provides a clear and concise overview.https://multiagentbook.com/labs/frameworks/
Explore different factors and receive a score to help you choose the perfect framework for your needs. Even you could go one step further to see how many lines would take by different frameworks to build a very basic hello world agent.
https://multiagentbook.com/labs/usecases/?usecase=hello-world-agent
🧩 Engineers Placeholder
Career framework
Everyone wants to progress in their career, but there's often confusion about which path to take or what areas of development to focus on to reach the next level. progression.fyi - Career framework inspiration from the world's best companies.
How roles like Engineer, Senior, Staff, Principal, etc., are defined across the industry. Learn competencies are typically expected at each stage and different engineering path could take.
Hope you found something useful this week. Feel free to share your feedback !
Until next time,
Egg Puff Engineer