Nov 2 Weekly Newsletter

AI Newsletter - Nov 2 highlights: Smaller models are better. Layer skip decoding is better than beam search. Tokenformer is a better architecture for scaling language models. Robots can sense the touch using a new sensors developed by meta. OpenAI and Google are trading punches with new releases within minutes of each other. SmolLM 2: Smaller Models, Bigger Impact Hugging Face has unveiled SmolLM v2, a groundbreaking release pushing the boundaries of small language models....

November 2, 2024 · 3 min · 463 words · Prakyath Kantharaju

Weekly Newsletter - October 26

AI Newsletter - October 2024 Sarvam: Breaking New Ground in Indian Language AI India’s linguistic diversity has long posed a challenge for AI development, but Sarvam-1 is changing the game. This 2-billion parameter language model specifically targets 10 major Indian languages alongside English, proving that thoughtful design can outperform brute force. The model’s innovative approach tackles two critical challenges: token efficiency and data quality. Traditional multilingual models struggle with Indic scripts, requiring up to 8 tokens per word compared to English’s 1....

October 26, 2024 · 4 min · 652 words · Prakyath Kantharaju

Weekly Newsletter - October 19

AI Newsletter - October 2024 1. Impact of Quantization on Large Language Models Recent extensive testing of the Llama 3.1 series models has shed light on the critical role of quantization in optimizing LLM deployments. The evaluation, conducted across various model sizes (8B, 70B, and 405B), compared three quantization schemes against the baseline 16-bit model: W8A8-INT: 8-bit integer quantization for weights and activations W8A8-FP: 8-bit floating-point quantization for weights and activations W4A16-INT: 4-bit integer quantization for weights, 16-bit precision for activations Key findings include:...

October 19, 2024 · 5 min · 895 words · Prakyath Kantharaju

Weekly Newsletter - Oct 11

Weekly Newsletter - October 11, 2024 AI Breakthroughs Take Center Stage in Nobel Prizes This year’s Nobel Prizes have put a spotlight on AI’s transformative impact across scientific disciplines. The Physics prize, awarded to Geoffrey Hinton and John Hopfield, celebrates their groundbreaking work on artificial neural networks – the very foundation of today’s deep learning revolution. Meanwhile, the Chemistry prize went to Demis Hassabis, John Jumper, and David Baker for their development of AlphaFold....

October 11, 2024 · 4 min · 682 words · Prakyath Kantharaju

AI Weekly Newsletter (New open models) - Sep 27

AI Weekly Newsletter (New open models) Highlights OpenAI experiences significant leadership changes Meta releases Llama 3 models with impressive benchmarks Google unveils updated Gemini models with performance improvements and price reductions AlphaChip transforms computer chip design using AI News OpenAI Leadership Changes Top executives leaving OpenAI: CTO Mira Murati Chief Research Officer Bob McGrew Research Leader Barret Zoph Llama 3.2 Release Meta releases new Llama 3.2 models: 11B model comparable/slightly better than Claude Haiku 90B model comparable/slightly better than GPT-4o-mini New 128k-context 1B and 3B models competing with Gemma 2 and Phi 3....

September 15, 2020 · 3 min · 463 words · Me