Sebastian Raschka's Blog

August 9, 2025

From GPT-2 to gpt-oss: Analyzing the Architectural Advances

OpenAI just released their new open-weight LLMs this week: gpt-oss-120b and gpt-oss-20b, their first open-weight models since GPT-2 in 2019. And yes, thanks to some clever optimizations, they can run locally. I spent the past few days reading through the code and technical reports to summarize the most interesting details.
 •  0 comments  •  flag
Share on Twitter
Published on August 09, 2025 04:00

July 18, 2025

The Big LLM Architecture Comparison

It has been seven years since the original GPT architecture was developed. At first glance, looking back at GPT-2 (2019) and forward to DeepSeek-V3 and Llama 4 (2024-2025), one might be surprised at how structurally similar these models still are. Comparing LLMs to determine the key ingredients that contribute to their good (or not-so-good) performance is notoriously challenging: datasets, training techniques, and hyperparameters vary widely and are often not well documented. However, I think that there is still a lot of value in examining the structural changes of the architectures themselves to see what LLM developers are up to in 2025.
 •  0 comments  •  flag
Share on Twitter
Published on July 18, 2025 23:00

June 30, 2025

LLM Research Papers: The 2025 List (January to June)

The latest in LLM research with a hand-curated, topic-organized list of over 200 research papers from 2025.
 •  0 comments  •  flag
Share on Twitter
Published on June 30, 2025 23:06

June 17, 2025

Understanding and Coding the KV Cache in LLMs from Scratch

KV caches are one of the most critical techniques for efficient inference in LLMs in production. KV caches are an important component for compute-efficient LLM inference in production. This article explains how they work conceptually and in code with a from-scratch, human-readable implementation.
 •  0 comments  •  flag
Share on Twitter
Published on June 17, 2025 01:00

May 9, 2025

Coding LLMs from the Ground Up: A Complete Course

Why build an LLM from scratch? It's probably the best and most efficient way to learn how LLMs really work. Plus, many readers have told me they had a lot of fun doing it.
 •  0 comments  •  flag
Share on Twitter
Published on May 09, 2025 17:00

April 18, 2025

The State of Reinforcement Learning for LLM Reasoning

A lot has happened this month, especially with the releases of new flagship models like GPT-4.5 and Llama 4. But you might have noticed that reactions to these releases were relatively muted. Why? One reason could be that GPT-4.5 and Llama 4 remain conventional models, which means they were trained without explicit reinforcement learning for reasoning. However, OpenAI's recent release of the o3 reasoning model demonstrates there is still considerable room for improvement when investing compute strategically, specifically via reinforcement learning methods tailored for reasoning tasks. While reasoning alone isn't a silver bullet, it reliably improves model accuracy and problem-solving capabilities on challenging tasks (so far). And I expect reasoning-focused post-training to become standard practice in future LLM pipelines. So, in this article, let's explore the latest developments in reasoning via reinforcement learning.
 •  0 comments  •  flag
Share on Twitter
Published on April 18, 2025 17:00

March 28, 2025

First Look at Reasoning From Scratch: Chapter 1

As you know, I've been writing a lot lately about the latest research on reasoning in LLMs. Before my next research-focused blog post, I wanted to offer something special to my paid subscribers as a thank-you for your ongoing support. So, I've started writing a new book on how reasoning works in LLMs, and here I'm sharing the first Chapter 1 with you. This ~15-page chapter is an introduction reasoning in the context of LLMs and provides an overview of methods like inference-time scaling and reinforcement learning. Thanks for your support! I hope you enjoy the chapter, and stay tuned for my next blog post on reasoning research!
 •  0 comments  •  flag
Share on Twitter
Published on March 28, 2025 23:03

March 7, 2025

Inference-Time Compute Scaling Methods to Improve Reasoning Models

This article explores recent research advancements in reasoning-optimized LLMs, with a particular focus on inference-time compute scaling that have emerged since the release of DeepSeek R1.
 •  0 comments  •  flag
Share on Twitter
Published on March 07, 2025 22:03

The State of LLM Reasoning Models

This article explores recent research advancements in reasoning-optimized LLMs, with a particular focus on inference-time compute scaling that have emerged since the release of DeepSeek R1.
 •  0 comments  •  flag
Share on Twitter
Published on March 07, 2025 22:03

February 4, 2025

Understanding Reasoning LLMs

In this article, I will describe the four main approaches to building reasoning models, or how we can enhance LLMs with reasoning capabilities. I hope this provides valuable insights and helps you navigate the rapidly evolving literature and hype surrounding this topic.
 •  0 comments  •  flag
Share on Twitter
Published on February 04, 2025 22:03

Sebastian Raschka's Blog

Sebastian Raschka
Sebastian Raschka isn't a Goodreads Author (yet), but they do have a blog, so here are some recent posts imported from their feed.
Follow Sebastian Raschka's blog with rss.