From bcab4969f401cbf92e1784e5e0f16ad5b2d894b3 Mon Sep 17 00:00:00 2001 From: Automated Publisher Date: Sat, 16 Aug 2025 21:07:37 +0000 Subject: [PATCH] =?UTF-8?q?=F0=9F=93=9A=20Auto-publish:=20Add/update=202?= =?UTF-8?q?=20blog=20posts?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Generated on: Sat Aug 16 21:07:37 UTC 2025 Source: md-personal repository --- ...a-comprehensive-guide-to-breville-barista-pro-maintenance.md | 2 +- content/posts/a-deep-dive-into-ppo-for-language-models.md | 2 +- 2 files changed, 2 insertions(+), 2 deletions(-) diff --git a/content/posts/a-comprehensive-guide-to-breville-barista-pro-maintenance.md b/content/posts/a-comprehensive-guide-to-breville-barista-pro-maintenance.md index 80c4ffd..1bdd444 100644 --- a/content/posts/a-comprehensive-guide-to-breville-barista-pro-maintenance.md +++ b/content/posts/a-comprehensive-guide-to-breville-barista-pro-maintenance.md @@ -1,6 +1,6 @@ --- title: "A Comprehensive Guide to Breville Barista Pro Maintenance" -date: 2025-08-16T20:48:28 +date: 2025-08-16T21:07:33 draft: false --- diff --git a/content/posts/a-deep-dive-into-ppo-for-language-models.md b/content/posts/a-deep-dive-into-ppo-for-language-models.md index 619d178..e5eb7e3 100644 --- a/content/posts/a-deep-dive-into-ppo-for-language-models.md +++ b/content/posts/a-deep-dive-into-ppo-for-language-models.md @@ -9,7 +9,7 @@ Large Language Models (LLMs) have demonstrated astonishing capabilities, but out You may have seen diagrams like the one below, which outlines the RLHF training process. It can look intimidating, with a web of interconnected models, losses, and data flows. -![](/images/a-deep-dive-into-ppo-for-language-models/64bfdb4b-678e-4bfc-8b62-0c05c243f6a9.png) +![[Pasted image 20250816140700.png]] This post will decode that diagram, piece by piece. We'll explore the "why" behind each component, moving from high-level concepts to the deep technical reasoning that makes this process work.