diff --git a/404.html b/404.html index 815489c..8296236 100644 --- a/404.html +++ b/404.html @@ -4,4 +4,4 @@ 2016 - 2025 Eric X. Liu -[c06e978] \ No newline at end of file +[0945204] \ No newline at end of file diff --git a/about/index.html b/about/index.html index d8443cb..e553ccd 100644 --- a/about/index.html +++ b/about/index.html @@ -4,4 +4,4 @@ 2016 - 2025 Eric X. Liu -[c06e978] \ No newline at end of file +[0945204] \ No newline at end of file diff --git a/categories/index.html b/categories/index.html index ded53da..fdf123f 100644 --- a/categories/index.html +++ b/categories/index.html @@ -4,4 +4,4 @@ 2016 - 2025 Eric X. Liu -[c06e978] \ No newline at end of file +[0945204] \ No newline at end of file diff --git a/images/a-technical-deep-dive-into-the-transformer-s-core-mechanics/.png b/images/a-technical-deep-dive-into-the-transformer-s-core-mechanics/.png new file mode 100644 index 0000000..a7d5e5d Binary files /dev/null and b/images/a-technical-deep-dive-into-the-transformer-s-core-mechanics/.png differ diff --git a/index.html b/index.html index ad76cc7..4d22342 100644 --- a/index.html +++ b/index.html @@ -4,4 +4,4 @@ 2016 - 2025 Eric X. Liu -[c06e978] \ No newline at end of file +[0945204] \ No newline at end of file diff --git a/index.xml b/index.xml index 0be3bae..7ea126a 100644 --- a/index.xml +++ b/index.xml @@ -1,4 +1,4 @@ -Eric X. Liu's Personal Page/Recent content on Eric X. Liu's Personal PageHugoenWed, 20 Aug 2025 04:16:22 +0000A Comprehensive Guide to Breville Barista Pro Maintenance/posts/a-comprehensive-guide-to-breville-barista-pro-maintenance/Wed, 20 Aug 2025 04:16:13 +0000/posts/a-comprehensive-guide-to-breville-barista-pro-maintenance/<p>Proper maintenance is critical for the longevity and performance of a Breville Barista Pro espresso machine. Consistent cleaning not only ensures the machine functions correctly but also directly impacts the quality of the espresso produced. This guide provides a detailed, technical breakdown of the essential maintenance routines, from automated cycles to daily upkeep.</p> +Eric X. Liu's Personal Page/Recent content on Eric X. Liu's Personal PageHugoenWed, 20 Aug 2025 04:32:59 +0000A Comprehensive Guide to Breville Barista Pro Maintenance/posts/a-comprehensive-guide-to-breville-barista-pro-maintenance/Wed, 20 Aug 2025 04:32:52 +0000/posts/a-comprehensive-guide-to-breville-barista-pro-maintenance/<p>Proper maintenance is critical for the longevity and performance of a Breville Barista Pro espresso machine. Consistent cleaning not only ensures the machine functions correctly but also directly impacts the quality of the espresso produced. This guide provides a detailed, technical breakdown of the essential maintenance routines, from automated cycles to daily upkeep.</p> <h4 id="understanding-the-two-primary-maintenance-cycles"> <strong>Understanding the Two Primary Maintenance Cycles</strong> <a class="heading-link" href="#understanding-the-two-primary-maintenance-cycles"> @@ -6,7 +6,15 @@ <span class="sr-only">Link to heading</span> </a> </h4> -<p>The Breville Barista Pro has two distinct, automated maintenance procedures: the <strong>Cleaning (Flush) Cycle</strong> and the <strong>Descale Cycle</strong>. It is important to understand that these are not interchangeable, as they address different types of buildup within the machine.</p>Fixing GPU Operator Pods Stuck in Init: Secure Boot, DKMS, and MOK on Proxmox + Debian/posts/secure-boot-dkms-and-mok-on-proxmox-debian/Sat, 09 Aug 2025 00:00:00 +0000/posts/secure-boot-dkms-and-mok-on-proxmox-debian/<p>I hit an issue where all GPU Operator pods on one node were stuck in Init after migrating from Legacy BIOS to UEFI. The common error was NVIDIA components waiting for “toolkit-ready,” while the toolkit init container looped with:</p> +<p>The Breville Barista Pro has two distinct, automated maintenance procedures: the <strong>Cleaning (Flush) Cycle</strong> and the <strong>Descale Cycle</strong>. It is important to understand that these are not interchangeable, as they address different types of buildup within the machine.</p>A Technical Deep Dive into the Transformer's Core Mechanics/posts/a-technical-deep-dive-into-the-transformer-s-core-mechanics/Wed, 20 Aug 2025 04:32:52 +0000/posts/a-technical-deep-dive-into-the-transformer-s-core-mechanics/<p>The Transformer architecture is the bedrock of modern Large Language Models (LLMs). While its high-level success is widely known, a deeper understanding requires dissecting its core components. This article provides a detailed, technical breakdown of the fundamental concepts within a Transformer block, from the notion of &ldquo;channels&rdquo; to the intricate workings of the attention mechanism and its relationship with other advanced architectures like Mixture of Experts.</p> +<h3 id="1-the-channel-a-foundational-view-of-d_model"> + 1. The &ldquo;Channel&rdquo;: A Foundational View of <code>d_model</code> + <a class="heading-link" href="#1-the-channel-a-foundational-view-of-d_model"> + <i class="fa-solid fa-link" aria-hidden="true" title="Link to heading"></i> + <span class="sr-only">Link to heading</span> + </a> +</h3> +<p>In deep learning, a &ldquo;channel&rdquo; can be thought of as a feature dimension. While this term is common in Convolutional Neural Networks for images (e.g., Red, Green, Blue channels), in LLMs, the analogous concept is the model&rsquo;s primary embedding dimension, commonly referred to as <code>d_model</code>.</p>Fixing GPU Operator Pods Stuck in Init: Secure Boot, DKMS, and MOK on Proxmox + Debian/posts/secure-boot-dkms-and-mok-on-proxmox-debian/Sat, 09 Aug 2025 00:00:00 +0000/posts/secure-boot-dkms-and-mok-on-proxmox-debian/<p>I hit an issue where all GPU Operator pods on one node were stuck in Init after migrating from Legacy BIOS to UEFI. The common error was NVIDIA components waiting for “toolkit-ready,” while the toolkit init container looped with:</p> <ul> <li>nvidia-smi failed to communicate with the NVIDIA driver</li> <li>modprobe nvidia → “Key was rejected by service”</li> diff --git a/posts/a-comprehensive-guide-to-breville-barista-pro-maintenance/index.html b/posts/a-comprehensive-guide-to-breville-barista-pro-maintenance/index.html index 7996f3c..22515c1 100644 --- a/posts/a-comprehensive-guide-to-breville-barista-pro-maintenance/index.html +++ b/posts/a-comprehensive-guide-to-breville-barista-pro-maintenance/index.html @@ -8,10 +8,10 @@ The Breville Barista Pro has two distinct, automated maintenance procedures: the Cleaning (Flush) Cycle and the Descale Cycle. It is important to understand that these are not interchangeable, as they address different types of buildup within the machine.">
\ No newline at end of file diff --git a/posts/a-deep-dive-into-ppo-for-language-models/index.html b/posts/a-deep-dive-into-ppo-for-language-models/index.html index 71797ab..c54c6d2 100644 --- a/posts/a-deep-dive-into-ppo-for-language-models/index.html +++ b/posts/a-deep-dive-into-ppo-for-language-models/index.html @@ -23,4 +23,4 @@ where δ_t = r_t + γV(s_{t+1}) - V(s_t)

  • γ (gam 2016 - 2025 Eric X. Liu -[c06e978] \ No newline at end of file +[0945204] \ No newline at end of file diff --git a/posts/a-technical-deep-dive-into-the-transformer-s-core-mechanics/index.html b/posts/a-technical-deep-dive-into-the-transformer-s-core-mechanics/index.html new file mode 100644 index 0000000..5def962 --- /dev/null +++ b/posts/a-technical-deep-dive-into-the-transformer-s-core-mechanics/index.html @@ -0,0 +1,39 @@ +A Technical Deep Dive into the Transformer's Core Mechanics · Eric X. Liu's Personal Page

    A Technical Deep Dive into the Transformer's Core Mechanics

    The Transformer architecture is the bedrock of modern Large Language Models (LLMs). While its high-level success is widely known, a deeper understanding requires dissecting its core components. This article provides a detailed, technical breakdown of the fundamental concepts within a Transformer block, from the notion of “channels” to the intricate workings of the attention mechanism and its relationship with other advanced architectures like Mixture of Experts.

    1. The “Channel”: A Foundational View of d_model + +Link to heading

    In deep learning, a “channel” can be thought of as a feature dimension. While this term is common in Convolutional Neural Networks for images (e.g., Red, Green, Blue channels), in LLMs, the analogous concept is the model’s primary embedding dimension, commonly referred to as d_model.

    An input text is first tokenized, and each token is mapped to a vector of size d_model (e.g., 4096). Each of the 4096 dimensions in this vector can be considered a “channel,” representing a different semantic or syntactic feature of the token.

    As this data, represented by a tensor of shape [batch_size, sequence_length, d_model], progresses through the layers of the Transformer, these channels are continuously transformed. However, a critical design choice is that the output dimension of every main sub-layer (like the attention block or the FFN block) is also d_model. This consistency is essential for enabling residual connections, where the input to a block is added to its output (output = input + SubLayer(input)). This technique is vital for training the extremely deep networks common today.

    2. The Building Blocks: Dimensions of Key Layers + +Link to heading

    A Transformer layer is primarily composed of two sub-layers: a Multi-Head Attention block and a position-wise Feed-Forward Network (FFN). The parameters for these are stored in several key weight matrices. Understanding their dimensions is crucial.

    Let’s define our variables:

    • d_model: The core embedding dimension.
    • d_ff: The inner dimension of the FFN, typically 4 * d_model.
    • h: The number of attention heads.
    • d_head: The dimension of each attention head, where d_model = h * d_head.

    The dimensions of the weight matrices are as follows:

    LayerWeight MatrixInput Vector ShapeOutput Vector ShapeWeight Matrix Dimension
    Attention Projections
    QueryW_Qd_modeld_model[d_model, d_model]
    KeyW_Kd_modeld_model[d_model, d_model]
    ValueW_Vd_modeld_model[d_model, d_model]
    OutputW_Od_modeld_model[d_model, d_model]
    Feed-Forward Network
    Layer 1 (Up-projection)W_ff1d_modeld_ff[d_model, d_ff]
    Layer 2 (Down-projection)W_ff2d_ffd_model[d_ff, d_model]

    3. Deconstructing Multi-Head Attention (MHA) + +Link to heading

    The core innovation of the Transformer is Multi-Head Attention. It allows the model to weigh the importance of different tokens in the sequence from multiple perspectives simultaneously. +

    3.1. The “Why”: Beyond a Single Attention + +Link to heading

    A single attention mechanism would force the model to average all types of linguistic relationships into one pattern. MHA avoids this by creating h parallel subspaces. Each “head” can specialize, with one head learning syntactic dependencies, another tracking semantic similarity, and so on. This creates a much richer representation.

    3.2. An Encoding/Decoding Analogy + +Link to heading

    A powerful way to conceptualize the attention calculation is as a two-stage process:

    1. Encoding Relationships: The first part of the calculation, softmax(Q @ K.T), can be seen as an encoding step. It does not use the actual “content” of the tokens (the V vectors). Instead, it uses the Queries and Keys to build a dynamic “relationship map” between tokens in the sequence. This map, a matrix of attention scores, answers the question: “For each token, how important is every other token right now?”
    2. Decoding via Information Retrieval: The second part, scores @ V, acts as a decoding step. It uses the relationship map to retrieve and synthesize information. For each token, it creates a new vector by taking a weighted sum of all the V vectors in the sequence, using the scores as the precise mixing recipe. It decodes the relational structure into a new, context-aware representation.

    3.3. The “How”: A Step-by-Step Flow + +Link to heading

    The MHA process is designed for maximum computational efficiency.

    1. Initial Projections: The input vectors (shape [seq_len, d_model]) are multiplied by W_Q, W_K, and W_V. These matrices are all [d_model, d_model] not to create one large query, but to efficiently compute the vectors for all h heads at once. The single large output vector is then reshaped into h separate vectors, each of size d_head.
    2. Attention Score Calculation: For each head i, a score matrix is calculated: scores_i = softmax( (Q_i @ K_i.T) / sqrt(d_head) ). Note that Q_i and K_i have dimensions [seq_len, d_head], so the resulting scores_i matrix has a dimension of [seq_len, seq_len].
    3. Weighted Value Calculation: The scores are used to create a weighted sum of the Value vectors for each head: output_i = scores_i @ V_i. Since scores_i is [seq_len, seq_len] and V_i is [seq_len, d_head], the resulting output_i has a dimension of [seq_len, d_head]. This is the final output of a single head.
    4. Concatenation and Final Projection: The outputs of all h heads are concatenated along the last dimension. This produces a single large matrix of shape [seq_len, h * d_head], which is equivalent to [seq_len, d_model]. This matrix is then passed through the final output projection layer, W_O (shape [d_model, d_model]), to produce the attention block’s final output. The W_O matrix learns the optimal way to mix the information from all the specialized heads into a single, unified representation.

    4. Optimizing Attention: GQA and MQA + +Link to heading

    During inference, storing the Key and Value vectors for all previous tokens (the KV Cache) is a major memory bottleneck. Grouped-Query Attention (GQA) and Multi-Query Attention (MQA) are architectural modifications that address this by allowing multiple Query heads to share the same Key and Value heads.

    Let’s use a concrete example, similar to Llama 2 7B:

    • d_model = 4096
    • h = 32 Q heads
    • d_head = 128
    • g = 8 KV head groups for GQA

    The key insight is that only the dimensions of the W_K and W_V matrices change, which in turn reduces the size of the KV cache. The W_Q and W_O matrices remain [4096, 4096].

    Attention TypeNo. of Q HeadsNo. of KV HeadsW_K & W_V DimensionRelative KV Cache Size
    MHA (Multi-Head)3232[4096, 32*128] = [4096, 4096]1x (Baseline)
    GQA (Grouped)328[4096, 8*128] = [4096, 1024]1/4x
    MQA (Multi-Query)321[4096, 1*128] = [4096, 128]1/32x

    GQA provides a robust balance, significantly reducing the memory and bandwidth requirements for the KV cache with negligible impact on model performance, making it a popular choice in modern LLMs.

    5. MHA vs. Mixture of Experts (MoE): A Clarification + +Link to heading

    While both MHA and MoE use the concept of “experts,” they are functionally and architecturally distinct.

    • MHA: The “experts” are the attention heads. All heads are active for every token to build a rich representation within the attention layer. This is akin to a board meeting where every member analyzes and contributes to every decision.
    • MoE: The “experts” are full Feed-Forward Networks. A routing network selects a small subset of these FFNs for each token. This is a scaling strategy to increase a model’s parameter count for greater capacity while keeping the computational cost fixed. It replaces the standard FFN block, whereas MHA is the attention block.

    By understanding these technical details, from the basic concept of a channel to the sophisticated interplay of heads and experts, one can build a more complete and accurate mental model of how LLMs truly operate.


    References + +Link to heading

    1. Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A. N., … & Polosukhin, I. (2017). Attention is all you need. Advances in neural information processing systems, 30.
    2. Shazeer, N., Mirhoseini, A., Maziarz, K., Davis, A., Le, Q., Hinton, G., & Dean, J. (2017). Outrageously large neural networks: The sparsely-gated mixture-of-experts layer. arXiv preprint arXiv:1701.06538.
    3. Ainslie, J., Ontanon, J., Cakka, E., Dosovitskiy, A., & Le, Q. V. (2023). GQA: Training Generalized Multi-Query Transformer Models from Multi-Head Checkpoints. arXiv preprint arXiv:2305.13245.
    \ No newline at end of file diff --git a/posts/espresso-theory-application-a-guide-for-the-breville-barista-pro/index.html b/posts/espresso-theory-application-a-guide-for-the-breville-barista-pro/index.html index 5001709..f2ce188 100644 --- a/posts/espresso-theory-application-a-guide-for-the-breville-barista-pro/index.html +++ b/posts/espresso-theory-application-a-guide-for-the-breville-barista-pro/index.html @@ -20,4 +20,4 @@ Our overarching philosophy is simple: isolate and change only one variable at a 2016 - 2025 Eric X. Liu -[c06e978] \ No newline at end of file +[0945204] \ No newline at end of file diff --git a/posts/how-rvq-teaches-llms-to-see-and-hear/index.html b/posts/how-rvq-teaches-llms-to-see-and-hear/index.html index f63a350..b9ede7c 100644 --- a/posts/how-rvq-teaches-llms-to-see-and-hear/index.html +++ b/posts/how-rvq-teaches-llms-to-see-and-hear/index.html @@ -18,4 +18,4 @@ The answer lies in creating a universal language—a bridge between the continuo 2016 - 2025 Eric X. Liu -[c06e978] \ No newline at end of file +[0945204] \ No newline at end of file diff --git a/posts/index.html b/posts/index.html index 79ed627..a552d3e 100644 --- a/posts/index.html +++ b/posts/index.html @@ -1,7 +1,8 @@ Posts · Eric X. Liu's Personal Page
    \ No newline at end of file +[0945204] \ No newline at end of file diff --git a/posts/index.xml b/posts/index.xml index 435aafa..6812c99 100644 --- a/posts/index.xml +++ b/posts/index.xml @@ -1,4 +1,4 @@ -Posts on Eric X. Liu's Personal Page/posts/Recent content in Posts on Eric X. Liu's Personal PageHugoenWed, 20 Aug 2025 04:16:22 +0000A Comprehensive Guide to Breville Barista Pro Maintenance/posts/a-comprehensive-guide-to-breville-barista-pro-maintenance/Wed, 20 Aug 2025 04:16:13 +0000/posts/a-comprehensive-guide-to-breville-barista-pro-maintenance/<p>Proper maintenance is critical for the longevity and performance of a Breville Barista Pro espresso machine. Consistent cleaning not only ensures the machine functions correctly but also directly impacts the quality of the espresso produced. This guide provides a detailed, technical breakdown of the essential maintenance routines, from automated cycles to daily upkeep.</p> +Posts on Eric X. Liu's Personal Page/posts/Recent content in Posts on Eric X. Liu's Personal PageHugoenWed, 20 Aug 2025 04:32:59 +0000A Comprehensive Guide to Breville Barista Pro Maintenance/posts/a-comprehensive-guide-to-breville-barista-pro-maintenance/Wed, 20 Aug 2025 04:32:52 +0000/posts/a-comprehensive-guide-to-breville-barista-pro-maintenance/<p>Proper maintenance is critical for the longevity and performance of a Breville Barista Pro espresso machine. Consistent cleaning not only ensures the machine functions correctly but also directly impacts the quality of the espresso produced. This guide provides a detailed, technical breakdown of the essential maintenance routines, from automated cycles to daily upkeep.</p> <h4 id="understanding-the-two-primary-maintenance-cycles"> <strong>Understanding the Two Primary Maintenance Cycles</strong> <a class="heading-link" href="#understanding-the-two-primary-maintenance-cycles"> @@ -6,7 +6,15 @@ <span class="sr-only">Link to heading</span> </a> </h4> -<p>The Breville Barista Pro has two distinct, automated maintenance procedures: the <strong>Cleaning (Flush) Cycle</strong> and the <strong>Descale Cycle</strong>. It is important to understand that these are not interchangeable, as they address different types of buildup within the machine.</p>Fixing GPU Operator Pods Stuck in Init: Secure Boot, DKMS, and MOK on Proxmox + Debian/posts/secure-boot-dkms-and-mok-on-proxmox-debian/Sat, 09 Aug 2025 00:00:00 +0000/posts/secure-boot-dkms-and-mok-on-proxmox-debian/<p>I hit an issue where all GPU Operator pods on one node were stuck in Init after migrating from Legacy BIOS to UEFI. The common error was NVIDIA components waiting for “toolkit-ready,” while the toolkit init container looped with:</p> +<p>The Breville Barista Pro has two distinct, automated maintenance procedures: the <strong>Cleaning (Flush) Cycle</strong> and the <strong>Descale Cycle</strong>. It is important to understand that these are not interchangeable, as they address different types of buildup within the machine.</p>A Technical Deep Dive into the Transformer's Core Mechanics/posts/a-technical-deep-dive-into-the-transformer-s-core-mechanics/Wed, 20 Aug 2025 04:32:52 +0000/posts/a-technical-deep-dive-into-the-transformer-s-core-mechanics/<p>The Transformer architecture is the bedrock of modern Large Language Models (LLMs). While its high-level success is widely known, a deeper understanding requires dissecting its core components. This article provides a detailed, technical breakdown of the fundamental concepts within a Transformer block, from the notion of &ldquo;channels&rdquo; to the intricate workings of the attention mechanism and its relationship with other advanced architectures like Mixture of Experts.</p> +<h3 id="1-the-channel-a-foundational-view-of-d_model"> + 1. The &ldquo;Channel&rdquo;: A Foundational View of <code>d_model</code> + <a class="heading-link" href="#1-the-channel-a-foundational-view-of-d_model"> + <i class="fa-solid fa-link" aria-hidden="true" title="Link to heading"></i> + <span class="sr-only">Link to heading</span> + </a> +</h3> +<p>In deep learning, a &ldquo;channel&rdquo; can be thought of as a feature dimension. While this term is common in Convolutional Neural Networks for images (e.g., Red, Green, Blue channels), in LLMs, the analogous concept is the model&rsquo;s primary embedding dimension, commonly referred to as <code>d_model</code>.</p>Fixing GPU Operator Pods Stuck in Init: Secure Boot, DKMS, and MOK on Proxmox + Debian/posts/secure-boot-dkms-and-mok-on-proxmox-debian/Sat, 09 Aug 2025 00:00:00 +0000/posts/secure-boot-dkms-and-mok-on-proxmox-debian/<p>I hit an issue where all GPU Operator pods on one node were stuck in Init after migrating from Legacy BIOS to UEFI. The common error was NVIDIA components waiting for “toolkit-ready,” while the toolkit init container looped with:</p> <ul> <li>nvidia-smi failed to communicate with the NVIDIA driver</li> <li>modprobe nvidia → “Key was rejected by service”</li> diff --git a/posts/mixture-of-experts-moe-models-challenges-solutions-in-practice/index.html b/posts/mixture-of-experts-moe-models-challenges-solutions-in-practice/index.html index 66b5034..8f70bdb 100644 --- a/posts/mixture-of-experts-moe-models-challenges-solutions-in-practice/index.html +++ b/posts/mixture-of-experts-moe-models-challenges-solutions-in-practice/index.html @@ -44,4 +44,4 @@ The Top-K routing mechanism, as illustrated in the provided ima 2016 - 2025 Eric X. Liu -[c06e978] \ No newline at end of file +[0945204] \ No newline at end of file diff --git a/posts/secure-boot-dkms-and-mok-on-proxmox-debian/index.html b/posts/secure-boot-dkms-and-mok-on-proxmox-debian/index.html index 13e11d3..56a0ae6 100644 --- a/posts/secure-boot-dkms-and-mok-on-proxmox-debian/index.html +++ b/posts/secure-boot-dkms-and-mok-on-proxmox-debian/index.html @@ -59,4 +59,4 @@ nvidia-smi failed to communicate with the NVIDIA driver modprobe nvidia → “K 2016 - 2025 Eric X. Liu -[c06e978] \ No newline at end of file +[0945204] \ No newline at end of file diff --git a/posts/supabase-deep-dive/index.html b/posts/supabase-deep-dive/index.html index 6fec890..d4b271c 100644 --- a/posts/supabase-deep-dive/index.html +++ b/posts/supabase-deep-dive/index.html @@ -90,4 +90,4 @@ Supabase enters this space with a radically different philosophy: transparency. 2016 - 2025 Eric X. Liu -[c06e978] \ No newline at end of file +[0945204] \ No newline at end of file diff --git a/posts/t5-the-transformer-that-zigged-when-others-zagged-an-architectural-deep-dive/index.html b/posts/t5-the-transformer-that-zigged-when-others-zagged-an-architectural-deep-dive/index.html index 55ebf8a..6bba40d 100644 --- a/posts/t5-the-transformer-that-zigged-when-others-zagged-an-architectural-deep-dive/index.html +++ b/posts/t5-the-transformer-that-zigged-when-others-zagged-an-architectural-deep-dive/index.html @@ -30,4 +30,4 @@ But to truly understand the field, we must look at the pivotal models that explo 2016 - 2025 Eric X. Liu -[c06e978] \ No newline at end of file +[0945204] \ No newline at end of file diff --git a/posts/useful/index.html b/posts/useful/index.html index ddd3c28..de320ed 100644 --- a/posts/useful/index.html +++ b/posts/useful/index.html @@ -9,4 +9,4 @@ One-minute read
    • [c06e978] \ No newline at end of file +[0945204] \ No newline at end of file diff --git a/sitemap.xml b/sitemap.xml index d382b1e..f5929cb 100644 --- a/sitemap.xml +++ b/sitemap.xml @@ -1 +1 @@ -/posts/a-comprehensive-guide-to-breville-barista-pro-maintenance/2025-08-20T04:16:22+00:00weekly0.5/2025-08-20T04:16:22+00:00weekly0.5/posts/2025-08-20T04:16:22+00:00weekly0.5/posts/secure-boot-dkms-and-mok-on-proxmox-debian/2025-08-14T06:50:22+00:00weekly0.5/posts/how-rvq-teaches-llms-to-see-and-hear/2025-08-08T17:36:52+00:00weekly0.5/posts/supabase-deep-dive/2025-08-04T03:59:37+00:00weekly0.5/posts/a-deep-dive-into-ppo-for-language-models/2025-08-16T21:13:18+00:00weekly0.5/posts/mixture-of-experts-moe-models-challenges-solutions-in-practice/2025-08-03T06:02:48+00:00weekly0.5/posts/t5-the-transformer-that-zigged-when-others-zagged-an-architectural-deep-dive/2025-08-03T03:41:10+00:00weekly0.5/posts/espresso-theory-application-a-guide-for-the-breville-barista-pro/2025-08-03T04:20:20+00:00weekly0.5/posts/useful/2025-08-03T08:37:28-07:00weekly0.5/about/2020-06-16T23:30:17-07:00weekly0.5/categories/weekly0.5/tags/weekly0.5 \ No newline at end of file +/posts/a-comprehensive-guide-to-breville-barista-pro-maintenance/2025-08-20T04:32:59+00:00weekly0.5/posts/a-technical-deep-dive-into-the-transformer-s-core-mechanics/2025-08-20T04:32:59+00:00weekly0.5/2025-08-20T04:32:59+00:00weekly0.5/posts/2025-08-20T04:32:59+00:00weekly0.5/posts/secure-boot-dkms-and-mok-on-proxmox-debian/2025-08-14T06:50:22+00:00weekly0.5/posts/how-rvq-teaches-llms-to-see-and-hear/2025-08-08T17:36:52+00:00weekly0.5/posts/supabase-deep-dive/2025-08-04T03:59:37+00:00weekly0.5/posts/a-deep-dive-into-ppo-for-language-models/2025-08-16T21:13:18+00:00weekly0.5/posts/mixture-of-experts-moe-models-challenges-solutions-in-practice/2025-08-03T06:02:48+00:00weekly0.5/posts/t5-the-transformer-that-zigged-when-others-zagged-an-architectural-deep-dive/2025-08-03T03:41:10+00:00weekly0.5/posts/espresso-theory-application-a-guide-for-the-breville-barista-pro/2025-08-03T04:20:20+00:00weekly0.5/posts/useful/2025-08-03T08:37:28-07:00weekly0.5/about/2020-06-16T23:30:17-07:00weekly0.5/categories/weekly0.5/tags/weekly0.5 \ No newline at end of file diff --git a/tags/index.html b/tags/index.html index 51d4fe0..a692fb4 100644 --- a/tags/index.html +++ b/tags/index.html @@ -4,4 +4,4 @@ 2016 - 2025 Eric X. Liu -[c06e978] \ No newline at end of file +[0945204] \ No newline at end of file