Exploiting Memorization: Understanding the CLM Objective for Knowledge Encoding in LLMs

This article provides the necessary background and notation for reasoning research, defining problems as tuples


This content originally appeared on HackerNoon and was authored by The Tech Reckoning is Upon Us!

Abstract and 1. Introduction

  1. Background

  2. Method

  3. Experiments

    4.1 Multi-hop Reasoning Performance

    4.2 Reasoning with Distractors

    4.3 Generalization to Real-World knowledge

    4.4 Run-time Analysis

    4.5 Memorizing Knowledge

  4. Related Work

  5. Conclusion, Acknowledgements, and References

\ A. Dataset

B. In-context Reasoning with Distractors

C. Implementation Details

D. Adaptive Learning Rate

E. Experiments with Large Language Models

2 Background

Notation We use f : X × θ → Y to refer to parameterised functions in which X is the set of possible inputs and θ are their possible weights (parameters). We use fθ : x 7→ f(x, θ) to easily refer to any f with a given set of parameters θ. We describe reasoning problems using tuples (K, x, y∗ , Y ) such that y ∈ Y is the correct answer for the question x given facts K, and use D to refer to sets of such problems. When it is clear from context, we drop Y and use only (K, x, y∗ ).

\ Language Modeling and Memorization In the causal language modeling (CLM) objective, a parameterized model fθ is trained to estimate the conditional probabilities of each token in a sequence given its predecessors: p(xt|x) Specifically, we train fθ to approximate p using the CLM loss:

\

\ This training objective allows language models to memorize individual training examples [10, 11], and we will exploit this ability to memorize and draw on contextual knowledge in our work.

\

\

:::info Authors:

(1) Zeming Chen, EPFL (zeming.chen@epfl.ch);

(2) Gail Weiss, EPFL (antoine.bosselut@epfl.ch);

(3) Eric Mitchell, Stanford University (eric.mitchell@cs.stanford.edu)';

(4) Asli Celikyilmaz, Meta AI Research (aslic@meta.com);

(5) Antoine Bosselut, EPFL (antoine.bosselut@epfl.ch).

:::


:::info This paper is available on arxiv under CC BY 4.0 DEED license.

:::

\


This content originally appeared on HackerNoon and was authored by The Tech Reckoning is Upon Us!


Print Share Comment Cite Upload Translate Updates
APA

The Tech Reckoning is Upon Us! | Sciencx (2025-10-24T00:49:47+00:00) Exploiting Memorization: Understanding the CLM Objective for Knowledge Encoding in LLMs. Retrieved from https://www.scien.cx/2025/10/24/exploiting-memorization-understanding-the-clm-objective-for-knowledge-encoding-in-llms/

MLA
" » Exploiting Memorization: Understanding the CLM Objective for Knowledge Encoding in LLMs." The Tech Reckoning is Upon Us! | Sciencx - Friday October 24, 2025, https://www.scien.cx/2025/10/24/exploiting-memorization-understanding-the-clm-objective-for-knowledge-encoding-in-llms/
HARVARD
The Tech Reckoning is Upon Us! | Sciencx Friday October 24, 2025 » Exploiting Memorization: Understanding the CLM Objective for Knowledge Encoding in LLMs., viewed ,<https://www.scien.cx/2025/10/24/exploiting-memorization-understanding-the-clm-objective-for-knowledge-encoding-in-llms/>
VANCOUVER
The Tech Reckoning is Upon Us! | Sciencx - » Exploiting Memorization: Understanding the CLM Objective for Knowledge Encoding in LLMs. [Internet]. [Accessed ]. Available from: https://www.scien.cx/2025/10/24/exploiting-memorization-understanding-the-clm-objective-for-knowledge-encoding-in-llms/
CHICAGO
" » Exploiting Memorization: Understanding the CLM Objective for Knowledge Encoding in LLMs." The Tech Reckoning is Upon Us! | Sciencx - Accessed . https://www.scien.cx/2025/10/24/exploiting-memorization-understanding-the-clm-objective-for-knowledge-encoding-in-llms/
IEEE
" » Exploiting Memorization: Understanding the CLM Objective for Knowledge Encoding in LLMs." The Tech Reckoning is Upon Us! | Sciencx [Online]. Available: https://www.scien.cx/2025/10/24/exploiting-memorization-understanding-the-clm-objective-for-knowledge-encoding-in-llms/. [Accessed: ]
rf:citation
» Exploiting Memorization: Understanding the CLM Objective for Knowledge Encoding in LLMs | The Tech Reckoning is Upon Us! | Sciencx | https://www.scien.cx/2025/10/24/exploiting-memorization-understanding-the-clm-objective-for-knowledge-encoding-in-llms/ |

Please log in to upload a file.




There are no updates yet.
Click the Upload button above to add an update.

You must be logged in to translate posts. Please log in or register.