Context Management Flashcards

(14 cards)

1
Q

Context Poisoning

A
  • Misinformation in the context (e.g., goals, summary)
  • can cause the model to become fixated on impossible or irrelevant objectives and lead to nonsensical strategies.
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Context Distraction

A
  • The context grows so long that the model over-focuses on history, “forgetting” what it learned in training.
  • Results in repeating past actions rather than synthesizing novel plans.
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Context Confusion

A
  • Irrelevant content forces the model to pay attention to it
  • leading to low-quality responses
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Context Quarantine

A

Isolating contexts in dedicated threads, often facilitated by subagents, to explore different aspects of a question in parallel and then condense the most important tokens for a lead agent.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Context Clash

A

New information or tools in the context directly conflict with other existing information, derailing the model’s reasoning.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Context Pruning

A
  • Actively removing irrelevant or unneeded information from the context
  • score sentences or chunks of conversation based on their relevance to the current goal or user query. Irrelevant or low-scoring content is dynamically removed.
  • use provence
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

Context Summarization

A

Using a separate LLM to condense conversation history.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Context Offloading

A

Storing information outside the LLM’s active context via a tool, often referred to as a “scratchpad”. Particularly effective for detailed tool output analysis, policy-heavy environments, and sequential decision-making.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

What is provence

A
  • lightweight context pruning model for RAG
  • fast, simple to use, small (1.75GB)
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Why should you use a map for conversation history?

A
  • Maintain a map of context for easier pruning.
  • separate instructions and goals from conversation history
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

How many tokens does it take for a model to experience context distraction?

A

Beyond 100k tokens in some cases, or around 32k for smaller models like Llama 3.1-405b)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Rolling Window

A
  • Truncation
  • This simple method involves keeping only the most recent ‘N’ messages or a fixed number of tokens, discarding older information entirely once the context window limit is reached.
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

RAG as a context management strategy

A

Rather than keeping all history in the active context, details are offloaded to an external database (like a vector database). When information is needed, the system retrieves only the necessary, relevant data based on the current query, ensuring the active context remains lean and focused.
Hierarchical Memory Management: This approach uses a tiered system where critical information (e.g., system instructions, key decisions) is preserved, recent history is kept in detail, and older or less relevant content is summarized or removed entirely.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Hierarchical Memory Management

A

This approach uses a tiered system where critical information (e.g., system instructions, key decisions) is preserved, recent history is kept in detail, and older or less relevant content is summarized or removed entirely)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly