Skip to content

Active Recall Session: Prompt Engineering — Instruction-following vs completion behavior #110

@github-actions

Description

@github-actions

Learning Objective

“What do I currently believe, and how could that belief be wrong?”

Bucket: Prompt Engineering
Focus: Instruction-following vs completion behavior
Date: 2025-12-26

(What system, concept, or failure surface am I reconstructing from memory?)


Active Recall (NO NOTES)

1. Definition + Boundary

Instruction following is the guidance given for handling the input, and completion behavior is whether the output should be incrementally given or given all at once at the completion.


2. Failure Statement

The system fails when the output does not align with the instructions pattern the model is expected to follow, nor does the model comply with the way that the completion should be given.


3. Mental Model

(Text diagram or Mermaid showing components, data flow, uncertainty, feedback)

Input -> instructions from prompt -> encode -> gather response -> decode -> completion -> output


4. Mechanism (Causal Chain)

(Write 3–5 linked causal statements explaining why this works)

  1. proper instructions set the boundaries that the model is expected to follow
  2. completion behavior is based on the expectations of the user group
  3. instructions give context to the model
  4. incremental completion can work well with chain-of-thought prompting

5. Constraints & Trade-offs

  • Constraint 1: the amount of information the model is expected to output can impact how precise the instructions are
  • Constraint 2: if the completion behavior is to wait until the full response is ready, it is important that the response does not take long, otherwise it can turn into a bad user experience.
  • Constraint 3: instructions that are ambiguous can lead to poor performance in being able to complete in a timely and accurate manner

6. Transfer Test

(How would this idea change in another domain or system?)

Could be similar to how telephone operators used to work. Give the operator incomplete information and the operator has bad instructions on how to find the information, the response time could take some time.


Self-Assessment (Rubric)

Dimension Score (0–4) Notes
Conceptual Clarity 3
Mental Model Integrity 3
Causal Understanding 4
Failure Awareness 2
Trade-off Judgment 1
Transfer Ability 3
Calibration & Honesty 3

Initial Confidence (0–100%): 80


Falsification Plan

(One experiment or eval that could prove me wrong)

A known bad prompt that immediately returns output response that is very accurate


Carry-Forward Insight

(One sentence for Future Me)


Delayed Recall (Fill Later)

  • What did I forget?
  • What was oversimplified?
  • What was wrong?

Completion Checklist

  • Explained aloud without notes
  • Identified ≥1 real failure mode
  • Made a falsifiable claim
  • Scored honestly

Confidence Delta Reflection (Fill After Review)

  • Initial confidence:
  • Reviewer signal (over / under / calibrated):
  • My assessment:
  • What I will adjust next time:

Metadata

Metadata

Assignees

Projects

No projects

Milestone

No milestone

Relationships

None yet

Development

No branches or pull requests

Issue actions