pith. machine review for the scientific record. sign in

arxiv: 2605.05400 · v1 · submitted 2026-05-06 · 💻 cs.SE · cs.AI· cs.HC

Recognition: unknown

Mise en Place for Agentic Coding: Deliberate Preparation as Context Engineering Methodology

Andrew Zigler

Authors on Pith no claims yet

Pith reviewed 2026-05-08 15:54 UTC · model grok-4.3

classification 💻 cs.SE cs.AIcs.HC
keywords agentic codingmise en placecontext engineeringAI coding agentspreparation methodologycontext fluencysoftware developmenthackathon case
0
0 comments X

The pith

Two hours of deliberate preparation lets concurrent AI agents build a full-stack platform with aligned code.

A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.

The paper argues that the dominant vibe coding workflow with AI agents creates misalignment because agents receive insufficient context and generate code that demands heavy debugging and refactoring. It proposes a three-phase mise en place methodology to address this by first externalizing domain knowledge into structured documents, then using human-agent dialogue to produce detailed designs, and finally converting those into dependency-aware task records. In one hackathon application, the approach allowed rapid parallel work on an educational platform. The work also names context fluency as a developer skill and ties the method to backward design and tacit knowledge externalization while calling for systematic tests of preparation techniques in AI-assisted development.

Core claim

The central claim is that a three-phase preparation methodology consisting of contextual grounding to capture domain expertise, collaborative specification through dialogue to create design artifacts, and task decomposition into structured dependency-aware records supplies AI coding agents with the context needed for aligned, efficient output. This was shown when roughly two hours of such preparation enabled concurrent agents to implement a full-stack educational platform rapidly during a competitive hackathon.

What carries the argument

The mise en place (MEP) methodology, a three-phase process that externalizes tacit knowledge into documents, refines specifications via dialogue, and decomposes work into dependency-aware tasks to give agents sufficient context for implementation.

Where Pith is reading between the lines

These are editorial extensions of the paper, not claims the author makes directly.

  • The same preparation structure could be tested on non-coding agent tasks such as data pipeline construction or research synthesis to see if context externalization improves results there.
  • Integration of the three phases into agent tooling might automate parts of contextual grounding and task tracking for larger projects.
  • The emphasis on dependency-aware task records suggests potential for agent teams to self-coordinate once initial decomposition is complete.
  • Developers without AI agents might still benefit from the externalization steps when handing off work to human collaborators.

Load-bearing premise

The success in the single hackathon case came from applying the mise en place preparation rather than from the particular project chosen or the capabilities of the agents involved.

What would settle it

A controlled comparison in which a team using only vibe coding completes a similar full-stack project in comparable time and with similar quality would show that the preparation phase is not necessary for the reported outcome.

Figures

Figures reproduced from arXiv: 2605.05400 by Andrew Zigler.

Figure 1
Figure 1. Figure 1: The MEP methodology. Three sequential prepara view at source ↗
Figure 2
Figure 2. Figure 2: Per-bead completion time distribution across the 43 view at source ↗
read the original abstract

The rapid adoption of AI coding agents has produced a dominant workflow pattern -- often called "vibe coding" -- that prioritizes speed of implementation over deliberate preparation. We argue that this approach creates a systematic alignment problem: agents that lack sufficient context produce code requiring extensive debugging and refactoring, consuming substantial development time. Drawing on the culinary concept of mise en place (everything in its place; abbreviated MEP), we propose a three-phase preparation methodology for agentic coding: (1) contextual grounding, where domain expertise and tacit knowledge are externalized into structured documents; (2) collaborative specification, where human-agent dialogue produces detailed design artifacts; and (3) task decomposition, where specifications are converted into structured, dependency-aware task records. We report on the application of MEP during a competitive hackathon, where roughly two hours of preparation enabled a rapid parallel implementation of a full-stack educational platform by concurrent AI agents. We introduce the concept of context fluency as an emerging developer skill -- the ability to create rich, structured context that agents can act on -- and connect it to established frameworks in backward design and tacit knowledge externalization. We conclude with a research agenda for empirically validating preparation-phase methodologies in AI-assisted software development.

Editorial analysis

A structured set of objections, weighed in public.

Desk editor's note, referee report, simulated authors' rebuttal, and a circularity audit. Tearing a paper down is the easy half of reading it; the pith above is the substance, this is the friction.

Referee Report

1 major / 1 minor

Summary. The paper argues that 'vibe coding' with AI agents produces misaligned code due to insufficient context and proposes a three-phase 'Mise en Place' (MEP) methodology—contextual grounding (externalizing domain knowledge), collaborative specification (human-agent design artifacts), and task decomposition (dependency-aware tasks)—to address it. Drawing on culinary and educational concepts, it introduces 'context fluency' as a developer skill, reports an anecdotal application during a competitive hackathon in which two hours of MEP preparation enabled concurrent AI agents to build a full-stack educational platform, and outlines a research agenda for empirical validation of preparation-focused workflows in agentic coding.

Significance. If validated, the MEP framing could usefully shift AI-assisted software engineering practice toward deliberate context engineering, potentially reducing downstream debugging costs and improving agent reliability in complex tasks. The conceptual linkage to backward design and tacit-knowledge externalization is a constructive contribution that situates the work within established SE and education literatures. However, the current single-case anecdotal report provides no quantitative baselines, metrics, or controls, so the practical significance remains speculative pending systematic evaluation.

major comments (1)
  1. [Hackathon application section] Hackathon application section: the central claim that 'roughly two hours of preparation enabled a rapid parallel implementation' rests on a single qualitative report without any quantitative metrics (lines of code, test coverage, wall-clock time, defect rates), baseline condition (e.g., same task without MEP), or controls for confounds such as task choice, underlying model capabilities, or operator skill. This prevents isolation of MEP's causal contribution and leaves alternative explanations equally plausible.
minor comments (1)
  1. [Abstract and Conclusion] The abstract and conclusion could more explicitly qualify the generalizability of the single hackathon observation and state that the reported outcome is illustrative rather than confirmatory.

Simulated Author's Rebuttal

1 responses · 0 unresolved

We thank the referee for their constructive comments on our manuscript. We agree that the hackathon application is presented as a qualitative demonstration and have revised the text to more explicitly articulate its limitations while clarifying the intended contribution of the work.

read point-by-point responses
  1. Referee: [Hackathon application section] Hackathon application section: the central claim that 'roughly two hours of preparation enabled a rapid parallel implementation' rests on a single qualitative report without any quantitative metrics (lines of code, test coverage, wall-clock time, defect rates), baseline condition (e.g., same task without MEP), or controls for confounds such as task choice, underlying model capabilities, or operator skill. This prevents isolation of MEP's causal contribution and leaves alternative explanations equally plausible.

    Authors: We agree that the hackathon report is a single qualitative case without quantitative metrics, baselines, or controls, and therefore cannot support causal claims about MEP's isolated effect. The manuscript presents this experience as an illustrative application of the proposed methodology in a competitive setting, not as a controlled evaluation. We have revised the hackathon section to state these limitations more explicitly, to avoid implying causal efficacy, and to reinforce that the primary contribution is the three-phase methodology together with the research agenda for future empirical validation. revision: yes

Circularity Check

0 steps flagged

No circularity: conceptual proposal with single observational case

full rationale

The manuscript advances a three-phase preparation methodology (contextual grounding, collaborative specification, task decomposition) drawn from the culinary mise en place concept and reports its use in one hackathon. No equations, fitted parameters, predictions, or derivations appear. The central claim is an observational report of preparation enabling parallel agent work; it does not reduce to self-definition, fitted inputs renamed as predictions, or load-bearing self-citations. Connections to backward design and tacit knowledge are external references, not internal reductions. The paper is self-contained as a methodology proposal.

Axiom & Free-Parameter Ledger

0 free parameters · 2 axioms · 1 invented entities

The central claim depends on the effectiveness of the proposed phases, which are grounded in assumptions about knowledge externalization and collaboration rather than prior empirical data.

axioms (2)
  • domain assumption Domain expertise and tacit knowledge can be effectively externalized into structured documents for AI agents.
    This underpins the contextual grounding phase.
  • domain assumption Human-agent dialogue can produce detailed and actionable design artifacts.
    Basis for collaborative specification phase.
invented entities (1)
  • context fluency no independent evidence
    purpose: A developer skill for creating rich, structured context that AI agents can effectively use.
    Newly introduced term without prior empirical validation in the abstract.

pith-pipeline@v0.9.0 · 5509 in / 1378 out tokens · 58740 ms · 2026-05-08T15:54:53.129350+00:00 · methodology

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Reference graph

Works this paper leans on

22 extracted references · 6 canonical work pages

  1. [1]

    Eman Abdullah Alomar et al. 2025. Promptware Engineering: The New Frontier of Software Engineering.arXiv preprintarXiv:2503.02400 (2025). doi:10.48550/ arXiv.2503.02400

  2. [2]

    Anthropic. 2025. Effective Context Engineering for AI Agents. Anthropic Doc- umentation. https://docs.anthropic.com/en/docs/build-with-claude/context- engineering

  3. [3]

    Brown, Benjamin Mann, Nick Ryder, Melanie Subbiah, Jared Kaplan, Prafulla Dhariwal, Arvind Neelakantan, Pranav Shyam, Girish Sastry, Amanda Askell, et al

    Tom B. Brown, Benjamin Mann, Nick Ryder, Melanie Subbiah, Jared Kaplan, Prafulla Dhariwal, Arvind Neelakantan, Pranav Shyam, Girish Sastry, Amanda Askell, et al . 2020. Language Models are Few-Shot Learners. InAdvances in Neural Information Processing Systems (NeurIPS)

  4. [4]

    Arghavan Moradi Dakhel et al. 2025. A Taxonomy of Human-AI Collaboration in Software Engineering.arXiv preprintarXiv:2501.08774 (2025). doi:10.48550/ arXiv.2501.08774

  5. [5]

    Jeffrey Emanuel. 2025. beads_rust: High-Performance Task Tracking for AI Agents. GitHub repository. https://github.com/Dicklesworthstone/beads_rust

  6. [6]

    Gartner. 2025. Context Engineering Is In, Prompt Engineering Is Out. Gartner Research Note. Gartner identified context engineering as a key shift in AI application development for 2025

  7. [7]

    GitHub. 2025. Spec Kit: Specification-Driven Development for AI Agents. GitHub repository. https://github.com/github/spec-kit

  8. [8]

    Hacks/Hackers and The Atlantic and Infactory. 2026. Building Future AI News Experiences with The Atlantic and Infactory. Hackathon event description. Approximately 12 teams, five-hour build window, $5,000 prize for top project

  9. [9]

    Dex Horthy. 2025. The RPI Methodology: Research, Plan, Implement for Agentic Coding. HumanLayer Blog. https://humanlayer.dev/blog/rpi-methodology

  10. [10]

    Geoffrey Huntley. 2025. The Ralph Loop: A Pattern for AI Agent Iteration. Blog post. https://ghuntley.com/ralph/

  11. [11]

    Andrej Karpathy. 2025. Vibe Coding. Twitter/X post. https://x.com/karpathy/ status/1886192184808149383 Term entered Merriam-Webster and Collins Dictio- nary in 2025

  12. [12]

    Pengfei Liu, Weizhe Yuan, Jinlan Fu, Zhengbao Jiang, Hiroaki Hayashi, and Graham Neubig. 2023. Pre-train, Prompt, and Predict: A Systematic Survey of Prompting Methods in Natural Language Processing.Comput. Surveys55, 9 (2023), 1–35. doi:10.1145/3560815

  13. [13]

    Hussein Mozannar et al . 2023. Reading Between the Lines: Modeling User Behavior and Costs in AI-Assisted Programming.arXiv preprintarXiv:2210.14306 (2023). doi:10.48550/arXiv.2210.14306

  14. [14]

    1995.The Knowledge-Creating Company: How Japanese Companies Create the Dynamics of Innovation

    Ikujiro Nonaka and Hirotaka Takeuchi. 1995.The Knowledge-Creating Company: How Japanese Companies Create the Dynamics of Innovation. Oxford University Press, New York

  15. [15]

    Sida Peng, Eirini Kalliamvakou, Peter Cihon, and Mert Demirer. 2023. The Impact of AI on Developer Productivity: Evidence from GitHub Copilot.arXiv preprint arXiv:2302.06590 (2023). doi:10.48550/arXiv.2302.06590

  16. [16]

    1958.Personal Knowledge: Towards a Post-Critical Philosophy

    Michael Polanyi. 1958.Personal Knowledge: Towards a Post-Critical Philosophy. University of Chicago Press, Chicago, IL

  17. [17]

    1966.The Tacit Dimension

    Michael Polanyi. 1966.The Tacit Dimension. University of Chicago Press, Chicago, IL

  18. [18]

    Vera Liao

    Marisa Vasconcelos, Jack Jamieson, Umang Bhatt, and Q. Vera Liao. 2024. Un- derstanding Trust in AI-Assisted Code Generation. InProceedings of the ACM Conference on Fairness, Accountability, and Transparency (FAccT). ACM, New York, NY

  19. [19]

    Veracode. 2025. State of Software Security: The Rise of AI Code. Industry report. Reports that 45% of AI-generated code contains security flaws, with a 10x spike in security findings from AI-generated code by June 2025

  20. [20]

    Chi, Quoc V

    Jason Wei, Xuezhi Wang, Dale Schuurmans, Maarten Bosma, Brian Ichter, Fei Xia, Ed H. Chi, Quoc V. Le, and Denny Zhou. 2022. Chain-of-Thought Prompting Elicits Reasoning in Large Language Models. InAdvances in Neural Information Processing Systems (NeurIPS)

  21. [21]

    1998.Understanding by Design

    Grant Wiggins and Jay McTighe. 1998.Understanding by Design. Association for Supervision and Curriculum Development, Alexandria, VA

  22. [22]

    Steve Yegge. 2025. Beads: External Memory for AI Agents. GitHub repository. https://github.com/steveyegge/beads