Recognition: unknown
Mise en Place for Agentic Coding: Deliberate Preparation as Context Engineering Methodology
Pith reviewed 2026-05-08 15:54 UTC · model grok-4.3
The pith
Two hours of deliberate preparation lets concurrent AI agents build a full-stack platform with aligned code.
A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.
Core claim
The central claim is that a three-phase preparation methodology consisting of contextual grounding to capture domain expertise, collaborative specification through dialogue to create design artifacts, and task decomposition into structured dependency-aware records supplies AI coding agents with the context needed for aligned, efficient output. This was shown when roughly two hours of such preparation enabled concurrent agents to implement a full-stack educational platform rapidly during a competitive hackathon.
What carries the argument
The mise en place (MEP) methodology, a three-phase process that externalizes tacit knowledge into documents, refines specifications via dialogue, and decomposes work into dependency-aware tasks to give agents sufficient context for implementation.
Where Pith is reading between the lines
- The same preparation structure could be tested on non-coding agent tasks such as data pipeline construction or research synthesis to see if context externalization improves results there.
- Integration of the three phases into agent tooling might automate parts of contextual grounding and task tracking for larger projects.
- The emphasis on dependency-aware task records suggests potential for agent teams to self-coordinate once initial decomposition is complete.
- Developers without AI agents might still benefit from the externalization steps when handing off work to human collaborators.
Load-bearing premise
The success in the single hackathon case came from applying the mise en place preparation rather than from the particular project chosen or the capabilities of the agents involved.
What would settle it
A controlled comparison in which a team using only vibe coding completes a similar full-stack project in comparable time and with similar quality would show that the preparation phase is not necessary for the reported outcome.
Figures
read the original abstract
The rapid adoption of AI coding agents has produced a dominant workflow pattern -- often called "vibe coding" -- that prioritizes speed of implementation over deliberate preparation. We argue that this approach creates a systematic alignment problem: agents that lack sufficient context produce code requiring extensive debugging and refactoring, consuming substantial development time. Drawing on the culinary concept of mise en place (everything in its place; abbreviated MEP), we propose a three-phase preparation methodology for agentic coding: (1) contextual grounding, where domain expertise and tacit knowledge are externalized into structured documents; (2) collaborative specification, where human-agent dialogue produces detailed design artifacts; and (3) task decomposition, where specifications are converted into structured, dependency-aware task records. We report on the application of MEP during a competitive hackathon, where roughly two hours of preparation enabled a rapid parallel implementation of a full-stack educational platform by concurrent AI agents. We introduce the concept of context fluency as an emerging developer skill -- the ability to create rich, structured context that agents can act on -- and connect it to established frameworks in backward design and tacit knowledge externalization. We conclude with a research agenda for empirically validating preparation-phase methodologies in AI-assisted software development.
Editorial analysis
A structured set of objections, weighed in public.
Referee Report
Summary. The paper argues that 'vibe coding' with AI agents produces misaligned code due to insufficient context and proposes a three-phase 'Mise en Place' (MEP) methodology—contextual grounding (externalizing domain knowledge), collaborative specification (human-agent design artifacts), and task decomposition (dependency-aware tasks)—to address it. Drawing on culinary and educational concepts, it introduces 'context fluency' as a developer skill, reports an anecdotal application during a competitive hackathon in which two hours of MEP preparation enabled concurrent AI agents to build a full-stack educational platform, and outlines a research agenda for empirical validation of preparation-focused workflows in agentic coding.
Significance. If validated, the MEP framing could usefully shift AI-assisted software engineering practice toward deliberate context engineering, potentially reducing downstream debugging costs and improving agent reliability in complex tasks. The conceptual linkage to backward design and tacit-knowledge externalization is a constructive contribution that situates the work within established SE and education literatures. However, the current single-case anecdotal report provides no quantitative baselines, metrics, or controls, so the practical significance remains speculative pending systematic evaluation.
major comments (1)
- [Hackathon application section] Hackathon application section: the central claim that 'roughly two hours of preparation enabled a rapid parallel implementation' rests on a single qualitative report without any quantitative metrics (lines of code, test coverage, wall-clock time, defect rates), baseline condition (e.g., same task without MEP), or controls for confounds such as task choice, underlying model capabilities, or operator skill. This prevents isolation of MEP's causal contribution and leaves alternative explanations equally plausible.
minor comments (1)
- [Abstract and Conclusion] The abstract and conclusion could more explicitly qualify the generalizability of the single hackathon observation and state that the reported outcome is illustrative rather than confirmatory.
Simulated Author's Rebuttal
We thank the referee for their constructive comments on our manuscript. We agree that the hackathon application is presented as a qualitative demonstration and have revised the text to more explicitly articulate its limitations while clarifying the intended contribution of the work.
read point-by-point responses
-
Referee: [Hackathon application section] Hackathon application section: the central claim that 'roughly two hours of preparation enabled a rapid parallel implementation' rests on a single qualitative report without any quantitative metrics (lines of code, test coverage, wall-clock time, defect rates), baseline condition (e.g., same task without MEP), or controls for confounds such as task choice, underlying model capabilities, or operator skill. This prevents isolation of MEP's causal contribution and leaves alternative explanations equally plausible.
Authors: We agree that the hackathon report is a single qualitative case without quantitative metrics, baselines, or controls, and therefore cannot support causal claims about MEP's isolated effect. The manuscript presents this experience as an illustrative application of the proposed methodology in a competitive setting, not as a controlled evaluation. We have revised the hackathon section to state these limitations more explicitly, to avoid implying causal efficacy, and to reinforce that the primary contribution is the three-phase methodology together with the research agenda for future empirical validation. revision: yes
Circularity Check
No circularity: conceptual proposal with single observational case
full rationale
The manuscript advances a three-phase preparation methodology (contextual grounding, collaborative specification, task decomposition) drawn from the culinary mise en place concept and reports its use in one hackathon. No equations, fitted parameters, predictions, or derivations appear. The central claim is an observational report of preparation enabling parallel agent work; it does not reduce to self-definition, fitted inputs renamed as predictions, or load-bearing self-citations. Connections to backward design and tacit knowledge are external references, not internal reductions. The paper is self-contained as a methodology proposal.
Axiom & Free-Parameter Ledger
axioms (2)
- domain assumption Domain expertise and tacit knowledge can be effectively externalized into structured documents for AI agents.
- domain assumption Human-agent dialogue can produce detailed and actionable design artifacts.
invented entities (1)
-
context fluency
no independent evidence
Reference graph
Works this paper leans on
- [1]
-
[2]
Anthropic. 2025. Effective Context Engineering for AI Agents. Anthropic Doc- umentation. https://docs.anthropic.com/en/docs/build-with-claude/context- engineering
2025
-
[3]
Brown, Benjamin Mann, Nick Ryder, Melanie Subbiah, Jared Kaplan, Prafulla Dhariwal, Arvind Neelakantan, Pranav Shyam, Girish Sastry, Amanda Askell, et al
Tom B. Brown, Benjamin Mann, Nick Ryder, Melanie Subbiah, Jared Kaplan, Prafulla Dhariwal, Arvind Neelakantan, Pranav Shyam, Girish Sastry, Amanda Askell, et al . 2020. Language Models are Few-Shot Learners. InAdvances in Neural Information Processing Systems (NeurIPS)
2020
- [4]
-
[5]
Jeffrey Emanuel. 2025. beads_rust: High-Performance Task Tracking for AI Agents. GitHub repository. https://github.com/Dicklesworthstone/beads_rust
2025
-
[6]
Gartner. 2025. Context Engineering Is In, Prompt Engineering Is Out. Gartner Research Note. Gartner identified context engineering as a key shift in AI application development for 2025
2025
-
[7]
GitHub. 2025. Spec Kit: Specification-Driven Development for AI Agents. GitHub repository. https://github.com/github/spec-kit
2025
-
[8]
Hacks/Hackers and The Atlantic and Infactory. 2026. Building Future AI News Experiences with The Atlantic and Infactory. Hackathon event description. Approximately 12 teams, five-hour build window, $5,000 prize for top project
2026
-
[9]
Dex Horthy. 2025. The RPI Methodology: Research, Plan, Implement for Agentic Coding. HumanLayer Blog. https://humanlayer.dev/blog/rpi-methodology
2025
-
[10]
Geoffrey Huntley. 2025. The Ralph Loop: A Pattern for AI Agent Iteration. Blog post. https://ghuntley.com/ralph/
2025
- [11]
-
[12]
Pengfei Liu, Weizhe Yuan, Jinlan Fu, Zhengbao Jiang, Hiroaki Hayashi, and Graham Neubig. 2023. Pre-train, Prompt, and Predict: A Systematic Survey of Prompting Methods in Natural Language Processing.Comput. Surveys55, 9 (2023), 1–35. doi:10.1145/3560815
-
[13]
Hussein Mozannar et al . 2023. Reading Between the Lines: Modeling User Behavior and Costs in AI-Assisted Programming.arXiv preprintarXiv:2210.14306 (2023). doi:10.48550/arXiv.2210.14306
-
[14]
1995.The Knowledge-Creating Company: How Japanese Companies Create the Dynamics of Innovation
Ikujiro Nonaka and Hirotaka Takeuchi. 1995.The Knowledge-Creating Company: How Japanese Companies Create the Dynamics of Innovation. Oxford University Press, New York
1995
-
[15]
Sida Peng, Eirini Kalliamvakou, Peter Cihon, and Mert Demirer. 2023. The Impact of AI on Developer Productivity: Evidence from GitHub Copilot.arXiv preprint arXiv:2302.06590 (2023). doi:10.48550/arXiv.2302.06590
-
[16]
1958.Personal Knowledge: Towards a Post-Critical Philosophy
Michael Polanyi. 1958.Personal Knowledge: Towards a Post-Critical Philosophy. University of Chicago Press, Chicago, IL
1958
-
[17]
1966.The Tacit Dimension
Michael Polanyi. 1966.The Tacit Dimension. University of Chicago Press, Chicago, IL
1966
-
[18]
Vera Liao
Marisa Vasconcelos, Jack Jamieson, Umang Bhatt, and Q. Vera Liao. 2024. Un- derstanding Trust in AI-Assisted Code Generation. InProceedings of the ACM Conference on Fairness, Accountability, and Transparency (FAccT). ACM, New York, NY
2024
-
[19]
Veracode. 2025. State of Software Security: The Rise of AI Code. Industry report. Reports that 45% of AI-generated code contains security flaws, with a 10x spike in security findings from AI-generated code by June 2025
2025
-
[20]
Chi, Quoc V
Jason Wei, Xuezhi Wang, Dale Schuurmans, Maarten Bosma, Brian Ichter, Fei Xia, Ed H. Chi, Quoc V. Le, and Denny Zhou. 2022. Chain-of-Thought Prompting Elicits Reasoning in Large Language Models. InAdvances in Neural Information Processing Systems (NeurIPS)
2022
-
[21]
1998.Understanding by Design
Grant Wiggins and Jay McTighe. 1998.Understanding by Design. Association for Supervision and Curriculum Development, Alexandria, VA
1998
-
[22]
Steve Yegge. 2025. Beads: External Memory for AI Agents. GitHub repository. https://github.com/steveyegge/beads
2025
discussion (0)
Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.