[{"data":1,"prerenderedAt":-1},["ShallowReactive",2],{"article-why-prompt-engineering-isnt-engineering-en":3,"tags-why-prompt-engineering-isnt-engineering-en":30,"related-lang-why-prompt-engineering-isnt-engineering-en":41,"related-posts-why-prompt-engineering-isnt-engineering-en":45,"series-research-6407bc31-5ff2-48c3-9760-a1cd5c8518b0":82},{"id":4,"title":5,"content":6,"summary":7,"source":8,"source_url":9,"author":10,"image_url":11,"keywords":12,"language":18,"translated_content":10,"views":19,"is_premium":20,"created_at":21,"updated_at":21,"cover_image":11,"published_at":22,"rewrite_status":23,"rewrite_error":10,"rewritten_from_id":24,"slug":25,"category":26,"related_article_id":27,"status":28,"google_indexed_at":29,"x_posted_at":10,"tweet_text":10,"title_rewritten_at":10,"title_original":10,"key_takeaways":10,"topic_cluster_id":10,"embedding":10,"is_canonical_seed":20},"6407bc31-5ff2-48c3-9760-a1cd5c8518b0","Why Prompt Engineering Isn’t Engineering","\u003Cp>Microsoft says prompt design is “more of an art than a science,” and that line matters. If a practice depends on taste, iteration, and a lot of trial and error, calling it engineering gets messy fast.\u003C\u002Fp>\u003Cp>This matters because the stakes are getting higher. Teams are already using prompts to ship products, automate support, and make decisions, yet the field still lacks the kind of methods, tests, and standards engineers usually rely on.\u003C\u002Fp>\u003Ch2>What engineering actually means\u003C\u002Fh2>\u003Cp>Before anyone can argue about prompt engineering, it helps to define engineering in the boring, useful way: a discipline grounded in science, formal methods, repeatable design, and accountability. Bodies like \u003Ca href=\"https:\u002F\u002Fwww.abet.org\u002F\" target=\"_blank\" rel=\"noopener\">ABET\u003C\u002Fa>, \u003Ca href=\"https:\u002F\u002Fwww.ieee.org\u002F\" target=\"_blank\" rel=\"noopener\">IEEE\u003C\u002Fa>, and the \u003Ca href=\"https:\u002F\u002Fwww.nspe.org\u002F\" target=\"_blank\" rel=\"noopener\">National Society of Professional Engineers\u003C\u002Fa> all point to the same core ideas. Engineering is about building systems under constraints, with measurable outcomes and real consequences.\u003C\u002Fp>\n\u003Cfigure class=\"my-6\">\u003Cimg src=\"https:\u002F\u002Fxxdpdyhzhpamafnrdkyq.supabase.co\u002Fstorage\u002Fv1\u002Fobject\u002Fpublic\u002Fcovers\u002Finline-1775122511647-9b56.png\" alt=\"Why Prompt Engineering Isn’t Engineering\" class=\"rounded-xl w-full\" loading=\"lazy\" \u002F>\u003C\u002Ffigure>\n\u003Cp>That legal and ethical weight is not abstract. In Germany, misusing the title “engineer” can lead to prison time. In Canada, the penalties can reach $25,000. In many US states, “Professional Engineer” is a protected title tied to licensure and exams. The title exists because bad engineering can hurt people.\u003C\u002Fp>\u003Cp>Prompt work does not sit in that same bucket yet. It looks more like a mix of writing, experimentation, and product tuning. That does not make it useless. It just means the label matters.\u003C\u002Fp>\u003Cul>\u003Cli>Engineering relies on formal methods, testable claims, and repeatable results.\u003C\u002Fli>\u003Cli>Prompt guides usually say things like “be specific” or “provide context.”\u003C\u002Fli>\u003Cli>Microsoft’s own docs describe prompt design as “more of an art than a science.”\u003C\u002Fli>\u003Cli>Protected engineering titles carry legal consequences in several countries.\u003C\u002Fli>\u003C\u002Ful>\u003Cp>There is also a historical pattern worth remembering. Software engineering was named at the NATO conference in 1968, and even then participants admitted the term described a need more than a reality. Civil engineering took centuries to formalize. New fields can earn the title over time, but they do that by building methods, standards, and testing culture.\u003C\u002Fp>\u003Cp>That is the real question for prompt engineering: is it on that path, or is it just a convenient label for a set of heuristics? Right now, the evidence points to heuristics.\u003C\u002Fp>\u003Ch2>What the vendor guides actually say\u003C\u002Fh2>\u003Cp>The easiest place to check for engineering-grade guidance is the vendors themselves. If \u003Ca href=\"https:\u002F\u002Fopenai.com\u002F\" target=\"_blank\" rel=\"noopener\">OpenAI\u003C\u002Fa>, \u003Ca href=\"https:\u002F\u002Fwww.anthropic.com\u002F\" target=\"_blank\" rel=\"noopener\">Anthropic\u003C\u002Fa>, \u003Ca href=\"https:\u002F\u002Fai.google.dev\u002F\" target=\"_blank\" rel=\"noopener\">Google\u003C\u002Fa>, and \u003Ca href=\"https:\u002F\u002Fazure.microsoft.com\u002Fen-us\u002Fproducts\u002Fai-services\" target=\"_blank\" rel=\"noopener\">Microsoft Azure AI\u003C\u002Fa> are publishing prompt advice, you would expect clear criteria, measurable thresholds, and testable procedures. Instead, most of the advice is qualitative.\u003C\u002Fp>\u003Cp>Across roughly 25 recommendations in those guides, only about four include anything close to quantifiable criteria. The rest are variations on the same themes: be specific, give context, use examples, and iterate. That is useful advice for beginners, but it is not a method.\u003C\u002Fp>\u003Cblockquote>“Prompt engineering is more of an art than a science.” — Microsoft, Azure OpenAI Service documentation\u003C\u002Fblockquote>\u003Cp>That quote is honest, and it also gives the game away. Engineering needs more than advice that sounds right. It needs definitions that hold up under scrutiny. What counts as “specific enough”? How much context is enough? What is the acceptance test? What failure rate is acceptable? The guides rarely answer those questions.\u003C\u002Fp>\u003Cp>The pipeline also gets fuzzy as advice moves away from research. Academic papers may include methods, sample sizes, and measured outcomes. Vendor docs compress that into developer-friendly guidance. Blog posts and social media then compress it again into “top tips.” By the time most people see it, the method has been stripped out.\u003C\u002Fp>\u003Cul>\u003Cli>OpenAI, Anthropic, Google, and Microsoft all publish prompt guidance.\u003C\u002Fli>\u003Cli>Most recommendations are qualitative rather than measurable.\u003C\u002Fli>\u003Cli>Microsoft explicitly calls prompt design “more of an art than a science.”\u003C\u002Fli>\u003Cli>Natural language requirements remain ambiguous without formal syntax.\u003C\u002Fli>\u003C\u002Ful>\u003Cp>That last point matters because engineering already solved this problem in other places. The Internet Engineering Task Force created \u003Ca href=\"https:\u002F\u002Fwww.rfc-editor.org\u002Frfc\u002Frfc2119\" target=\"_blank\" rel=\"noopener\">RFC 2119\u003C\u002Fa> to define words like MUST, SHOULD, and MAY. The whole point was to reduce ambiguity in requirements. If prompt work were a mature engineering discipline, you would expect it to adopt tools like that more widely.\u003C\u002Fp>\u003Ch2>The testing problem nobody can hand-wave away\u003C\u002Fh2>\u003Cp>Engineering without testing is just opinion with a spreadsheet. In software, we have spent decades building test-driven development, continuous integration, coverage analysis, fuzzing, and property-based testing. Those systems work because the software is usually deterministic: same input, same output.\u003C\u002Fp>\n\u003Cfigure class=\"my-6\">\u003Cimg src=\"https:\u002F\u002Fxxdpdyhzhpamafnrdkyq.supabase.co\u002Fstorage\u002Fv1\u002Fobject\u002Fpublic\u002Fcovers\u002Finline-1775122499222-xl9h.png\" alt=\"Why Prompt Engineering Isn’t Engineering\" class=\"rounded-xl w-full\" loading=\"lazy\" \u002F>\u003C\u002Ffigure>\n\u003Cp>Large language models are different. The same prompt can produce different answers on different runs. That means prompt testing has to be statistical, not just anecdotal. You need golden datasets, repeated trials, confidence intervals, and regression baselines. That is a much harder problem, and it requires more statistical literacy than most prompt guides admit.\u003C\u002Fp>\u003Cp>There are tools trying to fill the gap. \u003Ca href=\"https:\u002F\u002Fwww.promptfoo.dev\u002F\" target=\"_blank\" rel=\"noopener\">Promptfoo\u003C\u002Fa>, \u003Ca href=\"https:\u002F\u002Fwww.helicone.ai\u002F\" target=\"_blank\" rel=\"noopener\">Helicone\u003C\u002Fa>, \u003Ca href=\"https:\u002F\u002Fwww.langchain.com\u002Flangsmith\" target=\"_blank\" rel=\"noopener\">LangSmith\u003C\u002Fa>, and \u003Ca href=\"https:\u002F\u002Fwww.deepeval.com\u002F\" target=\"_blank\" rel=\"noopener\">DeepEval\u003C\u002Fa> all help teams evaluate prompts and model outputs. That is progress, but it is still early. The tooling is uneven, and the methods are not yet standardized across the industry.\u003C\u002Fp>\u003Cp>Regulated industries are already warning about the gap. The FAA has said that rigorous safety assurance methods must be developed for AI systems in aviation. The Federal Reserve’s SR 11-7 guidance says adaptive AI models may lose effectiveness over time. Those are not academic footnotes. They are institutions that deal with systems where failure has a price tag.\u003C\u002Fp>\u003Cul>\u003Cli>Traditional software testing assumes deterministic behavior.\u003C\u002Fli>\u003Cli>LLM evaluation needs repeated trials and statistical thresholds.\u003C\u002Fli>\u003Cli>Promptfoo, Helicone, LangSmith, and DeepEval are still maturing.\u003C\u002Fli>\u003Cli>The FAA and Federal Reserve both warn that current assurance methods are incomplete.\u003C\u002Fli>\u003C\u002Ful>\u003Cp>There is also a lifecycle problem. Writing the prompt is treated like the finish line, when it is really the start of maintenance. Prompts drift. Models change. Vendor updates alter behavior. If the prompt is a production artifact, it needs versioning, regression tests, and retirement plans. That is normal engineering thinking, and it is still missing from a lot of prompt work.\u003C\u002Fp>\u003Ch2>Some prompt advice is actively bad\u003C\u002Fh2>\u003Cp>The strongest reason to stop calling this engineering is that some of the most popular advice fails in controlled studies. Research from Wharton’s \u003Ca href=\"https:\u002F\u002Fgail.wharton.upenn.edu\u002F\" target=\"_blank\" rel=\"noopener\">Generative AI Lab\u003C\u002Fa> presented at EMNLP 2024 found that expert persona prompting can reduce factual accuracy. The same research found that chain-of-thought prompting can hurt performance on reasoning models. Those are not edge cases. They are techniques repeated in almost every prompt guide online.\u003C\u002Fp>\u003Cp>That should make people uncomfortable. If a widely recommended technique makes output worse in documented tests, then the field has a quality-control problem. Good engineering disciplines do not keep bad practices in circulation for years because they sound smart.\u003C\u002Fp>\u003Cp>There is a deeper lesson here too. Prompt effectiveness depends heavily on the model, the task, and the evaluation method. A trick that helps one system may harm another. That makes the whole practice feel less like engineering and more like tuning a musical instrument by ear while the instrument keeps changing shape.\u003C\u002Fp>\u003Cp>For teams building with AI, the practical takeaway is simple: treat prompts as experiments, not doctrine. Measure output quality. Keep version history. Test against real cases. Retire techniques that do not survive evaluation, even if they are popular on social media.\u003C\u002Fp>\u003Ch2>What to do instead\u003C\u002Fh2>\u003Cp>Prompt work is real work. It can improve systems, reduce error, and make models more useful. But the word “engineering” should be reserved for practices with formal methods, repeatable testing, and standards that outlive one product cycle.\u003C\u002Fp>\u003Cp>My prediction is simple: the teams that win with AI will stop treating prompts like magic phrases and start treating them like testable artifacts. The ones that keep relying on vibes will keep shipping inconsistent systems. If you are building with LLMs today, ask one question before you call the work engineering: what is the acceptance test?\u003C\u002Fp>\u003Cp>That question will tell you a lot.\u003C\u002Fp>","Prompt design is mostly heuristic, not formal engineering. The evidence shows weak standards, shaky testing, and a lot of guesswork.","the-infrastructure-mindset.ghost.io","https:\u002F\u002Fthe-infrastructure-mindset.ghost.io\u002Fprompt-engineering-is-not\u002F",null,"https:\u002F\u002Fxxdpdyhzhpamafnrdkyq.supabase.co\u002Fstorage\u002Fv1\u002Fobject\u002Fpublic\u002Fcovers\u002Finline-1775122511647-9b56.png",[13,14,15,16,17],"prompt engineering","LLM testing","AI evaluation","software engineering","Microsoft Azure AI","en",0,false,"2026-04-02T08:42:38.021219+00:00","2026-04-02T08:42:37.986+00:00","done","7e77e581-2608-44af-90a9-1c044bcbe73b","why-prompt-engineering-isnt-engineering-en","research","a130f571-69fd-4e39-8af3-4e9171bb1b56","published","2026-04-08T09:00:53.522+00:00",[31,33,35,37,39],{"name":17,"slug":32},"microsoft-azure-ai",{"name":13,"slug":34},"prompt-engineering",{"name":14,"slug":36},"llm-testing",{"name":15,"slug":38},"ai-evaluation",{"name":16,"slug":40},"software-engineering",{"id":27,"slug":42,"title":43,"language":44},"why-prompt-engineering-isnt-engineering-zh","為什麼 Prompt Engineering 不是工程","zh",[46,52,58,64,70,76],{"id":47,"slug":48,"title":49,"cover_image":50,"image_url":50,"created_at":51,"category":26},"94994abd-e24d-4fd1-b941-942d03d19acf","turboquant-seo-shift-small-sites-en","TurboQuant and the SEO Shift for Small Sites","https:\u002F\u002Fxxdpdyhzhpamafnrdkyq.supabase.co\u002Fstorage\u002Fv1\u002Fobject\u002Fpublic\u002Fcovers\u002Finline-1778840455122-jfce.png","2026-05-15T10:20:28.134545+00:00",{"id":53,"slug":54,"title":55,"cover_image":56,"image_url":56,"created_at":57,"category":26},"670a7f69-911f-41e8-a18b-7d3491253a19","turboquant-vllm-comparison-fp8-kv-cache-en","TurboQuant vs FP8: vLLM’s first broad test","https:\u002F\u002Fxxdpdyhzhpamafnrdkyq.supabase.co\u002Fstorage\u002Fv1\u002Fobject\u002Fpublic\u002Fcovers\u002Finline-1778839858405-b5ao.png","2026-05-15T10:10:37.219158+00:00",{"id":59,"slug":60,"title":61,"cover_image":62,"image_url":62,"created_at":63,"category":26},"5aef1c57-961f-49f7-8277-f83f7336799a","llmbda-calculus-agent-safety-rules-en","LLMbda calculus gives agents safety rules","https:\u002F\u002Fxxdpdyhzhpamafnrdkyq.supabase.co\u002Fstorage\u002Fv1\u002Fobject\u002Fpublic\u002Fcovers\u002Finline-1778825459914-obkf.png","2026-05-15T06:10:36.242145+00:00",{"id":65,"slug":66,"title":67,"cover_image":68,"image_url":68,"created_at":69,"category":26},"712a0357-f7cd-48f2-adde-c2691da0815f","low-complexity-beamspace-denoiser-mmwave-mimo-en","A simpler beamspace denoiser for mmWave MIMO","https:\u002F\u002Fxxdpdyhzhpamafnrdkyq.supabase.co\u002Fstorage\u002Fv1\u002Fobject\u002Fpublic\u002Fcovers\u002Finline-1778814646705-e7mx.png","2026-05-15T03:10:31.764301+00:00",{"id":71,"slug":72,"title":73,"cover_image":74,"image_url":74,"created_at":75,"category":26},"f595f949-6ea1-4b0e-a632-f1832ef26e36","ai-benchmark-wins-cyber-scare-defenders-en","Why AI benchmark wins in cyber should scare defenders","https:\u002F\u002Fxxdpdyhzhpamafnrdkyq.supabase.co\u002Fstorage\u002Fv1\u002Fobject\u002Fpublic\u002Fcovers\u002Finline-1778807444539-gz7f.png","2026-05-15T01:10:30.04579+00:00",{"id":77,"slug":78,"title":79,"cover_image":80,"image_url":80,"created_at":81,"category":26},"3ad202d1-9e5f-49c5-8383-02fcf1a23cf2","why-linux-security-needs-patch-wave-mindset-en","Why Linux security needs a patch-wave mindset","https:\u002F\u002Fxxdpdyhzhpamafnrdkyq.supabase.co\u002Fstorage\u002Fv1\u002Fobject\u002Fpublic\u002Fcovers\u002Finline-1778741441493-ikl6.png","2026-05-14T06:50:25.906256+00:00",[83,88,93,98,103,108,113,118,123,128],{"id":84,"slug":85,"title":86,"created_at":87},"a2715e72-1fe8-41b3-abb1-d0cf1f710189","ai-predictions-2026-big-changes-en","AI Predictions for 2026: Brace for Big Changes","2026-03-26T01:25:07.788356+00:00",{"id":89,"slug":90,"title":91,"created_at":92},"8404bd7b-4c2f-4109-9ec4-baf29d88af2b","ml-papers-of-the-week-github-research-desk-en","ML Papers of the Week Turns GitHub Into a Research Desk","2026-03-27T01:11:39.480259+00:00",{"id":94,"slug":95,"title":96,"created_at":97},"87897a94-8065-4464-a016-1f23e89e17cc","ai-ml-conferences-to-watch-in-2026-en","AI\u002FML Conferences to Watch in 2026","2026-03-27T01:51:54.184108+00:00",{"id":99,"slug":100,"title":101,"created_at":102},"6f1987cf-25f3-47a4-b3e6-db0997695be8","openclaw-agents-manipulated-self-sabotage-en","OpenClaw Agents Can Be Manipulated Into Failure","2026-03-28T03:03:18.899465+00:00",{"id":104,"slug":105,"title":106,"created_at":107},"a53571ad-735a-4178-9f93-cb09b699d99c","vega-driving-language-instructions-en","Vega: Driving with Natural Language Instructions","2026-03-28T14:54:04.698882+00:00",{"id":109,"slug":110,"title":111,"created_at":112},"a34581d6-f36e-46da-88bb-582fb3e7425c","personalizing-autonomous-driving-styles-en","Drive My Way: Personalizing Autonomous Driving Styles","2026-03-28T14:54:26.148181+00:00",{"id":114,"slug":115,"title":116,"created_at":117},"2bc1ad7f-26ce-4f02-9885-803b35fd229d","training-knowledge-bases-writeback-rag-en","Training Knowledge Bases with WriteBack-RAG","2026-03-28T14:54:45.643433+00:00",{"id":119,"slug":120,"title":121,"created_at":122},"71adc507-3c54-4605-bbe2-c966acd6187e","packforcing-long-video-generation-en","PackForcing: Efficient Long-Video Generation Method","2026-03-28T14:55:02.646943+00:00",{"id":124,"slug":125,"title":126,"created_at":127},"675942ef-b9ec-4c5f-a997-381250b6eacb","pixelsmile-facial-expression-editing-en","PixelSmile Framework Enhances Facial Expression Editing","2026-03-28T14:55:20.633463+00:00",{"id":129,"slug":130,"title":131,"created_at":132},"6954fa2b-8b66-4839-884b-e46f89fa1bc3","adaptive-block-scaled-data-types-en","IF4: Smarter 4-Bit Quantization That Adapts to Your Data","2026-03-31T06:00:36.65963+00:00"]