How to Get AI to Do What You Want: The Human Art of Smart Communication

Josh Fruhlinger
13 Min Read

As generative AI becomes ubiquitous in business software, prompt engineering has emerged as a fundamental capability for both developers and general knowledge workers.

Credit: Digineer Station / Shutterstock

Prompt engineering describes the art of designing inputs, or “prompts,” for a generative AI system to elicit superior outputs. While seemingly straightforward, the intricate, nondeterministic, and often opaque nature of large language models (LLMs) and other generative AI tools makes this a highly challenging endeavor, requiring considerable experimentation and intuition. This complexity exists even before tackling the subjective definition of what constitutes a “better output.”

Throughout the history of computer science, from the advent of COBOL, innovations have promised to democratize technology, enabling ordinary users to harness computing power without specialized skills or languages. With natural language AI chatbots, this aspiration appears to be realized. Yet, it turns out that various techniques—some intuitive, others less so—are crucial for optimizing generative AI performance. Mastering these techniques is rapidly becoming a vital skill in the age of AI.

Why is prompt engineering important?

Most individuals interact with generative AI tools like ChatGPT or Claude directly. For these users, prompt engineering offers a pathway to obtaining more effective responses from these systems. As these tools increasingly integrate into business software and processes, improving prompt quality becomes a significant advantage, much like early internet users learned the nuances of Google and other search engines.

However, prompt engineering holds even greater significance for developers constructing comprehensive AI ecosystems designed to simplify interactions for end-users. Enterprise AI applications are increasingly incorporating an orchestration layer positioned between the end-user and the foundational AI model. This layer integrates system prompts and utilizes retrieval augmented generation (RAG) tools to refine user inputs before they reach the core AI system.

For example, a medical AI application could allow doctors and nurses to merely input a list of patient symptoms. The application’s orchestration layer would then transform this list into an optimized prompt, applying prompt engineering principles and enriching it with RAG-derived information, with the aim of generating the most accurate diagnosis.

For developers, this orchestration layer represents a pivotal new domain of professional expertise in the AI era. Just as search engines, initially created for general users, eventually gave rise to a multi-billion dollar search engine optimization (SEO) industry, prompt engineering is similarly emerging as an essential and potentially highly rewarding skill.

Prompt engineering types and techniques

Prompt engineering methodologies vary in sophistication but share a common objective: to guide the AI model’s internal reasoning processes and mitigate its propensity for vagueness or generating incorrect information (hallucination). These techniques generally fall into several main categories:

Zero-shot prompting is often the simplest and default approach: you provide the model with a direct instruction—such as “Summarize this article,” “Explain this API,” or “Draft a patient note”—and the system generates a response based solely on its general training data. This is known as direct or zero-shot prompting (a term we’ll clarify shortly); it’s efficient for quick tasks but often lacks the consistency or structured output necessary for enterprise environments, where responses must adhere to specific formats and satisfy compliance or quality standards.One-shot and few-shot prompting enhance instructions by including examples, thereby demonstrating the desired format, reasoning style, or output structure. Below is an illustration of one-shot prompting using ChatGPT:

 One-shot prompting with ChatGPT
Foundry

This instance exemplifies one-shot prompting due to its single example. However, additional examples can be provided for few-shot (or even many-shot) prompts. Consequently, direct prompts lacking examples were later termed zero-shot.

Such prompts facilitate in-context learning, where examples guide the model towards improved performance. For instance, a model struggling with a zero-shot command like “Extract key risks from this report” might perform significantly better if provided with a few illustrative examples of the specific types of risks required. In production systems, these examples are frequently integrated into the system prompt or stored within an internal prompt template library, remaining invisible to the end-user.

Chain-of-thought prompting extends this by encouraging the model to decompose a problem into sequential intermediate steps. This method was first introduced in a 2022 research paper, which included the following example:

Source: “Chain-of-Thought Prompting Elicits Reasoning in Large Language Models,” Wei et al., 2022.
Foundry

Chain-of-thought prompts can feature elaborate demonstrations of the intended reasoning, as shown in the example. Nevertheless, it’s noteworthy that modern LLMs are increasingly capable of generating chain-of-thought reasoning independently with even subtle cues, such as simply adding “show your work” to the prompt. This technique proves particularly effective for complex reasoning tasks, including classification, diagnostics, planning, multi-step decision-making, or rules interpretation.

The functionality of these engineered prompts illuminates a crucial aspect of generative AI. While interfaces like ChatGPT create the illusion of conversing with an entity, the underlying model is fundamentally an engine for predicting the next token in a sequence.

When engaging in a natural conversational style, the model endeavors to predict the most probable next piece of dialogue based on its training data. However, as our examples illustrate, you can structure your prompt with multi-“character” dialogue frameworks, including both questions (Qs) and answers (As), and then instruct it to predict the subsequent A, or even the next Q. The model readily complies and does not inherently “identify” with either “character,” even capable of switching roles if prompted appropriately. Effective prompt engineering leverages this characteristic rather than attempting to coerce an LLM as if it were a human.

Zero-shot and few-shot examples can be embedded as system-level templates, and chain-of-thought reasoning can be enforced by the software layer, moving beyond user discretion. More sophisticated dialogue frameworks can shape model behavior, reducing risks and enhancing consistency. Collectively, these methods form the foundation of production-ready prompting, acting as an intermediary between end-users and the model.

Prompt engineering best practices

Clearly define roles and objectives. Explicitly state the model’s intended persona and the desired outcome of the output.
Utilize examples to establish patterns. One-shot and few-shot prompts aid the model in discerning structure, formatting, and reasoning styles.
Deconstruct complex tasks. Chain-of-thought prompts enhance reasoning by breaking down intricate queries into manageable steps.
Limit the model’s information sources. For RAG systems, instruct the model to exclusively draw upon retrieved documents or verified contextual information.
Specify output format requirements. Using JSON, XML, bullet points, or predefined templates reduces ambiguity and simplifies subsequent parsing.
Validate prompts across model iterations. Minor model updates can alter prompt behavior, making continuous evaluation imperative.
Document and version prompts systematically. Manage prompts as software artifacts to ensure consistency and maintain traceability.
Avoid excessively complex instructions. Shorter, clearer prompts minimize the likelihood of unintended model interpretations.

Prompt engineering challenges

Prompt engineering remains a rapidly developing field, presenting distinct challenges. A primary concern is prompt fragility: even minor linguistic alterations can significantly impact output quality. Prompts optimized for one model version may not perform identically with a newer version, necessitating continuous organizational maintenance to preserve stable outputs as models evolve.

Another related issue is opacity. Given that LLMs function as black-box systems, a well-constructed prompt does not guarantee sound reasoning; it merely increases the probability of the model interpreting instructions correctly. Research has underscored the discrepancy between expertly engineered prompts and reliably trustworthy outputs. In highly regulated sectors, a model that merely sounds convincing can pose substantial risks if its underlying prompt lacks sufficient constraints. (We’ve drawn parallels between prompt engineering and SEO, and both fragility and opacity are familiar hurdles for SEO professionals.)

Enterprise teams also encounter scalability challenges. Due to the nondeterministic nature of LLMs, a prompt effective for a single request may not consistently perform across thousands of queries, each with slightly varied inputs. As businesses expand their deployments, such inconsistencies can lead to productivity losses, compliance issues, or increased demands for human oversight.

Security risk constitutes another emerging concern. Prompt-injection attacks, where malicious user input or retrieved content manipulates internal prompt templates, are now recognized as tangible threats.

Prompt engineering courses

A further challenge in the prompt engineering landscape is the persistent skills gap. While enterprises acknowledge the significance of prompt engineering, the underlying technologies and methodologies are so nascent that few professionals possess practical experience in constructing robust prompt pipelines. This disparity fuels the demand for the expanding array of prompt engineering courses and certifications.

Companies themselves are increasingly initiating internal training programs as they deploy generative AI. Citi, for example, has mandated AI prompt training for approximately 175,000–180,000 employees authorized to use its AI tools, positioning it as a strategy to enhance AI proficiency across its workforce. Similarly, Deloitte’s AI Academy aims to educate over 120,000 professionals in generative AI and associated competencies.

Prompt engineering jobs

There is a growing need for professionals adept at designing prompt templates, developing orchestration layers, and integrating prompts with retrieval systems and data pipelines. Employers are increasingly seeking AI-skilled practitioners who not only grasp prompting but also understand its integration with retrieval systems and tool utilization.

These roles often encompass hybrid responsibilities: evaluating model updates, maintaining prompt libraries, assessing output quality, implementing safety protocols, and embedding prompts into multi-step agent workflows. As companies deepen AI integration into customer support, analytics, and operations, prompt engineers must collaborate with security, compliance, and UX teams to mitigate hallucination, model drift, or unforeseen system behaviors.

Despite some debate regarding the long-term viability of “prompt engineer” as a standalone job title, the foundational competencies—structured reasoning, workflow design, prompt orchestration, evaluation, and integration—are becoming integral to broader AI engineering disciplines. Demand for such talent remains robust, and compensation for AI skills continues its upward trend.

Prompt engineering guides

Readers keen on exploring practical techniques in more depth have access to several authoritative guides:

OpenAI’s Prompt Engineering Guide: This resource covers fundamental prompting patterns, emphasizing clarity, structure, role definitions, and reasoning instructions.

Google Cloud: What is Prompt Engineering: Provides an explanation of prompting basics and how prompt design fits within larger enterprise architectures.

IBM’s 2025 Guide to Prompt Engineering: Focuses on real-world enterprise applications, safety considerations, and the combination of prompt engineering with RAG and workflow automation.

DAIR-AI Prompt Engineering Guide: A collaborative community resource detailing contemporary prompting techniques, evaluation methodologies, and practical examples.

These resources can serve as an excellent starting point in this rapidly expanding field—however, nothing replaces direct, hands-on experience with prompts.

window.NREUM||(NREUM={});NREUM.info={“beacon”:”bam.nr-data.net”,”licenseKey”:”NRJS-51503b5fb963bca6c59″,”applicationID”:”1486670290″,”transactionName”:”NAZXbBAAXRZUU00KXA1MdFsWCFwLGkNQDVQPBhhIDRJH”,”queueTime”:0,”applicationTime”:495,”atts”:”GEFAGlgaTkkXURtZSB4e”,”errorBeacon”:”bam.nr-data.net”,”agent”:””}

                    Artificial IntelligenceGenerative AI                    

Share This Article
Leave a Comment

Leave a Reply

Your email address will not be published. Required fields are marked *