Introduction
- Understand what a prompting technique is
- Understand the contents of the Intermediate section
You have made it through the beginning stages of Prompt Engineering! Now you can dive into some intermediate techniques which can really take your prompting to the next level.
Here, you are going to shift your focus from the tasks that GenAI can solve, onto the prompting techniques themselves. According to The Prompt Report, "a prompting technique is a blueprint that describes how to structure a prompt, prompts, or dynamic sequencing of multiple prompts. A prompting technique may incorporate conditional or branching logic, parallelism, or other architectural considerations spanning multiple prompt". In the coming lessons, we will focus on more technical aspects of prompting such as prompt structure and design.
This module will expose you to moderately complex, research-based prompt engineering techniques. You'll learn how to implement these techniques to improve the performance of your GenAI applications. Some topics we will explore are Chain-of-Thought, Self-Consistency, and Generated knowledge. We will also revisit a technique we have already touched on, Role Prompting, and expand on its use. Along the way, you will also learn more about where prompting LLMs (Large Language Models) can fail.
By the end of this module, you will have a fundamental understanding of many of the world's most used prompting techniques and be able to apply them to a myriad of tasks.
Sander Schulhoff
Sander Schulhoff is the CEO of HackAPrompt and Learn Prompting. He created the first Prompt Engineering guide on the internet, two months before ChatGPT was released, which has taught 3 million people how to prompt ChatGPT. He also partnered with OpenAI to run the first AI Red Teaming competition, HackAPrompt, which was 2x larger than the White House's subsequent AI Red Teaming competition. Today, HackAPrompt partners with the Frontier AI labs to produce research that makes their models more secure. Sander's background is in Natural Language Processing and deep reinforcement learning. He recently led the team behind The Prompt Report, the most comprehensive study of prompt engineering ever done. This 76-page survey, co-authored with OpenAI, Microsoft, Google, Princeton, Stanford, and other leading institutions, analyzed 1,500+ academic papers and covered 200+ prompting techniques.
π’ Chain-of-Thought Prompting
π¦ Basic LLM Settings
π¦ Generated Knowledge
π¦ Least-to-Most Prompting
π¦ Dealing With Long Form Content
π¦ OpenAI Playground
π¦ Revisiting Roles
π¦ Self-Consistency
π¦ More About Prompt Elements
π’ Zero-Shot Chain-of-Thought
Footnotes
-
Schulhoff, S., Ilie, M., Balepur, N., Kahadze, K., Liu, A., Si, C., Li, Y., Gupta, A., Han, H., Schulhoff, S., Dulepet, P. S., Vidyadhara, S., Ki, D., Agrawal, S., Pham, C., Kroiz, G., Li, F., Tao, H., Srivastava, A., β¦ Resnik, P. (2024). The Prompt Report: A Systematic Survey of Prompting Techniques. https://arxiv.org/abs/2406.06608 β© β©2