From 22653896d3f3e4459952c8445c8e5861309f588d Mon Sep 17 00:00:00 2001 From: Simon Willison Date: Sun, 26 Feb 2023 14:12:35 -0800 Subject: [PATCH 1/3] Fixed spelling of my name --- guides/prompt-adversarial.md | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/guides/prompt-adversarial.md b/guides/prompt-adversarial.md index 623a52e..d94ca6d 100644 --- a/guides/prompt-adversarial.md +++ b/guides/prompt-adversarial.md @@ -17,7 +17,7 @@ Topics: --- ## Prompt Injection -Prompt injection aims to hijack the model output by using clever prompts that change its behavior. These attacks could be harmful -- Simon Williams define it ["as a form of security exploit"](https://simonwillison.net/2022/Sep/12/prompt-injection/). +Prompt injection aims to hijack the model output by using clever prompts that change its behavior. These attacks could be harmful -- Simon Willison defined it ["as a form of security exploit"](https://simonwillison.net/2022/Sep/12/prompt-injection/). Let's cover a basic example to demonstrate how prompt injection can be achieved. We will use a popular example shared by [Riley on Twitter](https://twitter.com/goodside/status/1569128808308957185?s=20). @@ -186,4 +186,4 @@ Models like ChatGPT and Claude have been aligned to avoid outputting content tha --- [Previous Section (Advanced Prompting)](./prompts-advanced-usage.md) -[Next Section (Miscellaneous Topics)](./prompt-miscellaneous.md) \ No newline at end of file +[Next Section (Miscellaneous Topics)](./prompt-miscellaneous.md) From 4e64cc503167c8e21fbf36bb6e10abd67d81ac49 Mon Sep 17 00:00:00 2001 From: Shuofei Qiao <920722741@qq.com> Date: Mon, 27 Feb 2023 10:46:49 +0800 Subject: [PATCH 2/3] add a survey --- README.md | 1 + 1 file changed, 1 insertion(+) diff --git a/README.md b/README.md index 0333d62..a4c43c9 100644 --- a/README.md +++ b/README.md @@ -58,6 +58,7 @@ The following are the latest papers (sorted by release date) on prompt engineeri - [Augmented Language Models: a Survey](https://arxiv.org/abs/2302.07842) (Feb 2023) - [A Survey for In-context Learning](https://arxiv.org/abs/2301.00234) (Dec 2022) - [Towards Reasoning in Large Language Models: A Survey](https://arxiv.org/abs/2212.10403) (Dec 2022) + - [Reasoning with Language Model Prompting: A Survey](https://arxiv.org/abs/2212.09597) (Dec 2022) - [Emergent Abilities of Large Language Models](https://arxiv.org/abs/2206.07682) (Jun 2022) - [A Taxonomy of Prompt Modifiers for Text-To-Image Generation](https://arxiv.org/abs/2204.13988) (Apr 2022) - [Pre-train, Prompt, and Predict: A Systematic Survey of Prompting Methods in Natural Language Processing](https://arxiv.org/abs/2107.13586) (Jul 2021) From b7a17eb9736c649b6147acbf403f7a755e332ad7 Mon Sep 17 00:00:00 2001 From: Santiago Lezica Date: Mon, 27 Feb 2023 13:07:34 -0300 Subject: [PATCH 3/3] Update prompts-advanced-usage.md It feels like the "let's talk about CoT next" paragraph should be just before the CoT heading. I was (briefly) confused when I came across it. --- guides/prompts-advanced-usage.md | 8 ++++---- 1 file changed, 4 insertions(+), 4 deletions(-) diff --git a/guides/prompts-advanced-usage.md b/guides/prompts-advanced-usage.md index c0d54be..9d94c60 100644 --- a/guides/prompts-advanced-usage.md +++ b/guides/prompts-advanced-usage.md @@ -79,9 +79,6 @@ The answer is True. That didn't work. It seems like basic standard prompting is not enough to get reliable responses for this type of reasoning problem. The example above provides basic information on the task, even with the examples. If you take a closer look at the task, it does involve more reasoning steps. -More recently, chain-of-thought (CoT) prompting has been popularized to address more complex arithmetic, -commonsense, and symbolic reasoning tasks. So let's talk about CoT next and see if we can solve the above task. - Following the findings from [Min et al. (2022)](https://arxiv.org/abs/2202.12837), here a few more tips about demonstrations/exemplars when doing few-shot: - the label space and the distribution of the input text specified by the demonstrations are both key (regardless of whether the labels are correct @@ -124,6 +121,9 @@ There is no consistency in the format above but the model still predicted the co Overall, it seems that providing examples is useful in some places. When zero-shot prompting and few-shot prompting are not sufficient, it might mean that the whatever was learned by the model isn't enough to do well at the task. From here it is recommended to start thinking about fine-tuning your own models. +More recently, chain-of-thought (CoT) prompting has been popularized to address more complex arithmetic, +commonsense, and symbolic reasoning tasks. So let's talk about CoT next and see if we can solve the above task. + --- ## Chain-of-Thought Prompting @@ -399,4 +399,4 @@ This paper touches on an important topic related to prompt engineering which is --- [Previous Section (Basic Prompting)](./prompts-basic-usage.md) -[Next Section (Adversarial Prompting)](./prompt-adversarial.md) \ No newline at end of file +[Next Section (Adversarial Prompting)](./prompt-adversarial.md)