Prompt Optimization of Large Language Model for Interactive Tasks without Gradient and Demonstrations

05/24/2023
by   Siqi Ouyang, et al.
0

Large language models (LLMs) have demonstrated remarkable language proficiency, but they face challenges when solving interactive tasks independently. Existing methods either rely on gradient access, which is often inaccessible in state-of-the-art LLMs like GPT-4, or necessitate diverse and high-quality in-context demonstrations. In this study, we propose LLM-PO, a novel approach that enables LLMs to address these tasks without gradient access or extensive demonstrations. The key idea is to maintain a text-based plan and ask LLMs to reflect on pros and cons of the current plan based on experience collected with it, to update the plan, and to collect more experiences with the new plan. Experiments on HotpotQA demonstrate that LLM-PO achieves higher or on par success rates compared to in-context learning (ICL) baselines while requiring less inference cost.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/26/2023

AdaPlanner: Adaptive Planning from Feedback with Language Models

Large language models (LLMs) have recently demonstrated the potential in...
research
01/27/2023

Large Language Models Are Implicitly Topic Models: Explaining and Finding Good Demonstrations for In-Context Learning

In recent years, pre-trained large language models have demonstrated rem...
research
07/01/2023

DoReMi: Grounding Language Model by Detecting and Recovering from Plan-Execution Misalignment

Large language models encode a vast amount of semantic knowledge and pos...
research
02/15/2022

Text-Based Action-Model Acquisition for Planning

Although there have been approaches that are capable of learning action ...
research
07/05/2023

Scaling In-Context Demonstrations with Structured Attention

The recent surge of large language models (LLMs) highlights their abilit...
research
05/24/2023

Adversarial Demonstration Attacks on Large Language Models

With the emergence of more powerful large language models (LLMs), such a...
research
05/24/2023

Coverage-based Example Selection for In-Context Learning

In-context learning (ICL), the ability of large language models to perfo...

Please sign up or login with your details

Forgot password? Click here to reset