[Yannic Kilcher] Tree of Thoughts: Deliberate Problem Solving with Large Language Models (Full Paper Review)
🎯 Загружено автоматически через бота:
🚫 Оригинал видео:
📺 Данное видео принадлежит каналу «Yannic Kilcher» (@YannicKilcher). Оно представлено в нашем сообществе исключительно в информационных, научных, образовательных или культурных целях. Наше сообщество не утверждает никаких прав на данное видео. Пожалуйста, поддержите автора, посетив его оригинальный канал.
✉️ Если у вас есть претензии к авторским правам на данное видео, пожалуйста, свяжитесь с нами по почте support@, и мы немедленно удалим его.
📃 Оригинальное описание:
#gpt4 #ai #prompt
Tree-of-Thought improves prompting of large language models (LLMs) by generalizing the concept of Chain-of-Thought prompting and introduces a tree search across language model thoughts, including state evaluation and backtracking. Experiments on toy tasks show large improvements over both classic and Chain-of-Thought prompting.
OUTLINE:
- Introduction
- From Chain-of-Thought to Tree-of-Thought
- Formalizing the algorithm
- Game of 24 & Creative writing
- Crosswords
- Is this a general problem solver?
- Ablation studies
- Conclusion
Paper:
Abstract:
Language models are increasingly being deployed for general problem solving across a wide range of tasks, but are still confined to token-level, left-to-right decision-making processes during inference. This means they can fall short in tasks that require exploration, strategic lookahead, or where initial decisions play a pivotal role. To surmount these challenges, we introduce a new framework for language model inference, Tree of Thoughts (ToT), which generalizes over the popular Chain of Thought approach to prompting language models, and enables exploration over coherent units of text (thoughts) that serve as intermediate steps toward problem solving. ToT allows LMs to perform deliberate decision making by considering multiple different reasoning paths and self-evaluating choices to decide the next course of action, as well as looking ahead or backtracking when necessary to make global choices. Our experiments show that ToT significantly enhances language models’ problem-solving abilities on three novel tasks requiring non-trivial planning or search: Game of 24, Creative Writing, and Mini Crosswords. For instance, in Game of 24, while GPT-4 with chain-of-thought prompting only solved 4% of tasks, our method achieved a success rate of 74%. Code repo with all prompts: this https URL.
Authors: Shunyu Yao, Dian Yu, Jeffrey Zhao, Izhak Shafran, Thomas L. Griffiths, Yuan Cao, Karthik Narasimhan
Links:
Homepage:
Merch:
YouTube:
Twitter:
Discord:
LinkedIn:
If you want to support me, the best thing to do is to share out the content :)
If you want to support me financially (completely optional and voluntary, but a lot of people have asked for this):
SubscribeStar:
Patreon:
Bitcoin (BTC): bc1q49lsw3q325tr58ygf8sudx2dqfguclvngvy2cq
Ethereum (ETH): 0x7ad3513E3B8f66799f507Aa7874b1B0eBC7F85e2
Litecoin (LTC): LQW2TRyKYetVC8WjFkhpPhtpbDM4Vw7r9m
Monero (XMR): 4ACL8AGrEo5hAir8A9CeVrW8pEauWvnp1WnSDZxW7tziCDLhZAGsgzhRQABDnFy8yuM9fWJDviJPHKRjV4FWt19CJZN9D4n
4 views
0
0
2 weeks ago 00:29:28 4
[Yannic Kilcher] Tree of Thoughts: Deliberate Problem Solving with Large Language Models (Full Paper Review)
1 year ago 00:17:24 2
Deep image reconstruction from human brain activity (Paper Explained)
2 years ago 01:21:28 6
Unsupervised Brain Models - How does Deep Learning inform Neuroscience? (w/ Patrick Mineault)
2 years ago 00:12:33 8
Machine Learning News you must know- June 2022. Devansh Machine Learning Made Simple.
3 years ago 00:56:02 3
How He Breaks Down Complex Machine Learning Research for YouTube (@Yannic Kilcher) - KNN Ep. 95
3 years ago 00:08:39 15
THIS AI REMOVES UNWANTED OBJECTS FROM IMAGES!
3 years ago 00:41:51 1
Full Self-Driving is HARD! Analyzing Elon Musk re: Tesla Autopilot on Lex Fridman’s Podcast
3 years ago 00:06:14 15
3D Modeling Without An Artist…Possible? 👩🎨
3 years ago 00:13:19 21
GPT-3 is a LIAR - Misinformation and fear-mongering around the TruthfulQA dataset
3 years ago 00:04:01 1
[ML News] DeepMind fails to get independence from Google
3 years ago 00:16:30 2
Divide and Contrast Explained!
3 years ago 00:10:49 10
My GitHub (Trash code I wrote during PhD)
4 years ago 00:13:53 99
AI made this music video | What happens when OpenAI’s CLIP meets BigGAN?
4 years ago 01:05:04 8
AI Weekly Update - March 8th, 2021 (#27)!
4 years ago 00:00:59 3
What is the Hardware Lottery? (Machine Learning Research Paper Explained by Yannic Kilcher) #Shorts
4 years ago 00:14:28 42
MEMES IS ALL YOU NEED - Deep Learning Meme Review - Episode 2 (Part 1 of 2)
4 years ago 01:30:36 12
Sara Hooker - The Hardware Lottery, Sparsity and Fairness
4 years ago 00:11:11 2
[News] OpenAI Model Generates Python Code
4 years ago 01:23:34 1
Programming Languages, Software Engineering and Machine Learning
4 years ago 01:02:33 6
Facebook Research - Unsupervised Translation of Programming Languages
4 years ago 02:33:32 15
Francois Chollet - On the Measure Of Intelligence
4 years ago 01:18:31 17
Francois Chollet: What is the future of artificial intelligence?
4 years ago 01:38:26 5
Harri Valpola: System 2 AI and Planning in Model-Based Reinforcement Learning
5 years ago 02:34:18 6
ICLR 2020: Yoshua Bengio and the Nature of Consciousness