[Andrej Karpathy] [1hr Talk] Intro to Large Language Models
🎯 Загружено автоматически через бота:
🚫 Оригинал видео:
📺 Данное видео принадлежит каналу «Andrej Karpathy» (@AndrejKarpathy). Оно представлено в нашем сообществе исключительно в информационных, научных, образовательных или культурных целях. Наше сообщество не утверждает никаких прав на данное видео. Пожалуйста, поддержите автора, посетив его оригинальный канал.
✉️ Если у вас есть претензии к авторским правам на данное видео, пожалуйста, свяжитесь с нами по почте support@, и мы немедленно удалим его.
📃 Оригинальное описание:
This is a 1 hour general-audience introduction to Large Language Models: the core technical component behind systems like ChatGPT, Claude, and Bard. What they are, where they are headed, comparisons and analogies to present-day operating systems, and some of the security-related challenges of this new computing paradigm.
As of November 2023 (this field moves fast!).
Context: This video is based on the slides of a talk I gave recently at the AI Security Summit. The talk was not recorded but a lot of people came to me after and told me they liked it. Seeing as I had already put in one long weekend of work to make the slides, I decided to just tune them a bit, record this round 2 of the talk and upload it here on YouTube. Pardon the random background, that’s my hotel room during the thanksgiving break.
Slides as PDF: (42MB)
Slides. as Keynote: (140MB)
Few things I wish I said (I’ll add items here as they come up):
The dreams and hallucinations do not get fixed with finetuning. Finetuning just “directs“ the dreams into “helpful assistant dreams“. Always be careful with what LLMs tell you, especially if they are telling you something from memory alone. That said, similar to a human, if the LLM used browsing or retrieval and the answer made its way into the “working memory“ of its context window, you can trust the LLM a bit more to process that information into the final answer. But TLDR right now, do not trust what LLMs say or do. For example, in the tools section, I’d always recommend double-checking the math/code the LLM did.
How does the LLM use a tool like the browser? It emits special words, e.g. |BROWSER|. When the code “above“ that is inferencing the LLM detects these words it captures the output that follows, sends it off to a tool, comes back with the result and continues the generation. How does the LLM know to emit these special words? Finetuning datasets teach it how and when to browse, by example. And/or the instructions for tool use can also be automatically placed in the context window (in the “system message”).
You might also enjoy my 2015 blog post “Unreasonable Effectiveness of Recurrent Neural Networks“. The way we obtain base models today is pretty much identical on a high level, except the RNN is swapped for a Transformer.
What is in the run.c file? A bit more full-featured 1000-line version hre:
Chapters:
Part 1: LLMs
Intro: Large Language Model (LLM) talk
LLM Inference
LLM Training
LLM dreams
How do they work?
Finetuning into an Assistant
Summary so far
Appendix: Comparisons, Labeling docs, RLHF, Synthetic data, Leaderboard
Part 2: Future of LLMs
LLM Scaling Laws
Tool Use (Browser, Calculator, Interpreter, DALL-E)
Multimodality (Vision, Audio)
Thinking, System 1/2
Self-improvement, LLM AlphaGo
LLM Customization, GPTs store
LLM OS
Part 3: LLM Security
LLM Security Intro
Jailbreaks
Prompt Injection
Data poisoning
LLM Security conclusions
End
Outro
11 views
0
0
2 months ago 01:55:57 5
[Andrej Karpathy] Building makemore Part 3: Activations & Gradients, BatchNorm
2 months ago 01:55:23 12
[Andrej Karpathy] Building makemore Part 4: Becoming a Backprop Ninja
2 months ago 00:56:21 4
[Andrej Karpathy] Building makemore Part 5: Building a WaveNet
2 months ago 01:56:19 72
[Andrej Karpathy] Let’s build GPT: from scratch, in code, spelled out.
2 months ago 00:59:47 11
[Andrej Karpathy] [1hr Talk] Intro to Large Language Models
2 months ago 04:01:25 13
[Andrej Karpathy] Let’s reproduce GPT-2 (124M)
2 months ago 02:13:34 2
[Andrej Karpathy] Let’s build the GPT Tokenizer
2 months ago 00:03:19 1
How To Study Hard - Richard Feynman
2 months ago 00:26:10 1
Attention in transformers, visually explained | Chapter 6, Deep Learning
2 months ago 00:27:14 1
How large language models work, a visual intro to transformers | Chapter 5, Deep Learning
3 months ago 00:27:13 10
But what is a GPT? Visual intro to transformers | Chapter 5, Deep Learning
3 months ago 00:44:17 1
No Priors Ep. 80 | With Andrej Karpathy from OpenAI and Tesla
4 months ago 00:18:11 67
НОВОСТИ ИИ: Подписка на ChatGPT за 2000$
4 months ago 00:06:53 1
Elon Musk says losers use LiDAR. [Explanation video]
5 months ago 00:47:27 1
Projeto Secreto da OpenAI: Descubra as Últimas Inovações da IA e Fique Super Atualizado no IA News#6
5 months ago 00:07:41 1
Educação 100% com IA, FBI invade celular, Hype das IA no fim, e muito mais
8 months ago 00:40:08 1
The Most Important Algorithm in Machine Learning
8 months ago 00:08:29 1
CATL’s sodium hybrid battery will be 30% cheaper & revolutionise the world
8 months ago 00:08:55 1
Tesla reveals timelline for massive electric Semi production at $ factory
10 months ago 00:26:53 1
Vedal & Neuro Build A Language Model From Scratch
10 months ago 00:16:39 1
Phi-1: A ’Textbook’ Model
11 months ago 00:20:13 1
GPT-5: Everything You Need to Know So Far
1 year ago 00:14:07 1
“Что в имени тебе моем?“ Учимся генерировать новые имена у звездного разработчика Tesla и OpenAI.
1 year ago 00:59:48 22
Введение в большие языковые модели от Andrej Karpathy