Популярное

Музыка Кино и Анимация Автомобили Животные Спорт Путешествия Игры Юмор

Интересные видео

2025 Сериалы Трейлеры Новости Как сделать Видеоуроки Diy своими руками

Топ запросов

смотреть а4 schoolboy runaway турецкий сериал смотреть мультфильмы эдисон
dTub
Скачать

Stanford CS25: V4 I Aligning Open Language Models

Автор: Stanford Online

Загружено: 2024-05-10

Просмотров: 29526

Описание:

April 18, 2024
Speaker: Nathan Lambert, Allen Institute for AI (AI2)

Aligning Open Language Models
Since the emergence of ChatGPT there has been an explosion of methods and models attempting to make open language models easier to use. This talk retells the major chapters in the evolution of open chat, instruct, and aligned models, covering the most important techniques, datasets, and models. Alpaca, QLoRA, DPO, PPO, and everything in between will be covered. The talk will conclude with predictions and expectations for the future of aligning open language models. Slides posted here: https://docs.google.com/presentation/...
All the models in the figures are in this HuggingFace collection: https://huggingface.co/collections/na...

About the speaker:
Nathan Lambert is a Research Scientist at the Allen Institute for AI focusing on RLHF and the author of Interconnects.ai. Previously, he helped build an RLHF research team at HuggingFace. He received his PhD from the University of California, Berkeley working at the intersection of machine learning and robotics. He was advised by Professor Kristofer Pister in the Berkeley Autonomous Microsystems Lab and Roberto Calandra at Meta AI Research.

More about the course can be found here: https://web.stanford.edu/class/cs25/

View the entire CS25 Transformers United playlist:    • Stanford CS25 - Transformers United  

Stanford CS25: V4 I Aligning Open Language Models

Поделиться в:

Доступные форматы для скачивания:

Скачать видео mp4

  • Информация по загрузке:

Скачать аудио mp3

Похожие видео

Stanford CS25: V4 I Demystifying Mixtral of Experts

Stanford CS25: V4 I Demystifying Mixtral of Experts

Andrej Karpathy: Software Is Changing (Again)

Andrej Karpathy: Software Is Changing (Again)

The Origin and Future of RLHF: the secret ingredient for ChatGPT - with Nathan Lambert

The Origin and Future of RLHF: the secret ingredient for ChatGPT - with Nathan Lambert

Stanford CS25: V5 I Transformers for Video Generation, Andrew Brown of Meta

Stanford CS25: V5 I Transformers for Video Generation, Andrew Brown of Meta

Stanford CS25: V4 I Behind the Scenes of LLM Pre-training: StarCoder Use Case

Stanford CS25: V4 I Behind the Scenes of LLM Pre-training: StarCoder Use Case

State of GPT | BRK216HFS

State of GPT | BRK216HFS

Stanford Webinar - Agentic AI: A Progression of Language Model Usage

Stanford Webinar - Agentic AI: A Progression of Language Model Usage

Stanford CS25: V4 I Overview of Transformers

Stanford CS25: V4 I Overview of Transformers

Reinforcement Learning from Human Feedback: From Zero to chatGPT

Reinforcement Learning from Human Feedback: From Zero to chatGPT

Stanford CS25: V4 I Hyung Won Chung of OpenAI

Stanford CS25: V4 I Hyung Won Chung of OpenAI

Stanford CS230 | Autumn 2025 | Lecture 7: Agents, Prompts, and RAG.

Stanford CS230 | Autumn 2025 | Lecture 7: Agents, Prompts, and RAG.

Краткое объяснение больших языковых моделей

Краткое объяснение больших языковых моделей

Stanford CS25: V4 I Transformers that Transform Well Enough to Support Near-Shallow Architectures

Stanford CS25: V4 I Transformers that Transform Well Enough to Support Near-Shallow Architectures

Stanford CS25: V3 I Generalist Agents in Open-Ended Worlds

Stanford CS25: V3 I Generalist Agents in Open-Ended Worlds

Let's build the GPT Tokenizer

Let's build the GPT Tokenizer

Stanford CS25: V3 I Low-level Embodied Intelligence w/ Foundation Models

Stanford CS25: V3 I Low-level Embodied Intelligence w/ Foundation Models

ДНК создал Бог? Самые свежие научные данные о строении. Как работает информация для жизни организмов

ДНК создал Бог? Самые свежие научные данные о строении. Как работает информация для жизни организмов

Let's build GPT: from scratch, in code, spelled out.

Let's build GPT: from scratch, in code, spelled out.

Stanford CS25: V4 I Jason Wei & Hyung Won Chung of OpenAI

Stanford CS25: V4 I Jason Wei & Hyung Won Chung of OpenAI

A little guide to building Large Language Models in 2024

A little guide to building Large Language Models in 2024

© 2025 dtub. Все права защищены.



  • Контакты
  • О нас
  • Политика конфиденциальности



Контакты для правообладателей: [email protected]