<?xml version='1.0' encoding='UTF-8'?>
<?xml-stylesheet href="/rss/stylesheet/" type="text/xsl"?>
<rss xmlns:content='http://purl.org/rss/1.0/modules/content/' xmlns:taxo='http://purl.org/rss/1.0/modules/taxonomy/' xmlns:rdf='http://www.w3.org/1999/02/22-rdf-syntax-ns#' xmlns:itunes='http://www.itunes.com/dtds/podcast-1.0.dtd' xmlns:googleplay="http://www.google.com/schemas/play-podcasts/1.0" xmlns:dc='http://purl.org/dc/elements/1.1/' xmlns:atom='http://www.w3.org/2005/Atom' xmlns:podbridge='http://www.podbridge.com/podbridge-ad.dtd' version='2.0'>
<channel>
  <title>Ding&apos;s posts</title>
  <language>en-us</language>
  <generator>microfeed.org</generator>
  <itunes:type>episodic</itunes:type>
  <itunes:explicit>false</itunes:explicit>
  <atom:link rel="self" href="https://microblog.lidingzeyu.com/rss/" type="application/rss+xml"/>
  <link>https://microblog-lidingzeyu-com.pages.dev</link>
  <description>
    <![CDATA[<p>click this <a href="mailto:ding@dingzeyu.li?subject=Coffee%20Chat!&amp;body=Hi%20Ding,%0A%0AI’d%20like%20to%20chat%20about%20XXX.%0A%0AHere%20are%20some%20time%20slots%20that%20work%20for%20me%20during%208am-5pm%20Pacific%20time:%0A%0ADate/Time%201:%0ADate/Time%202:%0ADate/Time%203:%0A%0ABest,%0A[Your%20Name]" rel="noopener noreferrer" target="_blank">link</a> to schedule a coffee chat!</p>]]>
  </description>
  <itunes:image href="https://microblog-cdn.lidingzeyu.com/microblog-lidingzeyu-com/production/images/channel-227f3a4b4759b5ff8b784b30ec4f8d8b.jpeg"/>
  <image>
    <title>Ding&apos;s posts</title>
    <url>https://microblog-cdn.lidingzeyu.com/microblog-lidingzeyu-com/production/images/channel-227f3a4b4759b5ff8b784b30ec4f8d8b.jpeg</url>
    <link>https://microblog-lidingzeyu-com.pages.dev</link>
  </image>
  <copyright>©2025</copyright>
  <itunes:category text="Technology"/>
  <item>
    <title>Deep Dive into LLMs like ChatGPT - YouTube</title>
    <guid>yNYn1hQ769n</guid>
    <pubDate>Fri, 07 Feb 2025 23:14:39 GMT</pubDate>
    <itunes:explicit>false</itunes:explicit>
    <description>
      <![CDATA[<p>I set aside some time today and yesterday and just finished watching Andrej Karpathy’s latest video on LLMs 101.&nbsp;<a href="https://youtu.be/7xTGNNLPyMI" rel="noopener noreferrer" target="_blank">https://youtu.be/7xTGNNLPyMI</a></p><p><br></p><p>In his video he went over LLM foundation model training, supervised finetuning, and the latest exploratory RL and RLHF training. And there is a&nbsp;<a href="https://youtu.be/7xTGNNLPyMI?t=8927" rel="noopener noreferrer" target="_blank">section</a>&nbsp;where he explains why DeepSeek R1 paper is a big deal (at a high level, yet super helpful). I feel like I have a much better understanding of how RL is helping LLM post-training and how RLHF is (and is not) really RL and its shortcomings.</p><p><br></p><p>I watched the whole video (3.5hours @ 2x) and would recommend it anyone who’s interested in LLM or working with LLM since it will help you understand its quirks.</p><p><br></p><p>Note, I watched many of Andrej’s previous videos. For example, the&nbsp;<a href="https://www.youtube.com/watch?v=zduSFxRajkE" rel="noopener noreferrer" target="_blank">one on tokenization</a>&nbsp;(which explains why asking GPT to output YAML instead of JSON saves more tokens, among other things.) He is a great educator.&nbsp;</p>]]>
    </description>
    <itunes:episodeType>full</itunes:episodeType>
    <enclosure url="https://www.youtube.com/watch?v=7xTGNNLPyMI" type="text/html"/>
  </item>
</channel>
</rss>