Transformers, large language models, and their use in physics
This seminar is offered for the MSc students of the Physics and Astronomy department of the Heidelberg university (Computational Physics specialisation). The seminar (SWS: 2; Leistungspunkte: 6; LSF) will be held for the first time in the winter semester 2023/24. The language of the seminar is English. To get credits, participants have to prepare one polished presentation and later submit a written report.
The teacher: Dr Dmitry Kobak is a group leader at the Tübingen University working on dimensionality reduction and self-supervised learning. During the winter semester 2023/24, he is a visiting professor in Heidelberg.
When/where: Thursday 14:15–15:45. INF 205, SR 11.
Schedule
- Nov 9: Andrij Karpathy, Let's build GPT, part 1
- Nov 16: Andrij Karpathy, Let's build GPT, part 2
- Nov 23: Subbarao Kambhampati, Avenging Polanyi's Revenge
- Nov 30: Jyot Makadiya: Evaluating cognitive abilities of LLMs
- Dec 7: Ken von Buenau: Training data memorization in LLMs
- Dec 14: Pit Neitemeier: Mechanistic understanding of LLMs: sparse coding
- Jan 11: Johnly Joshy: Mechanistic understanding of LLMs: attention circuits
- Jan 18: Aditya Rastogi: ViT, CLIP, and MedCLIP
- Jan 25: Philip Velie: Transformers for jet tagging at LHC
- Feb 1: Johannes Schmidt: Transformers for quantum chemistry
Some possible topics
- Introduction
- LLM intelligence
- Emergent abilities of LLMs
- Sparks of AGI
- GPT-4 cannot self-critique / plan
- Systematic cognitive evaluation
- World models debate
- Training data memorization
- In-context learning
- Understanding LLMs
- Mechanistic understanding via sparse coding
- Mechanistic understanding of attention layers
- Grammar learning
- Other LLM topics
- Tiny LLMs
- Time series modelling using LLMs
- Philosophy
- Transformers in other domains
- Vision transformers
- Transformers for LHC data analysis
- Transformers in quantum chemistry