Skip to main content
Search
Utility navigation
Calendar
Contact
Login
MAKE A GIFT
Main navigation
Home
Programs & Events
Research Programs
Workshops & Symposia
Public Lectures
Research Pods
Internal Program Activities
Algorithms, Society, and the Law
People
Scientific Leadership
Staff
Current Long-Term Visitors
Research Fellows
Postdoctoral Researchers
Scientific Advisory Board
Governance Board
Industry Advisory Council
Affiliated Faculty
Science Communicators in Residence
Law and Society Fellows
Participate
Apply to Participate
Plan Your Visit
Location & Directions
Postdoctoral Research Fellowships
Law and Society Fellowships
Science Communicator in Residence Program
Circles
Breakthroughs Workshops and Goldwasser Exploratory Workshops
Support
Annual Fund
Funders
Industrial Partnerships
News & Videos
News
Videos
About
Image
Alignment, Trust, Watermarking, and Copyright Issues in LLMs
Program
Special Year on Large Language Models and Transformers, Part 1
Location
Calvin lab auditorium
Date
Monday, Oct. 14
–
Thursday, Oct. 17, 2024
Back to calendar
Breadcrumb
Home
Workshop & Symposia
Schedule
Secondary tabs
The Workshop
Schedule
Videos
All talks listed in Pacific Time. Schedule subject to change.
Monday, Oct. 14, 2024
9:15
–
9:40 a.m.
Coffee and Check-In
9:40
–
9:45 a.m.
Opening Remarks
9:45
–
10:30 a.m.
Private Retrieval-Augmented Generation
Raluca Popa (UC Berkeley)
10:30
–
11 a.m.
Break
11
–
11:45 a.m.
Scalable Extraction of Training Data from (Production) Language Models
Nicholas Carlini (Google DeepMind)
11:45 a.m.
–
12:30 p.m.
Differentially private synthetic data for private LLM training
Andreas Terzis (Google Deep Mind)
12:30
–
1:45 p.m.
Lunch (on your own)
1:45
–
2 p.m.
A Sociotechnical Approach to A Safe, Responsible AI Future: A Path for Science‑ and Evidence‑based AI Policy
Dawn Song (UC Berkeley)
2
–
2:45 p.m.
Defense against prompt injection attacks
David Wagner (UC Berkeley)
2:45
–
3 p.m.
Break
3
–
3:45 p.m.
Pluralistic Alignment: A Roadmap, Recent Work, and Open Problems
Taylor Sorensen (University of Washington)
3:45
–
4 p.m.
Break
4
–
5 p.m.
Panel Discussion: AI Safety Regulation
Scott Aaronson (UT Austin & OpenAI)
,
Dan Hendrycks (Center for AI Safety)
,
Ion Stoica (UC Berkeley)
,
Martin Casado (a16z)
,
Joseph Gonzalez (UC Berkeley)
Tuesday, Oct. 15, 2024
9:15
–
9:45 a.m.
Coffee and Check-In
9:45
–
10:30 a.m.
Interactive Proofs, Debate, and AI Safety
Jonah Brown-Cohen (Google DeepMind)
10:30
–
11 a.m.
Break
11
–
11:45 a.m.
Prover-Verifier Games Improve Legibility of LLM outputs
Yining Chen (Open AI)
11:45 a.m.
–
12:30 p.m.
Models that prove their own correctness
Orr Paradise (UC Berkeley)
12:30
–
2:15 p.m.
Lunch (on your own)
2:15
–
3 p.m.
On Mitigating Backdoors
Jonathan Shafer (MIT)
3
–
3:15 p.m.
Break
3:15
–
4 p.m.
Formal backdoor detection games and deceptive alignment
Jacob Hilton (Alignment Research Center)
4
–
5 p.m.
Reception
Wednesday, Oct. 16, 2024
9:15
–
9:45 a.m.
Coffee and Check-In
9:45
–
10:30 a.m.
AI Interactions: Misuse, Markets, and Managers
Jacob Steinhardt (UC Berkeley)
10:30
–
11 a.m.
Break
11
–
11:45 a.m.
Beyond Preferences in AI Alignment: Towards Richer Models of Human Reasons and Decisions
Tan Zhi Xuan (MIT)
11:45 a.m.
–
12:30 p.m.
Differential privacy in the clean room: Copyright protections for generative AI
Aloni Cohen (University of Chicago)
12:30
–
2 p.m.
Lunch (on your own)
2
–
2:45 p.m.
Veridical Data Science and Alignment in Medical AI
Bin Yu (UC Berkeley)
2:45
–
3 p.m.
Break
3
–
3:45 p.m.
What should we align with?
Frauke Kreuter (LMU Munich and University of Maryland)
3:45
–
4:30 p.m.
Open Technical Questions in Generative AI Copyright
Yangsibo Huang (Google)
Thursday, Oct. 17, 2024
9
–
9:30 a.m.
Coffee and Check-In
9:30
–
9:45 a.m.
Watermarking: where to?
Scott Aaronson (UT Austin)
9:45
–
10:30 a.m.
Pseudorandom Error-Correcting Codes with Applications to Watermarking Generative AI
Miranda Christ (Columbia University)
10:30
–
11 a.m.
Break
11
–
11:45 a.m.
Edit Distance Robust Watermarks: beyond substitution channels
Noah Golowich (MIT)
11:45 a.m.
–
12:30 p.m.
Distortion-free mechanisms for language model provenance
Rohith Kuditipudi (Stanford University)
12:30
–
2 p.m.
Lunch (on your own)
2:15
–
3 p.m.
Will Copyright Derail Generative AI Technologies?
Pam Samuelson (UC Berkeley)
3
–
3:15 p.m.
Break
3:15
–
4 p.m.
From Risk to Resilience: Risk Assessment, Safety Alignment, and Guardrails for Foundation Models
Bo Li (University of Illinois at Urbana–Champaign)
4
–
5 p.m.
Panel Discussion: Will AI make us stupid and what can we do about it?
Umesh Vazirani (UC Berkeley)
,
Vered Shemtov (Stanford)
,
Trevor Darrell (UC Berkeley)
,
Maroussia Lévesque (Harvard Law School)
,
Richard Zemel (Columbia University)
Share this page
Copy URL of this page
link to homepage
Close
Main navigation
Home
Programs & Events
Research Programs
Workshops & Symposia
Public Lectures
Research Pods
Internal Program Activities
Algorithms, Society, and the Law
People
Scientific Leadership
Staff
Current Long-Term Visitors
Research Fellows
Postdoctoral Researchers
Scientific Advisory Board
Governance Board
Industry Advisory Council
Affiliated Faculty
Science Communicators in Residence
Law and Society Fellows
Participate
Apply to Participate
Plan Your Visit
Location & Directions
Postdoctoral Research Fellowships
Law and Society Fellowships
Science Communicator in Residence Program
Circles
Breakthroughs Workshops and Goldwasser Exploratory Workshops
Support
Annual Fund
Funders
Industrial Partnerships
News & Videos
News
Videos
About
Utility navigation
Calendar
Contact
Login
MAKE A GIFT
link to homepage
Close
Search