Summary of Semester

Here’s a summary of the topics for the semester: Week 1: Introduction Attention, Transformers, and BERT Training LLMs, Risks and Rewards Week 2: Alignment Introduction to AI Alignment and Failure Cases Redteaming Jail-breaking LLMs Week 3: Prompting and Bias Prompt Engineering Marked Personas Week 4: Capabilities of LLMs LLM Capabilities Medical Applications of LLMs Week 5: Hallucination Hallucination Risks Potential Solutions Week 6: Visit from Anton Korinek Week 7: Generative Adversarial Networks and DeepFakes

Read More…

Week 14b: Ethical AI

Presenting Team: Aparna Kishore, Elena Long, Erzhen Hu, Jingping Wan Blogging Team: Haolin Liu, Haochen Liu, Ji Hyun Kim, Stephanie Schoch, Xueren Ge Note: since the topics were unrelated, Week 14 is split into two posts: Monday, November 27: Multimodal Models Wednesday, November 29: Ethical AI Wednesday, November 29: Ethical AI Ben Shneiderman. Bridging the Gap Between Ethics and Practice: Guidelines for Reliable, Safe, and Trustworthy Human-centered AI Systems. ACM Transactions on Interactive Intelligent Systems, October 2020.

Read More…

Week 14a: Multimodal Models

Presenting Team: Aparna Kishore, Elena Long, Erzhen Hu, Jingping Wan Blogging Team: Haolin Liu, Haochen Liu, Ji Hyun Kim, Stephanie Schoch, Xueren Ge Note: since the topics were unrelated, Week 14 is split into two posts: Monday, November 27: Multimodal Models Wednesday, November 29: Ethical AI Monday, November 27: Multimodal Models Today’s topic is how to improve model performance by combining multiple modes. We will first introduce the multimodal foundations and then center around CLIP, which is the most famous vision-language model.

Read More…

Week 13: Regulating Dangerous Technologies

The slides are here: Regulating Dangerous Technologies (I’ve included some slides in the posted slides that I didn’t present in class but you might find interesting, including some excerpts from a talk I gave in 2018 on Mutually Assured Destruction and the Impending AI Apocalypse.) Since one of the groups made the analogy to tobacco products, I also will take the liberty of pointing to a talk I gave at Google making a similar analogy: The Dragon in the Room.

Read More…

Week 12: LLM Agents

Presenting Team: Liu Zhe, Peng Wang, Sikun Guo, Yinhan He, Zhepei Wei Blogging Team: Anshuman Suri, Jacob Christopher, Kasra Lekan, Kaylee Liu, My Dinh Monday, November 13: LLM Agents LLM agents are the “next big thing”, with the potential to directly impact important fields like healthcare and education. Essentially, they are LLM-based systems that have the ability to use external tools, such as Internet browsing access and calculators, to augment their abilities.

Read More…

Week 11: Watermarking on Generative Models

Presenting Team: Tseganesh Beyene Kebede, Zihan Guan, Xindi Guo, Mengxuan Hu Blogging Team: Ajwa Shahid, Caroline Gihlstorf, Changhong Yang, Hyeongjin Kim, Sarah Boyce Monday, November 6: Watermarking LLM Outputs Recent instances of AI-generated text passing for human text and the writing of students being misattributed to AI suggest the need for a tool to distinguish between human-written and AI-generated text. The presenters also noted that the increase in the amount of AI-generated text online is a risk for training future LLMs on this data.

Read More…

Week 10: Data Selection for LLMs

(see bottom for assigned readings and questions) Presenting Team: Haolin Liu, Xueren Ge, Ji Hyun Kim, Stephanie Schoch Blogging Team: Aparna Kishore, Elena Long, Erzhen Hu, Jingping Wan Monday, 30 October: Data Selection for Fine-tuning LLMs Question: Would more models help? We’ve discussed so many risks and issues of GenAI so far and one question is that it can be difficult for us to come up with a possible solution to these problems.

Read More…

Week 9: Interpretability

(see bottom for assigned readings and questions) Presenting Team: Anshuman Suri, Jacob Christopher, Kasra Lekan, Kaylee Liu, My Dinh Blogging Team: Hamza Khalid, Liu Zhe, Peng Wang, Sikun Guo, Yinhan He, Zhepei Wei Monday, 23 October: Interpretability: Overview, Limitations, & Challenges Definition of Interpretability Interpretability in the context of artificial intelligence (AI) and machine learning refers to the extent to which a model’s decisions, predictions, or internal workings can be understood and explained by humans.

Read More…

Week 8: Machine Translation

(see bottom for assigned readings and questions) Machine Translation (Week 8) Presenting Team: Ajwa Shahid, Caroline Gihlstorf, Changhong Yang, Hyeongjin Kim, Sarah Boyce Blogging Team: Xindi Guo, Mengxuan Hu, Tseganesh Beyene Kebede, Zihan Guan Monday, 16 Oct: Diving into the History of Machine Translation Let’s kick off this topic with an activity that involves translating an English sentence into a language of your choice and subsequently composing pseudocode to describe the process.

Read More…

Week 7: GANs and DeepFakes

(see bottom for assigned readings and questions) Presenting Team: Aparna Kishore, Elena Long, Erzhen Hu, Jingping Wan Blogging Team: Haochen Liu, Haolin Liu, Ji Hyun Kim, Stephanie Schoch, Xueren Ge Monday, 9 October: Generative Adversarial Networks and DeepFakes Today's topic is how to utilize generative adversarial networks to create fake images and how to identify the images generated by these models. Generative Adversarial Network (GAN) is a revolutionary deep learning framework that pits two neural networks against each other in a creative showdown.

Read More…

All Posts by Category or Tags.