Skip to content

Commit

Permalink
Update index.md
Browse files Browse the repository at this point in the history
  • Loading branch information
TaraSPande authored Dec 7, 2024
1 parent d1fc47d commit e780625
Showing 1 changed file with 1 addition and 1 deletion.
2 changes: 1 addition & 1 deletion index.md
Original file line number Diff line number Diff line change
Expand Up @@ -74,7 +74,7 @@ Large language models (LLMs) have revolutionized a wide range of domains. In par
| Nov 11 | **No Class - Veterans Day** | |
| Nov 18 | **Open-Source and Science in the Era of Foundation Models** <br> Percy Liang, Stanford University <br> <a href="https://rdi.berkeley.edu/llm-agents/assets/percyliang.pdf">Slides</a> [Original Recording](https://bcourses.berkeley.edu/courses/1535641/external_tools/90481) [Edited Video](https://www.youtube.com/live/f3KKx9LWntQ) | - [Cybench: A Framework for Evaluating Cybersecurity Capabilities and Risks of Language Models](https://arxiv.org/abs/2408.08926) |
| Nov 25 | **Measuring Agent capabilities and Anthropic's RSP** <br> Ben Mann, Anthropic <br> <a href="https://rdi.berkeley.edu/llm-agents/assets/antrsp.pdf">Slides</a> [Original Recording](https://bcourses.berkeley.edu/courses/1535641/external_tools/90481) [Edited Video](https://www.youtube.com/live/6y2AnWol7oo) | - [Announcing our updated Responsible Scaling Policy](https://www.anthropic.com/news/announcing-our-updated-responsible-scaling-policy) <br> - [Developing a computer use model](https://www.anthropic.com/news/developing-computer-use) |
| Dec 2 | **Towards Building Safe & Trustworthy AI Agents and A Path for Science‑ and Evidence‑based AI Policy** <br> Dawn Song, UC Berkeley <br> Slides posted soon. [Edited Video](https://www.youtube.com/live/QAgR4uQ15rc) | - [Decoding Trust: Comprehensive Assessment of Trustworthiness in GPT Models](https://decodingtrust.github.io/) <br> - [Extracting Training Data from Large Language Models](https://arxiv.org/abs/2012.07805) <br> - [Robust Physical-World Attacks on Deep Learning Models](https://arxiv.org/abs/1707.08945) <br> *All readings are optional this week.* |
| Dec 2 | **Towards Building Safe & Trustworthy AI Agents and A Path for Science‑ and Evidence‑based AI Policy** <br> Dawn Song, UC Berkeley <br> Slides posted soon. [Edited Video](https://www.youtube.com/live/QAgR4uQ15rc) | - [A Path for Science‑ and Evidence‑based AI Policy](https://understanding-ai-safety.org/) <br> - [DecodingTrust: A Comprehensive Assessment of Trustworthiness in GPT Models](https://arxiv.org/abs//2306.11698) <br> - [Representation Engineering: A Top-Down Approach to AI Transparency](https://arxiv.org/abs/2310.01405) <br> - [Extracting Training Data from Large Language Models](https://www.usenix.org/system/files/sec21-carlini-extracting.pdf) <br> - [The Secret Sharer: Evaluating and Testing Unintended Memorization in Neural Networks](https://www.usenix.org/system/files/sec19-carlini.pdf) <br> *All readings are optional this week.* |


## Enrollment and Grading
Expand Down

0 comments on commit e780625

Please sign in to comment.