AI Safety
Description
With the advancement of systems like GPT, artificial intelligence (AI) techniques are anticipated to significantly impact various aspects of individuals' lives. While these AI techniques have demonstrated remarkable, occasionally superhuman, performance across numerous applications, there is a growing concern regarding their safety and security. It has been shown AI systems are subject to a range of attacks, ranging from adversarial attacks (i.e., perturbing an input slightly causes an AI to make completely wrong predictions), backdoor attacks (i.e., backdoors can be easily embedded in neural networks), and privacy-violating attacks such as membership inference attacks (i.e., an adversary may reliably infer whether a certain sample is used during training or not). In addition, AI systems can inherit or amplify biases present in their training data, potentially leading to unfair or discriminatory outcomes. Furthermore, many AI models, including GPT, are complex and not easily interpretable. It makes understanding how these models make decisions highly nontrivial, even though it is crucial for trust and accountability. This course aims to present a systematic view on the range of AI safety problems that have been identified, analyse their root causes, and study potential approaches to mitigate the safety and security risks. In particular, we will focus on answering two key questions. First, given an AI system, how do we systematically evaluate its safety risk? Second, given an AI system that potentially has safety issues, how do we systematically mitigate the risks? This course will feature real-life AI safety issues on popular AI systems such as ImageNet, GPT and so on.
Requisites
Prerequisites: (CS101/IS111/IS200/COR-IS1704) & (CS420/CS421/IS460) - Pre-req
Co-requisites: None
Anti-requisites: None
Attributes
Department: SCIS
Course Level: Undergraduate
Tracks: CS/IS: Artificial Intelligence Track IS/T4BS: Business Analytics Track
Areas: Business Options Digital Business Electives IS Depth Electives IT Solution Development Electives
Learning Outcomes
Graduate Learning Outcomes
Disciplinary Knowledge, Critical thinking & problem solving, Collaboration and leadership, Ethics and social responsibility, Self-directed learning
Competencies
Software Design, Formal Proof Construction, Research, Security Assessment and Testing, Software Testing