Personalized Dialogue Generation with Persona-Adaptive Attention
Qiushi Huang, Yu Zhang, et al.
AAAI 2023
Deep neural networks (DNNs) can easily be manipulated (by an adversary) to output drastically different predictions and can be done so in a controlled and directed way. This process is known as adversarial attack and is considered one of the major hurdles in using DNNs in high-stakes and real-world applications. Although developing methods to secure DNNs against adversaries is now a primary research focus, it suffers from limitations such as lack of optimization generality and lack of optimization scalability. My research highlights will offer a holistic understanding of optimization foundations for robust AI, peer into their emerging challenges, and present recent solutions developed by my research group.
Qiushi Huang, Yu Zhang, et al.
AAAI 2023
Ravi Tejwani, Yen-ling Kuo, et al.
AAAI 2023
Chih-kai Ting, Karl Munson, et al.
AAAI 2023
Shuang Chen, Herbert Freeman
International Journal of Pattern Recognition and Artificial Intelligence