大数跨境
0
0

AI研讨会 | Towards Trustworthy and Equitable AI in the LLM Era

AI研讨会 | Towards Trustworthy and Equitable AI in the LLM Era 港科广 I 人工智能
2025-11-03
0
导读:Speaker: Dr. Kailong WANG, Associate Professor, Huazhong University of Science and Technology






 研讨会信息

🎤 Speaker: Dr. Kailong WANG,

Associate Professor,

School of CSE at Huazhong University of Science and Technology,

(Ph.D. in Computer Science at National University of Singapore)


📰 Title: Towards Trustworthy and Equitable AI in the LLM Era

(Language: English)


⏰ Time: 2:30 – 3:30 PM, 

📆 Date: Nov. 11, 2025 (Tuesday)


📍 Venue: C2-108, GZ Campus

Online Zoom Meeting:

https://hkust-gz-edu-cn.zoom.us/j/99781303741?pwd=fqOLbfx8MPbCBHHflobKGQzQQoeblH.1

Meeting ID: 997 8130 3741

Passcode: ait



 研讨会概要 



Abstract




As large language models (LLMs) increasingly underpin applications across education, healthcare, finance, and governance, ensuring their trustworthiness and reliability has become a pressing research frontier. This talk presents a systematic exploration of LLM safety and societal alignment, centered on two representative studies. The first focuses on coverage-guided jailbreak detection, which uses neuron activation patterns to find abnormal model behaviors. This method improves jailbreak detection accuracy, helps prioritize risky test cases, and guides test case generation for robust model evaluation. The second focuses on logic-based hallucination detection, which applies logic reasoning and metamorphic testing to spot factual inconsistencies in LLM outputs. The end-to-end detection framework automatically builds benchmark datasets, verifies answers against trusted knowledge bases, and detects fact-conflicting hallucinations with minimal human effort. Building on these foundations, I will outline future directions towards ensuring LLM trustworthiness and reliability, including developing explainable testing frameworks, conducting empirical studies on LLM bias and its influence on human behavior, and designing formally verified reasoning mechanisms. Together, these efforts aim to advance the Greater Bay Area’s vision for safe, transparent, and human-aligned AI ecosystems.





 分享者简介 

Image




Dr. Kailong WANG


Associate Professor,

School of Computer Science & Technology,

Huazhong University of Science and Technology


Dr. Kailong Wang is currently an Associate Professor in the School of CSE at Huazhong University of Science and Technology. He received his Ph.D. in Computer Science from the National University of Singapore and his B.Eng. (First Class Honours) from Nanyang Technological University. His research focuses on trustworthy artificial intelligence, particularly the safety, security, and societal alignment of large language models. Dr. Wang has published extensively in top-tier venues such as OOPSLA, ICSE, FSE, TOSEM, NDSS, ICML and MM. He has received multiple distinctions, including the National Distinguished Postdoctoral Researcher Award and the Outstanding Young Researcher Award (Hubei Province). His broader vision aims to advance safe, transparent, and human-aligned AI to support the Greater Bay Area’s growing innovation ecosystem.




Image

扫描加关注

获取更多AI学域消息

 SCAN & FOLLOW US! 






【声明】内容源于网络
0
0
港科广 I 人工智能
香港科技大学(广州)信息枢纽人工智能学域AI Thrust, Information Hub, HKUST(GZ)
内容 269
粉丝 0
港科广 I 人工智能 香港科技大学(广州)信息枢纽人工智能学域AI Thrust, Information Hub, HKUST(GZ)
总阅读40
粉丝0
内容269