Summary of Multi-round Jailbreak Attack on Large Language Models, by Yihua Zhou et al.
Multi-round jailbreak attack on large language modelsby Yihua Zhou, Xiaochuan ShiFirst submitted to arxiv on:…
Multi-round jailbreak attack on large language modelsby Yihua Zhou, Xiaochuan ShiFirst submitted to arxiv on:…
Improving Data Efficiency via Curating LLM-Driven Rating Systemsby Jinlong Pang, Jiaheng Wei, Ankit Parag Shah,…
Improving the Language Understanding Capabilities of Large Language Models Using Reinforcement Learningby Bokai Hu, Sai…
Can Structured Data Reduce Epistemic Uncertainty?by Shriram M S, Sushmitha S, Gayathri K S, Shahina…
Evaluating Semantic Variation in Text-to-Image Synthesis: A Causal Perspectiveby Xiangru Zhu, Penglei Sun, Yaoxian Song,…
Thinking LLMs: General Instruction Following with Thought Generationby Tianhao Wu, Janice Lan, Weizhe Yuan, Jiantao…
Derail Yourself: Multi-turn LLM Jailbreak Attack through Self-discovered Cluesby Qibing Ren, Hao Li, Dongrui Liu,…
ForgeryGPT: Multimodal Large Language Model For Explainable Image Forgery Detection and Localizationby Jiawei Liu, Fanrui…
Improving Semantic Understanding in Speech Language Models via Brain-tuningby Omer Moussa, Dietrich Klakow, Mariya TonevaFirst…
Controllable Safety Alignment: Inference-Time Adaptation to Diverse Safety Requirementsby Jingyu Zhang, Ahmed Elgohary, Ahmed Magooda,…