Summary of Multi-round Jailbreak Attack on Large Language Models, by Yihua Zhou et al.
Multi-round jailbreak attack on large language modelsby Yihua Zhou, Xiaochuan ShiFirst submitted to arxiv on:…
Multi-round jailbreak attack on large language modelsby Yihua Zhou, Xiaochuan ShiFirst submitted to arxiv on:…
Can Structured Data Reduce Epistemic Uncertainty?by Shriram M S, Sushmitha S, Gayathri K S, Shahina…
Preserve or Modify? Context-Aware Evaluation for Balancing Preservation and Modification in Text-Guided Image Editingby Yoonjeon…
Improving Data Efficiency via Curating LLM-Driven Rating Systemsby Jinlong Pang, Jiaheng Wei, Ankit Parag Shah,…
Improving the Language Understanding Capabilities of Large Language Models Using Reinforcement Learningby Bokai Hu, Sai…
Thinking LLMs: General Instruction Following with Thought Generationby Tianhao Wu, Janice Lan, Weizhe Yuan, Jiantao…
Derail Yourself: Multi-turn LLM Jailbreak Attack through Self-discovered Cluesby Qibing Ren, Hao Li, Dongrui Liu,…
ForgeryGPT: Multimodal Large Language Model For Explainable Image Forgery Detection and Localizationby Jiawei Liu, Fanrui…
Evaluating Semantic Variation in Text-to-Image Synthesis: A Causal Perspectiveby Xiangru Zhu, Penglei Sun, Yaoxian Song,…
Improving Semantic Understanding in Speech Language Models via Brain-tuningby Omer Moussa, Dietrich Klakow, Mariya TonevaFirst…