Summary of Can Language Models Explain Their Own Classification Behavior?, by Dane Sherburn et al.
Can Language Models Explain Their Own Classification Behavior?by Dane Sherburn, Bilal Chughtai, Owain EvansFirst submitted…
Can Language Models Explain Their Own Classification Behavior?by Dane Sherburn, Bilal Chughtai, Owain EvansFirst submitted…
MedConceptsQA: Open Source Medical Concepts QA Benchmarkby Ofir Ben Shoham, Nadav RappoportFirst submitted to arxiv…
Automating Code Adaptation for MLOps – A Benchmarking Study on LLMsby Harsh Patel, Buvaneswari A.…
Reddit-Impacts: A Named Entity Recognition Dataset for Analyzing Clinical and Social Effects of Substance Use…
NaturalCodeBench: Examining Coding Performance Mismatch on HumanEval and Natural User Promptsby Shudan Zhang, Hanlin Zhao,…
Evaluating Text Summaries Generated by Large Language Models Using OpenAI’s GPTby Hassan Shakil, Atqiya Munawara…
Utilizing GPT to Enhance Text Summarization: A Strategy to Minimize Hallucinationsby Hassan Shakil, Zeydy Ortiz,…
Large Language Models Reveal Information Operation Goals, Tactics, and Narrative Framesby Keith Burghardt, Kai Chen,…
Anchored Answers: Unravelling Positional Bias in GPT-2’s Multiple-Choice Questionsby Ruizhe Li, Yanjun GaoFirst submitted to…
Overconfidence is Key: Verbalized Uncertainty Evaluation in Large Language and Vision-Language Modelsby Tobias Groot, Matias…