Summary of Unleash the Power Of Pre-trained Language Models For Irregularly Sampled Time Series, by Weijia Zhang et al.
Unleash The Power of Pre-Trained Language Models for Irregularly Sampled Time Seriesby Weijia Zhang, Chenlong…
Unleash The Power of Pre-Trained Language Models for Irregularly Sampled Time Seriesby Weijia Zhang, Chenlong…
Navigating Data Scarcity using Foundation Models: A Benchmark of Few-Shot and Zero-Shot Learning Approaches in…
A Spitting Image: Modular Superpixel Tokenization in Vision Transformersby Marius Aasan, Odd Kolbjørnsen, Anne Schistad…
CROME: Cross-Modal Adapters for Efficient Multimodal LLMby Sayna Ebrahimi, Sercan O. Arik, Tejas Nama, Tomas…
Perceptual Similarity for Measuring Decision-Making Style and Policy Diversity in Gamesby Chiu-Chou Lin, Wei-Chen Chiu,…
Synthetic Patient-Physician Dialogue Generation from Clinical Notes Using LLMby Trisha Das, Dina Albassam, Jimeng SunFirst…
Efficient and Versatile Robust Fine-Tuning of Zero-shot Modelsby Sungyeon Kim, Boseung Jeong, Donghyun Kim, Suha…
On zero-shot learning in neural state estimation of power distribution systemsby Aleksandr Berezin, Stephan Balduin,…
An Evaluation of Standard Statistical Models and LLMs on Time Series Forecastingby Rui Cao, Qiao…
Leveraging Large Language Models with Chain-of-Thought and Prompt Engineering for Traffic Crash Severity Analysis and…