Summary of Vimi: Grounding Video Generation Through Multi-modal Instruction, by Yuwei Fang et al.
VIMI: Grounding Video Generation through Multi-modal Instructionby Yuwei Fang, Willi Menapace, Aliaksandr Siarohin, Tsai-Shien Chen,…
VIMI: Grounding Video Generation through Multi-modal Instructionby Yuwei Fang, Willi Menapace, Aliaksandr Siarohin, Tsai-Shien Chen,…
Double-Ended Synthesis Planning with Goal-Constrained Bidirectional Searchby Kevin Yu, Jihye Roh, Ziang Li, Wenhao Gao,…
Enhanced Safety in Autonomous Driving: Integrating Latent State Diffusion Model for End-to-End Navigationby Detian Chu,…
Noise-Free Explanation for Driving Action Predictionby Hongbo Zhu, Theodor Wulff, Rahul Singh Maharjan, Jinpei Han,…
Large Language Model Recall Uncertainty is Modulated by the Fan Effectby Jesse Roberts, Kyle Moore,…
Interactively Diagnosing Errors in a Semantic Parserby Constantine Nakos, Kenneth D. ForbusFirst submitted to arxiv…
Knowledge Management in the Companion Cognitive Architectureby Constantine Nakos, Kenneth D. ForbusFirst submitted to arxiv…
AI-driven multi-omics integration for multi-scale predictive modeling of causal genotype-environment-phenotype relationshipsby You Wu, Lei XieFirst…
Exploring the Capability of ChatGPT to Reproduce Human Labels for Social Computing Tasks (Extended Version)by…
InsightBench: Evaluating Business Analytics Agents Through Multi-Step Insight Generationby Gaurav Sahu, Abhay Puri, Juan Rodriguez,…