Summary of Perft: Parameter-efficient Routed Fine-tuning For Mixture-of-expert Model, by Yilun Liu et al.
PERFT: Parameter-Efficient Routed Fine-Tuning for Mixture-of-Expert Modelby Yilun Liu, Yunpu Ma, Shuo Chen, Zifeng Ding,…
PERFT: Parameter-Efficient Routed Fine-Tuning for Mixture-of-Expert Modelby Yilun Liu, Yunpu Ma, Shuo Chen, Zifeng Ding,…
PatchCTG: Patch Cardiotocography Transformer for Antepartum Fetal Health Monitoringby M. Jaleed Khan, Manu Vatish, Gabriel…
Interaction Asymmetry: A General Principle for Learning Composable Abstractionsby Jack Brady, Julius von Kügelgen, Sébastien…
Efficient Federated Finetuning of Tiny Transformers with Resource-Constrained Devicesby Kilian Pfeiffer, Mohamed Aboelenien Ahmed, Ramin…
Learning Memory Mechanisms for Decision Making through Demonstrationsby William Yue, Bo Liu, Peter StoneFirst submitted…
Circuit Complexity Bounds for RoPE-based Transformer Architectureby Bo Chen, Xiaoyu Li, Yingyu Liang, Jiangxuan Long,…
Navigation with QPHIL: Quantizing Planner for Hierarchical Implicit Q-Learningby Alexi Canesse, Mathieu Petitbois, Ludovic Denoyer,…
Automatic Album Sequencingby Vincent Herrmann, Dylan R. Ashley, Jürgen SchmidhuberFirst submitted to arxiv on: 12…
MSEG-VCUQ: Multimodal SEGmentation with Enhanced Vision Foundation Models, Convolutional Neural Networks, and Uncertainty Quantification for…
LAuReL: Learned Augmented Residual Layerby Gaurav Menghani, Ravi Kumar, Sanjiv KumarFirst submitted to arxiv on:…