Summary of Efficient Llm Inference Using Dynamic Input Pruning and Cache-aware Masking, by Marco Federici et al.
Efficient LLM Inference using Dynamic Input Pruning and Cache-Aware Maskingby Marco Federici, Davide Belli, Mart…
Efficient LLM Inference using Dynamic Input Pruning and Cache-Aware Maskingby Marco Federici, Davide Belli, Mart…
Enhancing Neural Network Robustness Against Fault Injection Through Non-linear Weight Transformationsby Ninnart Fuengfusin, Hakaru TamukohFirst…
Neural Networks Use Distance Metricsby Alan OurslandFirst submitted to arxiv on: 26 Nov 2024CategoriesMain: Machine…
Transformers are Deep Optimizers: Provable In-Context Learning for Deep Model Trainingby Weimin Wu, Maojiang Su,…
Agnostic Learning of Arbitrary ReLU Activation under Gaussian Marginalsby Anxin Guo, Aravindan VijayaraghavanFirst submitted to…
Deriving Activation Functions Using Integrationby Allen Hao Huang, Imanol SchlagFirst submitted to arxiv on: 20…
Hysteresis Activation Function for Efficient Inferenceby Moshe Kimhi, Idan Kashani, Avi Mendelson, Chaim BaskinFirst submitted…
Dense ReLU Neural Networks for Temporal-spatial Modelby Zhi Zhang, Carlos Misael Madrid Padilla, Xiaokai Luo,…
Least Squares Training of Quadratic Convolutional Neural Networks with Applications to System Theoryby Zachary Yetman…
On the Principles of ReLU Networks with One Hidden Layerby Changcun HuangFirst submitted to arxiv…