Shiwei Liu
Orcid: 0000-0001-6195-771XAffiliations:
- University of Oxford, Mathematical Institute, UK
- University of Texas at Austin, TX, USA
- Eindhoven University of Technology, Eindhoven, The Netherlands (PhD)
According to our database1,
Shiwei Liu
authored at least 66 papers
between 2019 and 2024.
Collaborative distances:
Collaborative distances:
Timeline
Legend:
Book In proceedings Article PhD thesis Dataset OtherLinks
Online presence:
-
on twitter.com
-
on orcid.org
-
on github.com
On csauthors.net:
Bibliography
2024
CoRR, 2024
(PASS) Visual Prompt Locates Good Structure Sparsity through a Recurrent HyperNetwork.
CoRR, 2024
From GaLore to WeLore: How Low-Rank Weights Non-uniformly Emerge from Low-Rank Gradients.
CoRR, 2024
Q-GaLore: Quantized GaLore with INT4 Projection and Layer-Adaptive Low-Rank Gradients.
CoRR, 2024
MSRS: Training Multimodal Speech Recognition Models from Scratch with Sparse Mask Optimization.
CoRR, 2024
Towards Efficient Pareto Set Approximation via Mixture of Experts Based Model Fusion.
CoRR, 2024
OwLore: Outlier-weighed Layerwise Sampled Low-Rank Projection for Memory-Efficient LLM Fine-tuning.
CoRR, 2024
Found in the Middle: How Language Models Use Long Contexts Better via Plug-and-Play Positional Encoding.
CoRR, 2024
Q-Hitter: A Better Token Oracle for Efficient LLM Inference via Sparse-Quantized KV Cache.
Proceedings of the Seventh Annual Conference on Machine Learning and Systems, 2024
Proceedings of the Forty-first International Conference on Machine Learning, 2024
Proceedings of the Forty-first International Conference on Machine Learning, 2024
Outlier Weighed Layerwise Sparsity (OWL): A Missing Secret Sauce for Pruning LLMs to High Sparsity.
Proceedings of the Forty-first International Conference on Machine Learning, 2024
Junk DNA Hypothesis: Pruning Small Pre-Trained Weights Irreversibly and Monotonically Impairs "Difficult" Downstream Tasks in LLMs.
Proceedings of the Forty-first International Conference on Machine Learning, 2024
Proceedings of the Forty-first International Conference on Machine Learning, 2024
Proceedings of the Twelfth International Conference on Learning Representations, 2024
Proceedings of the Twelfth International Conference on Learning Representations, 2024
Proceedings of the Twelfth International Conference on Learning Representations, 2024
FFN-SkipLLM: A Hidden Gem for Autoregressive Decoding with Adaptive Feed Forward Skipping.
Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing, 2024
Is C4 Dataset Optimal for Pruning? An Investigation of Calibration Data for LLM Pruning.
Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing, 2024
2023
Int. J. Comput. Vis., October, 2023
Trans. Mach. Learn. Res., 2023
The Counterattack of CNNs in Self-Supervised Learning: Larger Kernel Size might be All You Need.
CoRR, 2023
E2ENet: Dynamic Sparse Feature Fusion for Accurate and Efficient 3D Medical Image Segmentation.
CoRR, 2023
CoRR, 2023
Outlier Weighed Layerwise Sparsity (OWL): A Missing Secret Sauce for Pruning LLMs to High Sparsity.
CoRR, 2023
Junk DNA Hypothesis: A Task-Centric Angle of LLM Pre-trained Weights through Sparsity.
CoRR, 2023
Ten Lessons We Have Learned in the New "Sparseland": A Short Handbook for Sparse Neural Network Researchers.
CoRR, 2023
Proceedings of the Machine Learning and Knowledge Discovery in Databases: Research Track, 2023
Proceedings of the Machine Learning and Knowledge Discovery in Databases: Research Track, 2023
Proceedings of the Advances in Neural Information Processing Systems 36: Annual Conference on Neural Information Processing Systems 2023, 2023
The Emergence of Essential Sparsity in Large Pre-trained Models: The Weights that Matter.
Proceedings of the Advances in Neural Information Processing Systems 36: Annual Conference on Neural Information Processing Systems 2023, 2023
Proceedings of the Advances in Neural Information Processing Systems 36: Annual Conference on Neural Information Processing Systems 2023, 2023
Proceedings of the Advances in Neural Information Processing Systems 36: Annual Conference on Neural Information Processing Systems 2023, 2023
Instant Soup: Cheap Pruning Ensembles in A Single Pass Can Draw Lottery Tickets from Large Models.
Proceedings of the International Conference on Machine Learning, 2023
Graph Ladling: Shockingly Simple Parallel GNN Training without Intermediate Communication.
Proceedings of the International Conference on Machine Learning, 2023
Proceedings of the International Conference on Machine Learning, 2023
Proceedings of the Eleventh International Conference on Learning Representations, 2023
Proceedings of the Eleventh International Conference on Learning Representations, 2023
Proceedings of the Eleventh International Conference on Learning Representations, 2023
Proceedings of the Eleventh International Conference on Learning Representations, 2023
Proceedings of the IEEE/CVF International Conference on Computer Vision, 2023
Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2023
Lottery Pools: Winning More by Interpolating Tickets without Increasing Training or Inference Cost.
Proceedings of the Thirty-Seventh AAAI Conference on Artificial Intelligence, 2023
2022
Mach. Learn., 2022
CoRR, 2022
Superposing Many Tickets into One: A Performance Booster for Sparse Neural Network Training.
CoRR, 2022
Proceedings of the Advances in Neural Information Processing Systems 35: Annual Conference on Neural Information Processing Systems 2022, 2022
You Can Have Better Graph Neural Networks by Not Training Weights at All: Finding Untrained GNNs Tickets.
Proceedings of the Learning on Graphs Conference, 2022
The Unreasonable Effectiveness of Random Pruning: Return of the Most Naive Baseline for Sparse Training.
Proceedings of the Tenth International Conference on Learning Representations, 2022
Deep Ensembling with No Overhead for either Training or Testing: The All-Round Blessings of Dynamic Sparsity.
Proceedings of the Tenth International Conference on Learning Representations, 2022
2021
Neural Comput. Appl., 2021
Sparse evolutionary deep learning with over one million artificial neurons on commodity hardware.
Neural Comput. Appl., 2021
FreeTickets: Accurate, Robust and Efficient Deep Ensemble by Training with Dynamic Sparsity.
CoRR, 2021
Proceedings of the Advances in Neural Information Processing Systems 34: Annual Conference on Neural Information Processing Systems 2021, 2021
Do We Actually Need Dense Over-Parameterization? In-Time Over-Parameterization in Sparse Training.
Proceedings of the 38th International Conference on Machine Learning, 2021
Proceedings of the 38th International Conference on Machine Learning, 2021
Proceedings of the Asian Conference on Machine Learning, 2021
2020
Proceedings of the Machine Learning and Knowledge Discovery in Databases, 2020
Proceedings of the Twenty-Ninth International Joint Conference on Artificial Intelligence, 2020
Network Performance Optimization with Real Time Traffic Prediction in Data Center Network.
Proceedings of the European Conference on Optical Communications, 2020
2019
CoRR, 2019