Lu Yin

Affiliations:
  • Eindhoven University of Technology, Eindhoven, The Netherlands


According to our database1, Lu Yin authored at least 34 papers between 2020 and 2024.

Collaborative distances:
  • Dijkstra number2 of four.
  • Erdős number3 of four.

Timeline

Legend:

Book 
In proceedings 
Article 
PhD thesis 
Dataset
Other 

Links

On csauthors.net:

Bibliography

2024
Full-Rank No More: Low-Rank Weight Training for Modern Speech Recognition Models.
CoRR, 2024

Are Sparse Neural Networks Better Hard Sample Learners?
CoRR, 2024

Robust Active Learning (RoAL): Countering Dynamic Adversaries in Active Learning with Elastic Weight Consolidation.
CoRR, 2024

From GaLore to WeLore: How Low-Rank Weights Non-uniformly Emerge from Low-Rank Gradients.
CoRR, 2024

Q-GaLore: Quantized GaLore with INT4 Projection and Layer-Adaptive Low-Rank Gradients.
CoRR, 2024

Dynamic Data Pruning for Automatic Speech Recognition.
CoRR, 2024

MSRS: Training Multimodal Speech Recognition Models from Scratch with Sparse Mask Optimization.
CoRR, 2024

OwLore: Outlier-weighed Layerwise Sampled Low-Rank Projection for Memory-Efficient LLM Fine-tuning.
CoRR, 2024

A Structural-Clustering Based Active Learning for Graph Neural Networks.
Proceedings of the Advances in Intelligent Data Analysis XXII, 2024

Advancing Dynamic Sparse Training by Exploring Optimization Opportunities.
Proceedings of the Forty-first International Conference on Machine Learning, 2024

Outlier Weighed Layerwise Sparsity (OWL): A Missing Secret Sauce for Pruning LLMs to High Sparsity.
Proceedings of the Forty-first International Conference on Machine Learning, 2024

Junk DNA Hypothesis: Pruning Small Pre-Trained Weights Irreversibly and Monotonically Impairs "Difficult" Downstream Tasks in LLMs.
Proceedings of the Forty-first International Conference on Machine Learning, 2024

NeurRev: Train Better Sparse Neural Network Practically via Neuron Revitalization.
Proceedings of the Twelfth International Conference on Learning Representations, 2024

FFN-SkipLLM: A Hidden Gem for Autoregressive Decoding with Adaptive Feed Forward Skipping.
Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing, 2024

Is C4 Dataset Optimal for Pruning? An Investigation of Calibration Data for LLM Pruning.
Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing, 2024

2023
Supervised Feature Selection with Neuron Evolution in Sparse Neural Networks.
Trans. Mach. Learn. Res., 2023

E2ENet: Dynamic Sparse Feature Fusion for Accurate and Efficient 3D Medical Image Segmentation.
CoRR, 2023

Outlier Weighed Layerwise Sparsity (OWL): A Missing Secret Sauce for Pruning LLMs to High Sparsity.
CoRR, 2023

Junk DNA Hypothesis: A Task-Centric Angle of LLM Pre-trained Weights through Sparsity.
CoRR, 2023

REST: Enhancing Group Robustness in DNNs Through Reweighted Sparse Training.
Proceedings of the Machine Learning and Knowledge Discovery in Databases: Research Track, 2023

Enhancing Adversarial Training via Reweighting Optimization Trajectory.
Proceedings of the Machine Learning and Knowledge Discovery in Databases: Research Track, 2023

Dynamic Sparsity Is Channel-Level Sparsity Learner.
Proceedings of the Advances in Neural Information Processing Systems 36: Annual Conference on Neural Information Processing Systems 2023, 2023

Are Large Kernels Better Teachers than Transformers for ConvNets?
Proceedings of the International Conference on Machine Learning, 2023

Lottery Pools: Winning More by Interpolating Tickets without Increasing Training or Inference Cost.
Proceedings of the Thirty-Seventh AAAI Conference on Artificial Intelligence, 2023

2022
Superposing Many Tickets into One: A Performance Booster for Sparse Neural Network Training.
CoRR, 2022

Superposing many tickets into one: A performance booster for sparse neural network training.
Proceedings of the Uncertainty in Artificial Intelligence, 2022

You Can Have Better Graph Neural Networks by Not Training Weights at All: Finding Untrained GNNs Tickets.
Proceedings of the Learning on Graphs Conference, 2022

Semantic-Based Few-Shot Classification by Psychometric Learning.
Proceedings of the Advances in Intelligent Data Analysis XX, 2022

2021
Semantic-Based Few-Shot Learning by Interactive Psychometric Testing.
CoRR, 2021

Sparse Training via Boosting Pruning Plasticity with Neuroregeneration.
Proceedings of the Advances in Neural Information Processing Systems 34: Annual Conference on Neural Information Processing Systems 2021, 2021

Do We Actually Need Dense Over-Parameterization? In-Time Over-Parameterization in Sparse Training.
Proceedings of the 38th International Conference on Machine Learning, 2021

Hierarchical Semantic Segmentation using Psychometric Learning.
Proceedings of the Asian Conference on Machine Learning, 2021

2020
Knowledge Elicitation Using Deep Metric Learning and Psychometric Testing.
Proceedings of the Machine Learning and Knowledge Discovery in Databases, 2020

Beyond Labels: Knowledge Elicitation using Deep Metric Learning and Psychometric Testing.
Proceedings of the Twenty-Ninth International Joint Conference on Artificial Intelligence, 2020


  Loading...