Baeseong Park

According to our database1, Baeseong Park authored at least 17 papers between 2019 and 2024.

Collaborative distances:
  • Dijkstra number2 of four.
  • Erdős number3 of four.

Timeline

Legend:

Book 
In proceedings 
Article 
PhD thesis 
Dataset
Other 

Links

On csauthors.net:

Bibliography

2024
HyperCLOVA X Technical Report.
CoRR, 2024

DropBP: Accelerating Fine-Tuning of Large Language Models by Dropping Backward Propagation.
CoRR, 2024

LUT-GEMM: Quantized Matrix Multiplication based on LUTs for Efficient Inference in Large-Scale Generative Language Models.
Proceedings of the Twelfth International Conference on Learning Representations, 2024

2023
Sparsity-Aware Memory Interface Architecture using Stacked XORNet Compression for Accelerating Pruned-DNN Models.
Proceedings of the Sixth Conference on Machine Learning and Systems, 2023

Winning Both the Accuracy of Floating Point Activation and the Simplicity of Integer Arithmetic.
Proceedings of the Eleventh International Conference on Learning Representations, 2023

TF-MVP: Novel Sparsity-Aware Transformer Accelerator with Mixed-Length Vector Pruning.
Proceedings of the 60th ACM/IEEE Design Automation Conference, 2023

2022
nuQmm: Quantized MatMul for Efficient Inference of Large-Scale Generative Language Models.
CoRR, 2022

Encoding Weights of Irregular Sparsity for Fixed-to-Fixed Model Compression.
Proceedings of the Tenth International Conference on Learning Representations, 2022

AlphaTuning: Quantization-Aware Parameter-Efficient Adaptation of Large-Scale Pre-Trained Language Models.
Proceedings of the Findings of the Association for Computational Linguistics: EMNLP 2022, 2022

2021
Modulating Regularization Frequency for Efficient Compression-Aware Model Training.
CoRR, 2021

Sequential Encryption of Sparse Neural Networks Toward Optimum Representation of Irregular Sparsity.
CoRR, 2021

Q-Rater: Non-Convex Optimization for Post-Training Uniform Quantization.
CoRR, 2021

2020
BiQGEMM: matrix multiplication with lookup table for binary-coding-based quantized DNNs.
Proceedings of the International Conference for High Performance Computing, 2020

FleXOR: Trainable Fractional Quantization.
Proceedings of the Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020, 2020

Extremely Low Bit Transformer Quantization for On-Device Neural Machine Translation.
Proceedings of the Findings of the Association for Computational Linguistics: EMNLP 2020, 2020

Structured Compression by Weight Encryption for Unstructured Pruning and Quantization.
Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2020

2019
Structured Compression by Unstructured Pruning for Sparse Quantized Neural Networks.
CoRR, 2019


  Loading...