Xiuying Wei
Google Scholar / GitHub / Email /
xiuying.wei [at] epfl.ch, weixiuying966 [at] gmail.com
                                  

Hi, this is Xiuying Wei. She is now a second-year PhD student in the CLAIRE lab at EPFL, supervised by Prof. Caglar Gulcehre. Her research interests include efficient machine learning and natural language processing. She is now mainly focusing on model architecture design but is also trying to touch more research fields.

Before that, she earned her master's degree at Beihang University and started her research about deep learning, specifically quantization. She earned her bachelor's degree at Shandong University and spent much time in algorithm competition.

Publications

Building on Efficient Foundations: Effectively Training LLMs with Structured Feedforward Layers
Xiuying Wei, Skander Moalla, Razvan Pascanu, Caglar Gulcehre.
Neural Information Processing Systems (NeurIPS), 2024

Investigate three structured linear parameterizations in transformer language models: 1)scaling law study and model size scaling, 2)efficiency and pre-merge technique, 3)optimization and self-guided training

[Paper] [Code]

Outlier Suppression+: Accurate quantization of large language models by equivalent and effective shifting and scaling
Xiuying Wei, Yunchen Zhang, Yuhang Li, Xianguo Zhang, Ruihao Gong, Jinyang Guo, Xianglong Liu.
EMNLP23
[Paper] [Code]

Lossy and Lossless (L2) Post-training Model Size Compression
Yumeng Xue, Shihao Bai, Xiuying Wei, Ruihao Gong, Jianlei Yang
International Conference on Computer Vision (ICCV), 2023

Integrate lossless and lossy compression techniques in a post-training setting.

[Paper]

Outlier Suppression: Pushing the Limit of Low-bit Transformer Language Models
Xiuying Wei, Yunchen Zhang, Xiangguo Zhang, Ruihao Gong, Shanghang Zhang, Qi Zhang, Fengwei Yu, and Xianglong Liu
Neural Information Processing Systems (NeurIPS), 2022 (Spotlight)

Identify outlier phenomenons (channel concentration and token discrepancy) for quantizing transformer language models. Propose a framework to suppress these outliers.

[Paper] [Code]

QDrop: Randomly Dropping Quantization For Extremely Low-bit Post-training quantization.
Xiuying Wei, Ruihao Gong, Yuhang Li, Xianglong Liu, and Fengwei Yu
International Conference on Learning Representations (ICLR), 2022, with 8688 scores by reviewers

Investigate how the activation quantization affects weight tuning. Build the relationship between activation quantization and flatness of quantized weights. Propose to randomly drop the activation quantization to achieve a flatter optimized weights.

[Paper] [Code]


Honors and Awards
Misc
In her spare time, she reads a lot of novels and loves to explore the lifestyle.