Shizhe Diao

Orcid: 0000-0002-3325-9209

According to our database1, Shizhe Diao authored at least 48 papers between 2017 and 2024.

Collaborative distances:
  • Dijkstra number2 of four.
  • Erdős number3 of four.

Timeline

Legend:

Book 
In proceedings 
Article 
PhD thesis 
Dataset
Other 

Links

On csauthors.net:

Bibliography

2024
Grounded-VideoLLM: Sharpening Fine-grained Temporal Grounding in Video Large Language Models.
CoRR, 2024

CodeGraph: Enhancing Graph Reasoning of LLMs with Code.
CoRR, 2024

TheoremLlama: Transforming General-Purpose LLMs into Lean4 Experts.
CoRR, 2024

VeraCT Scan: Retrieval-Augmented Fake News Detection with Justifiable Reasoning.
CoRR, 2024

PLUM: Preference Learning Plus Test Cases Yields Better Code Language Models.
CoRR, 2024

LISA: Layerwise Importance Sampling for Memory-Efficient Large Language Model Fine-Tuning.
CoRR, 2024

Can We Verify Step by Step for Incorrect Answer Detection?
CoRR, 2024

The Instinctive Bias: Spurious Images lead to Hallucination in MLLMs.
CoRR, 2024

UniTime: A Language-Empowered Unified Model for Cross-Domain Time Series Forecasting.
Proceedings of the ACM on Web Conference 2024, 2024

R-Tuning: Instructing Large Language Models to Say 'I Don't Know'.
Proceedings of the 2024 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 1: Long Papers), 2024

LMFlow: An Extensible Toolkit for Finetuning and Inference of Large Foundation Models.
Proceedings of the 2024 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies: System Demonstrations, 2024

SaySelf: Teaching LLMs to Express Confidence with Self-Reflective Rationales.
Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing, 2024

TheoremLlama: Transforming General-Purpose LLMs into Lean4 Experts.
Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing, 2024

FIRST: Teach A Reliable Large Language Model Through Efficient Trustworthy Distillation.
Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing, 2024

Mitigating the Alignment Tax of RLHF.
Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing, 2024

The Instinctive Bias: Spurious Images lead to Illusion in MLLMs.
Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing, 2024

ConstraintChecker: A Plugin for Large Language Models to Reason on Commonsense Knowledge Bases.
Proceedings of the 18th Conference of the European Chapter of the Association for Computational Linguistics, 2024

Arithmetic Control of LLMs for Diverse User Preferences: Directional Preference Alignment with Multi-Objective Rewards.
Proceedings of the 62nd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), 2024

Plum: Prompt Learning using Metaheuristics.
Proceedings of the Findings of the Association for Computational Linguistics, 2024

Active Prompting with Chain-of-Thought for Large Language Models.
Proceedings of the 62nd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), 2024

2023
RAFT: Reward rAnked FineTuning for Generative Foundation Model Alignment.
Trans. Mach. Learn. Res., 2023

Black-Box Prompt Learning for Pre-trained Language Models.
Trans. Mach. Learn. Res., 2023

R-Tuning: Teaching Large Language Models to Refuse Unknown Questions.
CoRR, 2023

Plum: Prompt Learning using Metaheuristic.
CoRR, 2023

MarineGPT: Unlocking Secrets of Ocean to the Public.
CoRR, 2023

Mitigating the Alignment Tax of RLHF.
CoRR, 2023

On the Difference of BERT-style and CLIP-style Text Encoders.
CoRR, 2023

RAFT: Reward rAnked FineTuning for Generative Foundation Model Alignment.
CoRR, 2023

Active Prompting with Chain-of-Thought for Large Language Models.
CoRR, 2023

Hashtag-Guided Low-Resource Tweet Classification.
Proceedings of the ACM Web Conference 2023, 2023

Write and Paint: Generative Vision-Language Models are Unified Modal Learners.
Proceedings of the Eleventh International Conference on Learning Representations, 2023

Towards Unifying Medical Vision-and-Language Pre-training via Soft Prompts.
Proceedings of the IEEE/CVF International Conference on Computer Vision, 2023

Automatic Prompt Augmentation and Selection with Chain-of-Thought from Labeled Data.
Proceedings of the Findings of the Association for Computational Linguistics: EMNLP 2023, 2023

DetGPT: Detect What You Need via Reasoning.
Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, 2023

Doolittle: Benchmarks and Corpora for Academic Writing Formalization.
Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, 2023

Mixture-of-Domain-Adapters: Decoupling and Injecting Domain Knowledge to Pre-trained Language Models' Memories.
Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), 2023

On the Difference of BERT-style and CLIP-style Text Encoders.
Proceedings of the Findings of the Association for Computational Linguistics: ACL 2023, 2023

2022
ExtremeBERT: A Toolkit for Accelerating Pretraining of Customized BERT.
CoRR, 2022

Normalizing Flow with Variational Latent Representation.
CoRR, 2022

Prefix Language Models are Unified Modal Learners.
CoRR, 2022

VLUE: A Multi-Task Benchmark for Evaluating Vision-Language Models.
CoRR, 2022

Black-box Prompt Learning for Pre-trained Language Models.
CoRR, 2022

VLUE: A Multi-Task Multi-Dimension Benchmark for Evaluating Vision-Language Pre-training.
Proceedings of the International Conference on Machine Learning, 2022

2021
Efficient Neural Network Training via Forward and Backward Propagation Sparsification.
Proceedings of the Advances in Neural Information Processing Systems 34: Annual Conference on Neural Information Processing Systems 2021, 2021

Taming Pre-trained Language Models with N-gram Representations for Low-Resource Domain Adaptation.
Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing, 2021

TILGAN: Transformer-based Implicit Latent GAN for Diverse and Coherent Text Generation.
Proceedings of the Findings of the Association for Computational Linguistics: ACL/IJCNLP 2021, 2021

2020
ZEN: Pre-training Chinese Text Encoder Enhanced by N-gram Representations.
Proceedings of the Findings of the Association for Computational Linguistics: EMNLP 2020, 2020

2017
GubaLex: Guba-Oriented Sentiment Lexicon for Big Texts in Finance.
Proceedings of the 13th International Conference on Semantics, Knowledge and Grids, 2017


  Loading...