Knowledge-driven distractor generation for cloze-style multiple choice questions S Ren, KQ Zhu Proceedings of the AAAI conference on artificial intelligence 35 (5), 4339-4347, 2021 | 50 | 2021 |
Multi-turn response selection using dialogue dependency relations Q Jia, Y Liu, S Ren, KQ Zhu, H Tang EMNLP 2020, 2020 | 41 | 2020 |
Symbol-LLM: Towards foundational symbol-centric interface for large language models F Xu, Z Wu, Q Sun, S Ren, F Yuan, S Yuan, Q Lin, Y Qiao, J Liu arXiv preprint arXiv:2311.09278, 2023 | 27 | 2023 |
On the efficacy of eviction policy for key-value constrained generative language model inference S Ren, KQ Zhu arXiv preprint arXiv:2402.06262, 2024 | 14 | 2024 |
Zero-shot faithfulness evaluation for text summarization with foundation language model Q Jia, S Ren, Y Liu, KQ Zhu arXiv preprint arXiv:2310.11648, 2023 | 11 | 2023 |
Taxonomy of abstractive dialogue summarization: scenarios, approaches, and future directions Q Jia, Y Liu, S Ren, KQ Zhu ACM Computing Surveys 56 (3), 1-38, 2023 | 10 | 2023 |
Context Compression for Auto-regressive Transformers with Sentinel Tokens SR , Qi Jia, Kenny Q. Zhu EMNLP 2023, 2023 | 10* | 2023 |
Leaner and Faster: Two-Stage Model Compression for Lightweight Text-Image Retrieval S Ren, KQ Zhu NAACL 2022, 2022 | 6 | 2022 |
EMO: Earth Mover Distance Optimization for Auto-Regressive Language Modeling S Ren, Z Wu, KQ Zhu arXiv preprint arXiv:2310.04691, 2023 | 4 | 2023 |
Low-rank prune-and-factorize for language model compression S Ren, KQ Zhu arXiv preprint arXiv:2306.14152, 2023 | 4 | 2023 |
A controlled study on long context extension and generalization in llms Y Lu, JN Yan, S Yang, JT Chiu, S Ren, F Yuan, W Zhao, Z Wu, AM Rush arXiv preprint arXiv:2409.12181, 2024 | 2 | 2024 |
Pruning pre-trained language models with principled importance and self-regularization S Ren, KQ Zhu arXiv preprint arXiv:2305.12394, 2023 | 2 | 2023 |
Specializing Pre-trained Language Models for Better Relational Reasoning via Network Pruning S Ren, K Zhu Findings of the Association for Computational Linguistics: NAACL 2022, 2195-2207, 2022 | 2 | 2022 |
Boosting LLM via Learning from Data Iteratively and Selectively Q Jia, S Ren, Z Qin, F Xue, J Ni, Y You arXiv preprint arXiv:2412.17365, 2024 | | 2024 |
Combating Short Circuit Behavior in Natural Language Reasoning: Crossover and Mutation Operations for Enhanced Robustness S Huanga, S Renb, KQ Zhuc | | 2023 |