Seuraa
Jue Wang
Jue Wang
Together AI; ZJU
Vahvistettu sähköpostiosoite verkkotunnuksessa zju.edu.cn - Kotisivu
Nimike
Viittaukset
Viittaukset
Vuosi
Holistic evaluation of language models
P Liang, R Bommasani, T Lee, D Tsipras, D Soylu, M Yasunaga, Y Zhang, ...
arXiv preprint arXiv:2211.09110, 2022
9342022
Two are Better than One: Joint Entity and Relation Extraction with Table-Sequence Encoders
J Wang, W Lu
Empirical Methods in Natural Language Processing, 1706--1721, 2020
2422020
Pyramid: A Layered Model for Nested Named Entity Recognition
J Wang, L Shou, K Chen, G Chen
Association for Computational Linguistics, 5918–5928, 2020
1632020
Deja vu: Contextual sparsity for efficient llms at inference time
Z Liu, J Wang, T Dao, T Zhou, B Yuan, Z Song, A Shrivastava, C Zhang, ...
1582023
Continual Federated Learning Based on Knowledge Distillation.
Y Ma, Z Xie, J Wang, K Chen, L Shou
IJCAI, 2182-2188, 2022
432022
Draft & verify: Lossless large language model acceleration via self-speculative decoding
J Zhang, J Wang, H Li, L Shou, K Chen, G Chen, S Mehrotra
arXiv preprint arXiv:2309.08168, 2023
362023
Skill-it! a data-driven skills framework for understanding and training language models
M Chen, N Roberts, K Bhatia, J Wang, C Zhang, F Sala, C Ré
Advances in Neural Information Processing Systems 36, 2024
322024
CocktailSGD: Fine-tuning Foundation Models over 500Mbps Networks
J Wang, Y Lu, B Yuan, B Chen, P Liang, C De Sa, C Re, C Zhang
252023
Compress, then prompt: Improving accuracy-efficiency trade-off of llm inference with transferable prompt
Z Xu, Z Liu, B Chen, Y Tang, J Wang, K Zhou, X Hu, A Shrivastava
arXiv preprint arXiv:2305.11186, 2023
252023
Skipbert: Efficient inference with shallow layer skipping
J Wang, K Chen, G Chen, L Shou, J McAuley
Proceedings of the 60th Annual Meeting of the Association for Computational …, 2022
212022
Fine-tuning language models over slow networks using activation quantization with guarantees
J Wang, B Yuan, L Rimanic, Y He, T Dao, B Chen, C Ré, C Zhang
Advances in Neural Information Processing Systems 35, 19215-19230, 2022
16*2022
Mixture-of-Agents Enhances Large Language Model Capabilities
J Wang, J Wang, B Athiwaratkun, C Zhang, J Zou
arXiv preprint arXiv:2406.04692, 2024
142024
Effective Slot Filling via Weakly-Supervised Dual-Model Learning
J Wang, K Chen, L Shou, S Wu, G Chen
AAAI Conference on Artificial Intelligence 35 (16), 13952-13960, 2021
52021
Effective Continual Learning for Text Classification with Lightweight Snapshots
J Wang, D Dong, L Shou, K Chen, G Chen
AAAI Conference on Artificial Intelligence 37, 10122-10130, 2023
32023
Semi-supervised few-shot learning for dual question-answer extraction
J Wang, K Chen, L Shou, S Wu, S Mehrotra
arXiv preprint arXiv:1904.03898, 2019
22019
Learning Label-Adaptive Representation for Large-Scale Multi-Label Text Classification
C Peng, H Wang, J Wang, L Shou, K Chen, G Chen, C Yao
IEEE/ACM Transactions on Audio, Speech, and Language Processing, 2024
12024
SMILE: A Cost-Effective System for Serving Massive Pretrained Language Models in The Cloud
J Wang, K Chen, L Shou, D Jiang, G Chen
Companion of the 2023 International Conference on Management of Data, 135-138, 2023
2023
Järjestelmä ei voi suorittaa toimenpidettä nyt. Yritä myöhemmin uudelleen.
Artikkelit 1–17