Internlm2 technical report Z Cai, M Cao, H Chen, K Chen, K Chen, X Chen, X Chen, Z Chen, Z Chen, ... arXiv preprint arXiv:2403.17297, 2024 | 210 | 2024 |
Unified demonstration retriever for in-context learning X Li, K Lv, H Yan, T Lin, W Zhu, Y Ni, G Xie, X Wang, X Qiu arXiv preprint arXiv:2305.04320, 2023 | 111 | 2023 |
Full parameter fine-tuning for large language models with limited resources K Lv, Y Yang, T Liu, Q Gao, Q Guo, X Qiu arXiv preprint arXiv:2306.09782, 2023 | 101 | 2023 |
Cont: Contrastive neural text generation C An, J Feng, K Lv, L Kong, X Qiu, X Huang Advances in Neural Information Processing Systems 35, 2197-2210, 2022 | 54* | 2022 |
Adalomo: Low-memory optimization with adaptive learning rate K Lv, H Yan, Q Guo, H Lv, X Qiu arXiv preprint arXiv:2310.10195, 2023 | 13 | 2023 |
Collie: Collaborative training of large language models in an efficient way K Lv, S Zhang, T Gu, S Xing, J Hong, K Chen, X Liu, Y Yang, H Guo, T Liu, ... Proceedings of the 2023 Conference on Empirical Methods in Natural Language …, 2023 | 6 | 2023 |
Longwanjuan: Towards systematic measurement for long text quality K Lv, X Liu, Q Guo, H Yan, C He, X Qiu, D Lin arXiv preprint arXiv:2402.13583, 2024 | 4 | 2024 |
ReAttention: Training-Free Infinite Context with Finite Attention Scope X Liu, R Li, Q Guo, Z Liu, Y Song, K Lv, H Yan, L Li, Q Liu, X Qiu arXiv preprint arXiv:2407.15176, 2024 | 3* | 2024 |