[go: up one dir, main page]

Follow
Peiyi Wang
Peiyi Wang
DeepSeek AI
Verified email at stu.pku.edu.cn
Title
Cited by
Cited by
Year
DeepSeek-R1 incentivizes reasoning in LLMs through reinforcement learning
D Guo, D Yang, H Zhang, J Song, P Wang, Q Zhu, R Xu, R Zhang, S Ma, ...
Nature 645 (8081), 633-638, 2025
7683*2025
DeepSeekMath: Pushing the Limits of Mathematical Reasoning in Open Language Models
Z Shao, P Wang, Q Zhu, R Xu, J Song, M Zhang, YK Li, Y Wu, D Guo
arXiv preprint arXiv:2402.03300, 2024
3958*2024
Deepseek-v3 technical report
A Liu, B Feng, B Xue, B Wang, B Wu, C Lu, C Zhao, C Deng, C Zhang, ...
arXiv preprint arXiv:2412.19437, 2024
34082024
Large Language Models are not Fair Evaluators
P Wang, L Li, L Chen, D Zhu, B Lin, Y Cao, Q Liu, T Liu, Z Sui
ACL2024, 2023
7472023
Deepseek llm: Scaling open-source language models with longtermism
X Bi, D Chen, G Chen, S Chen, D Dai, C Deng, H Ding, K Dong, Q Du, ...
arXiv preprint arXiv:2401.02954, 2024
7282024
Math-Shepherd: Verify and Reinforce LLMs Step-by-step without Human Annotations
P Wang, L Li, Z Shao, RX Xu, D Dai, Y Li, D Chen, Y Wu, Z Sui
ACL2024, 2023
718*2023
Deepseek-v2: A strong, economical, and efficient mixture-of-experts language model
A Liu, B Feng, B Wang, B Wang, B Liu, C Zhao, C Dengr, C Ruan, D Dai, ...
arXiv preprint arXiv:2405.04434, 2024
6782024
Deepseek-coder-v2: Breaking the barrier of closed-source models in code intelligence
Q Zhu, D Guo, Z Shao, D Yang, P Wang, R Xu, Y Wu, Y Li, H Gao, S Ma, ...
arXiv preprint arXiv:2406.11931, 2024
405*2024
MIT: A Large-Scale Dataset towards Multi-Modal Multilingual Instruction Tuning
L Li, Y Yin, S Li, L Chen, P Wang, S Ren, M Li, Y Yang, J Xu, X Sun, ...
arXiv preprint arXiv:2306.04387, 2023
318*2023
Unlocking efficiency in large language model inference: A comprehensive survey of speculative decoding
H Xia, Z Yang, Q Dong, P Wang, Y Li, T Ge, T Liu, W Li, Z Sui
Findings of ACL2024, 2024
1982024
Incorporating Hierarchy into Text Encoder: a Contrastive Learning Approach for Hierarchical Text Classification
Z Wang, P Wang, L Huang, X Sun, H Wang
ACL2022, 2022
1902022
VLFeedback: A Large-Scale AI Feedback Dataset for Large Vision-Language Models Alignment
L Li, Z Xie, M Li, S Chen, P Wang, L Chen, Y Yang, B Wang, L Kong, Q Liu
EMNLP 2024, 2024
172*2024
Speculative decoding: Exploiting speculative execution for accelerating seq2seq generation
H Xia, T Ge, P Wang, SQ Chen, F Wei, Z Sui
Findings of EMNLP 2023, 2022
150*2022
Inference-time scaling for generalist reward modeling
Z Liu, P Wang, R Xu, S Ma, C Ruan, P Li, Y Liu, Y Wu
arXiv preprint arXiv:2504.02495, 2025
1162025
PCA-Bench: Evaluating Multimodal Large Language Models in Perception-Cognition-Action Chain
L Chen, Y Zhang, S Ren, H Zhao, Z Cai, Y Wang, P Wang, X Meng, T Liu, ...
Findings of ACL2024, 2024
96*2024
HPT: Hierarchy-aware Prompt Tuning for Hierarchical Text Classification
Z Wang, P Wang, T Liu, Y Cao, Z Sui, H Wang
EMNLP2022, 2022
872022
A Two-Stream AMR-enhanced Model for Document-level Event Argument Extraction
R Xu, P Wang, T Liu, S Zeng, B Chang, Z Sui
NAACL2022, 2022
852022
Making large language models better reasoners with alignment
P Wang, L Li, L Chen, F Song, B Lin, Y Cao, T Liu, Z Sui
arXiv preprint arXiv:2309.02144, 2023
772023
An enhanced span-based decomposition method for few-shot sequence labeling
P Wang, R Xu, T Liu, Q Zhou, Y Cao, B Chang, Z Sui
NAACL2022, 2021
732021
Llm critics help catch bugs in mathematics: Towards a better mathematical verifier with natural language feedback
B Gao, Z Cai, R Xu, P Wang, C Zheng, R Lin, K Lu, D Liu, C Zhou, W Xiao, ...
Findings of the Association for Computational Linguistics: ACL 2025, 14588-14604, 2025
442025
The system can't perform the operation now. Try again later.
Articles 1–20