--- library_name: transformers license: apache-2.0 base_model: Qwen/Qwen2.5-7B-Instruct tags: - llama-factory - verl - grpo-training model-index: - name: Psyche-R1 results: [] language: - zh ---
Psyche-R1 logo

Psyche-R1

We propose the first Chinese psychological reasoning LLM that unifies empathy, expertise, and reasoning. This model is a fine-tuned version of [Qwen/Qwen2.5-7B-Instruct](https://huggingface.co/Qwen/Qwen2.5-7B-Instruct) on our proposed dataset encompassing psychological questions paired with detailed rationales, and empathetic single-turn dialogues. We conduct a hybrid training strategy, including SFT and GRPO training. We will present detailed training hyperparameters later. It achieves comparable performance to DeepSeek-R1 on several psychology benchmarks, including psychology counselor examination benchmark (PCEB) proposed by [Hu et al. (2024)](https://ieeexplore.ieee.org/abstract/document/10772313), and CPsyExam test set proposed by [Zhao et al. (2024)](https://aclanthology.org/anthology-files/anthology-files/pdf/coling/2025.coling-main.745.pdf). It also demonstates better performance in empathy on [SoulChat2.0 test set (Xie et al. 2025)](https://aclanthology.org/2025.acl-long.55.pdf). ## Training procedure ### SFT Training hyperparameters The following hyperparameters were used during training: - learning_rate: 1e-05 - seed: 42 - distributed_type: multi-GPU - num_devices: 8 - gradient_accumulation_steps: 16 - total_train_batch_size: 256 - total_eval_batch_size: 8 - optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments - lr_scheduler_type: cosine - lr_scheduler_warmup_ratio: 0.1 - num_epochs: 2.0 ### GRPO Training hyperparameters The following hyperparameters were used during training: - learning_rate: 1e-06 - seed: 42 - distributed_type: multi-GPU - num_devices: 8 - total_train_batch_size: 128 - total_eval_batch_size: 8 - ppo_mini_batch_size: 32 - ppo_micro_batch_size_per_gpu: 20 - kl_loss_coef: 0.001 - lr_scheduler_warmup_steps: 10 - num_epochs: 2.0 ## Usage For quick start, please see [MindIntLab-HFUT/Psyche-R1](https://github.com/MindIntLab-HFUT/Psyche-R1) on GitHub. ## Citation If this work is helpful, please kindly cite as: ```bibtex @article{dai2025psyche, title={Psyche-R1: Towards Reliable Psychological LLMs through Unified Empathy, Expertise, and Reasoning}, author={Dai, Chongyuan and Hu, Jinpeng and Shi, Hongchang and Li, Zhuo and Yang, Xun and Wang, Meng}, journal={arXiv preprint arXiv:2508.10848}, year={2025} } ```