Hi! I am Ruilin Luo. I am currently a third-year master’s student in the Intelligent Interaction Group (IIGroup) at Tsinghua University, supervised by Prof. Yujiu Yang. I received my bachelor’s degree from Huazhong University of Science and Technology, supervised by Prof. Kun He. Prior to that, I spent three happy years of high school at NO.1 Middle School Affiliated to Central China Normal University.

My research interests lie in LLM/MLLM Reasoning and Graph Representation Learning.

🎓 Education

Tsinghua University

  • Sept. 2023 - Present, Master’s Degree in Electronic and Information Engineering, National Scholarship for Graduate Students(2025)

Huazhong University of Science and Technology

  • Sept. 2019 - June 2023, Bachelor’s Degree in Computer Science and Technology, GPA 3.96/4.00.

🎆 News

  • Sept. 2025: Qwen3-VL is released! Try stronger reasoning on Qwen Chat.
  • Sept. 2025: URSA has been accepted by NeurIPS 2025. See you in San Diego! A nice research internship at ByteDance. Congrats to Zhuofan.
  • Aug. 2025: FairTAG has been accepted by EMNLP 2025. See you in Suzhou! The arXiv preprint and code are coming soon.
  • May 2025: Two papers have been accepted to ICML 2025 and ACL 2025, respectively. Congrats to Zicheng and Tianle.
  • Sept. 2024: PTD-SQL has been accepted by EMNLP 2024. See you in Miami. A nice research internship at Tencent.
  • May 2024: UniBi has been accepted by ECML-PKDD 2024 as an oral paper. See you in Vilnius, Lithuania.
  • Feb. 2024: PReSA has been accepted by COLING 2024. See you in Torino, Italy.

💻 Interships

Alibaba Group, QwenVL Team, Hangzhou

  • Mar. 2025 - Present, Research Intern, Multimodal Large Language Model Reasoning.

ByteDance, Beijing

  • Sept. 2024 - Mar. 2025, Research Intern, Multimodal Large Language Model Reasoning.

Tencent, Beijing

  • Jan. 2024 - July 2024, Research Intern, Large Language Model Code Agent.

📝 Publications

Papers on Large Language Model:

ICLR 2026 under review
sym

From Narrow to Panoramic Vision: Attention-Guided Cold-Start Reshapes Multimodal Reasoning

Ruilin Luo*, Chufan Shi*, Yizhen Zhang*, Cheng Yang, Songtao Jiang, Tongkun Guan, Ruizhe Chen, Ruihang Chu, Peng Wang, Mingkun Yang, Lei Wang, Yujiu Yang, Junyang Lin, Zhibo Yang

  • We propose the VAS attention metric and find that the reasoning performance of Multimodal Large Reasoning Models (MLRM) is highly correlated with VAS.
  • We introduce the AVAR training pipeline, which enhances the effectiveness of multimodal reasoning data training by scaling VAS through three stages: data construction, cold-start initialization, and reinforcement learning.
  • Our model achieves state-of-the-art results on reasoning and perception benchmarks at the 7B scale.
NeurIPS 2025
sym

Unlocking Multimodal Mathematical Reasoning via Process Reward Model

Ruilin Luo, Zhuofan Zheng, Yifan Wang, Xinzhe Ni, Zicheng Lin, Songtao Jiang, Yiyao Yu, Chufan Shi, Ruihang Chu, Lei Wang, Jin zeng, Yujiu Yang

[Paper] | [Code] | [HuggingFace]

  • We are the first to propose leveraging a process reward model (PRM) to provide process-level optimization in multimodal mathematical reasoning.
  • We introduce a training framework addressing three key stages: reasoning data scarcity, reward data scaling, and online PRM-integrated reinforcement learning (RL).
  • Our data is used to train Seed1.5-VL, and we open-source URSA-8B-PS-GRPO, a model whose reasoning capability matches that of the InternVL3 series.
EMNLP 2024
sym

PTD-SQL: Partitioning and Targeted Drilling with LLMs in Text-to-SQL

Ruilin Luo, Liyuan Wang, Binghuai Lin, Zicheng Lin, Yujiu Yang

[Paper] | [Code]

  • We propose a code agent framework for the text-to-SQL task that leverages structured grammars for question-type classification, followed by question-type-specific automated question bank construction and a two-layer few-shot retrieval mechanism.
  • We observe that stronger models achieve greater improvements on harder question types when provided with better-matched few-shot chain-of-thought (CoT) prompts, reflecting a learning pattern aligned with human cognition.
  • Our method achieves state-of-the-art (SOTA) performance on benchmarks such as Spider and BIRD.

Papers on LLM for Graph Representation Learning:

EMNLP 2025
sym

Fair Text-Attributed Graph Representation Learning

Ruilin Luo, Tianle Gu, Lin Wang, Yunfeng Zhou, Lei Wang, Yujiu Yang

  • We are the first to identify the bias amplification effect between language model embeddings and GNNs in TAG representation learning.
  • We propose solutions from the perspectives of fine-tuning and offline reinforcement learning.
  • We provide a theoretical foundation for both the amplification of unfairness and our proposed mitigation approaches.
ECML-PKDD 2024 Oral
sym

Prior Bilinear Based Models for Knowledge Graph Completion

Ruilin Luo*, Jiayi Li*, Jiaqi Sun, Jing Xiao, Yujiu Yang

[Paper]

  • We propose the principle of identity in knowledge graph completion (KGC).
  • We introduce a bilinear KGC method that explicitly models the law of identity and outperforms classical approaches such as RESCAL and ComplEx.
  • We provide a theoretical derivation for modeling the law of identity.
COLING 2024
sym

Prior Relational Schema Assists Effective Contrastive Learning for Inductive Knowledge Graph Completion

Ruilin Luo, Jiayi Li, Jianghangfan Zhang, Jing Xiao, Yujiu Yang

[Paper] | [Code]


🏅 Honors and Awards

  • National Scholarship
  • Tsinghua University Comprehensive First-Class Scholarship
  • Huazhong University of Science and Technology Outstanding Student

🎵 Miscellaneous

  • Favorite singers/groups: Wang Leehom, JJ Lin, Jacky Cheung, David Tao, Eric Chou, LE SSERAFIM, (G)I-DLE, BIGBANG, IZ*ONE
  • Hobbies: Cycling, fitness, concert-going, esports (DOTA 2, CS2), singing, badminton