Ruihong Qiu

邱瑞鸿 /`ray hong chill/

I am currently an Assistant Professor and an ARC DECRA fellow (2025-2028) in School of Electrical Engineering and Computer Science (EECS) at The University of Queensland (UQ). Here is my short bio.

My current research mainly focuses on:

  1. LLM post-training.
  2. Agentic RL.
  3. Diffusion LLM.

Previously, I worked on IR, RecSys, GNNs, and GFMs.

Recruitment

I am actively looking for (1-2) self-motivated PhD students in Year 2027, all fully funded!

Recent News

  • 10.2025 Our paper, “Beyond Static LLM Policies: Imitation-Enhanced Reinforcement Learning for Recommendation” is selected as Best Paper Finalist, ICDM 2025.

  • 08.2024 ARC DECRA project funded, “Lifelong Paradigms for Versatile, Robust and Agile Recommender Systems” (2025-2028)

  • 06.2024 Our team gives a talk, “Effective Representation Learning for Legal Case Retrieval”, at IR Seminar, the University of Glasgow. [slides]

  • 05.2024 Our team gives a talk, “Effective Representation Learning for Legal Case Retrieval”, at THUIR, Tsinghua University. [slides]

  • 03.2024 Give a talk, “Graph Learning Methods in Session-based Recommendations and Legal Case Retrieval”, at IRonGraphs Workshop at ECIR 2024. [slides]
  • Past news

Selected Research

Google Scholar page includes the full publication list.

LLM Post-training

PontTuset Break the Block: Dynamic-size Reasoning Blocks for Diffusion Large Language Models via Monotonic Entropy Descent with Reinforcement Learning
Yan Jiang, Ruihong Qiu, Zi Huang
ICML 2026
arXiv / code

We identify a monotonic entropy descent scenario in the block decoding for dLLMs's correct generation. We further develop a RL post-training method to achieve this descent with a dynamic size block mechanism.

PontTuset TRN-R1-Zero: Text-rich Network Reasoning via LLMs with Reinforcement Learning Only
Yilun Liu, Ruihong Qiu, Zi Huang
ACL 2026 (Main, Oral)
arXiv / code

We introduce TRN-R1-Zero, a RL-based LLM post-training method for effective reasoning over text-rich networks, such as citation, hyperlink, social and co-purchase domains.

PontTuset Block-R1: Rethinking the Role of Block Size in Multi-domain Reinforcement Learning for Diffusion Large Language Models
Yan Jiang, Ruihong Qiu, Zi Huang
Preprint
arXiv / code

We introduce a new dataset on block size for dLLMs and a new RL post-training frameworks for dLLMs.

PontTuset When to Commit? Towards Variable-Size Self-Contained Blocks for Discrete Diffusion Language Models
Danny Wang, Ruihong Qiu, Zi Huang
Preprint
arXiv

We formally define future-aware and no-future criteria for self-contained blocks and a novel variable-size self-contained blocks methods for dLLMs.

Out-of-distribution on Graphs

PontTuset What Information Matters? Graph Out-of-Distribution Detection via Tri-Component Information Decomposition
Danny Wang, Ruihong Qiu, Guangdong Bai, Zi Huang
ICML 2026
arXiv / code

We propose TIDE, an information decomposition framework that separates feature, structure, and joint signals in graph neural networks, using an information bottleneck objective to improve OOD detection by filtering spurious information.

PontTuset GFMate: Empowering Graph Foundation Models with Pre-training-agnostic Test-time Prompt Tuning
Yan Jiang, Ruihong Qiu, Zi Huang
ICML 2026
arXiv / code

We propose GFMate, a pre-training-agnostic test-time prompt tuning framework that uses centroid and layer prompts to adapt GFMs by leveraging both labelled and unlabelled target data.

PontTuset Text Meets Topology: Rethinking Out-of-distribution Detection in Text-Rich Networks
Danny Wang, Ruihong Qiu, Guangdong Bai, Zi Huang
EMNLP 2025 (Main)
arXiv / code

We introduce TextTopoOOD, a framework for modeling diverse OOD scenarios on text-rich networks, and propose TNT-OOD, a novel detection method that captures the intricate interplay between text and topology.

PontTuset GOLD: Graph Out-of-Distribution Detection via Implicit Adversarial Latent Generation
Danny Wang, Ruihong Qiu, Guangdong Bai, Zi Huang
ICLR 2025 (Spotlight)
arXiv / OpenReview / code

We propose the GOLD framework for graph OOD detection, an implicit adversarial learning pipeline with synthetic OOD exposure without pre-trained models.

PontTuset Does Homophily Help in Robust Test-time Node Classification?
Yan Jiang, Ruihong Qiu, Zi Huang
WSDM 2026 (Oral)
arXiv / code

We propose the GrapHoST framework for graph learning to conduct test-time graph transformation based on homophily to enhance the robustness of graph models.

Graph Condensation

PontTuset GCondenser: Benchmarking Graph Condensation
Yilun Liu, Ruihong Qiu, Zi Huang
CIKM 2025
arXiv / code

We introduce a benchmark for graph condensation with a thorough methodology development method and an extensive evaluation protocol.

PontTuset PUMA: Efficient Continual Graph Learning with Graph Condensation
Yilun Liu, Ruihong Qiu, Yanran Tang, Hongzhi Yin, Zi Huang
TKDE 2024
arXiv / code

We extend the Condense-and-Train (CaT) continual graph learning algorithm with a more efficient and effective, psudo-label guided memory bank (PUMA🐆) framework.

PontTuset CaT: Balanced Continual Graph Learning with Graph Condensation
Yilun Liu, Ruihong Qiu, Zi Huang
ICDM 2023
arXiv / code

We introduce a Condense-and-Train (CaT🐱) memory-based continual graph learning algorithm using graph condensation to construct a more representative memory bank. And a Train-in-Memory continual learning scheme can further alleviate the imbalanced training issue in Class Incremental Learning.

PontTuset Cassette: Case-to-Case Structural Distillation for Efficient Legal Case Retrieval
Yanran Tang, Ruihong Qiu, Hongzhi Yin, Xue Li, Zi Huang
TOIS 2026
arXiv / code

We develop a distillation strategy for case-to-case graph-based method (see our CaseLink below) acceleration for legal case retrieval.

PontTuset CaseLink: Inductive Graph Learning for Legal Case Retrieval
Yanran Tang, Ruihong Qiu, Hongzhi Yin, Xue Li, Zi Huang
SIGIR 2024
arXiv / code

We introduce an inductive graph learning paradigm for legal case retrieval to tackle the challenge of unseen testing query and candidate cases.

PontTuset CaseGNN++: Graph Contrastive Learning for Legal Case Retrieval with Graph Augmentation
Yanran Tang, Ruihong Qiu, Yilun Liu, Xue Li, Zi Huang
TOIS 2024 (under review)
arXiv / code

We introduce an extended CaseGNN++ method with graph augmentations based on the CaseGNN framework.

PontTuset CaseGNN: Graph Neural Networks for Legal Case Retrieval with Text-Attributed Graphs
Yanran Tang, Ruihong Qiu, Yilun Liu, Xue Li, Zi Huang
ECIR 2024
arXiv / code

We introduce a structural modelling of law case for effective retrieval with the aid of summarisation from LLM and grpah neural networks.

Team

  • Xiaoguang Qiao, UQ EECS PhD (1.2026-, co-advise with Helen Huang and Yujun Cai)
  • Xiaoqian Hu, UQ EECS PhD (10.2025-, co-advise with Sen Wang)
  • Chenke Xu, UQ EECS PhD (7.2025-, co-advise with Xue Li)
  • Boyu Luo, UQ EECS PhD (7.2024-, co-advise with Helen Huang and Guangdong Bai)
  • Yan Jiang, UQ EECS PhD (1.2024-, co-advise with Helen Huang and Guangdong Bai)
  • Danny Wang, UQ EECS PhD (1.2024-, co-advise with Helen Huang and Guangdong Bai)
  • Hrishikesh Patel, UQ EECS PhD (4.2023, co-advise with Sen Wang)
  • Yilun Liu, UQ EECS PhD (1.2023-, co-advise with Helen Huang)

Alumni

  • Yi Zhang, UQ EECS PhD -> SMU Postdoc (7.2024-12.2025, co-advise with Sen Wang and Jiajun Liu)
  • Jingyu Ge, UQ ACWEB PhD -> UQ ACWEB Postdoc (1.2022-12.2025, co-advise with Zhiguo Yuan, Helen Huang, and Jiuling Li)

Service

MISC

I speak Cantonese, Mandarin and English.

Updated on 16/05/2026.