CV

Click the button on the right to get the PDF version. (updated in Nov. 3rd 2025)

Basics

Name Qi Zheng
University Tongji Univeristy, Shanghai, China
Degree PhD Student (expected Aug. 2026)
Email zhengqi97@tongji.edu.cn
Phone / wechat (+86)15300859602
Url https://zhuoshu.github.io/

Education

  • 2022.09 - now

    Shanghai, China

    PhD
    Tongji University
    The Key Laboratory of Embedded System and Service Computing, Ministry of Education
    • Spatial-Temporal Data Science
  • 2019.09 - 2022.09

    Shanghai, China

    Master
    Tongji University
    The Key Laboratory of Embedded System and Service Computing, Ministry of Education
    • Spatial-Temporal Data Science
  • 2015.09 - 2019.09

    Shanghai, China

    Bachelor
    Tongji University
    Department of Computer Science and Technology

Selected Publications

  • 2025
    ST-ReP: Learning Predictive Representations Efficiently for Spatial-Temporal Forecasting
    Proceedings of the AAAI Conference on Artificial Intelligence, AAAI 2025

    Addressed the lack of cross-domain transferability in supervised spatio-temporal forecasting. Designed a self-supervised pretraining framework combining temporal masking and a efficient spatial-temporal encoder. Compared with concurrent SOTA supervised baselines (GPT-ST), ST-ReP achieved achieved 20.24% lower MAE, 67.22% lower GPU usage and 28.07% lower training time. Contribution: Pioneered self-supervised spatio-temporal representation pretraining, bridging efficient forecasting and future foundation-model paradigms.*

  • 2025
    TLAST: A Time-Lag Aware Spatial-Temporal Transformer for Traffic Flow Forecasting
    IEEE Transactions on Intelligent Transportation Systems, 2025

    Solved the inefficiency of Transformers and the neglect of propagation time-lag among sensors. Introduced Time-Lag Embedding and Spatial Proxy Attention (SPA) to capture asynchronous correlations with O(N) complexity. Compared with the SOTA model at publication time (STAEformer), TLAST achieved 2.81% lower MAE, 92.50% lower GPU usage and 91.32% lower training time. Contribution: Established a scalable Transformer paradigm that unifies lag-aware modeling with linear complexity.

  • 2023
    TAGnn: Time Adjoint Graph Neural Network for Traffic Forecasting
    Database Systems for Advanced Applications: 28th International Conference, DASFAA 2023

    Tackled the limitation that prior GNN-based models neglect explicit temporal priors and cross-time spatial correlations. Proposed a time-encoding module and a time-adjoint graph convolution capturing multi-scale temporal-spatial semantics with a concise architecture. Compared with the SOTA model at that time (ASTGNN), TAGnn achieved 9.27% lower MAE and 89.84% lower training time. Contribution: Demonstrated that explicit temporal priors and cross-time adjacency enable interpretable, lightweight, and accurate forecasting.

  • 2022
    DSTAGCN: Dynamic Spatial-Temporal Adjacent Graph Convolutional Network for Traffic Forecasting
    IEEE Transactions on Big Data, 2023

    Addressed the challenge of dynamic and uncertain spatial dependencies in traffic forecasting. Proposed a fuzzy-neural dynamic adjacency generator to construct cross-time graphs and capture evolving spatial-temporal relations using a single GCN layer. Compared with the SOTA method at publication time (STSGCN), DSTAGCN achieved 8.71% lower MAE and 50.47% lower training time. Contribution: Introduced uncertainty-aware, time-varying adjacency learning, realizing both higher predictive accuracy and training efficiency.

Languages

Chinense
Native speaker
English
CET-6

Interests

Spatial-Temporal Data Mining
Expanding-node forecasting
OOD generalization
Traffic Forecasting
Time Series Modeling
Multivariate Time Series forecasting
Deep Learning
Self-supervised learning
High efficiency
LLMs
Domain LLMs
Prompting Learning

Projects

Volunteer

Awards