Zecheng Tang

Zecheng Tang (汤泽成)




Brief Bio

I am currently a fourth-year PhD candidate at Soochow University (exp. grad. Jul 2027), supervised by Assoc. Prof. Juntao Li and Prof. Min Zhang. My research focuses on Machine Learning and Foundation Model (with Algorithm-Architecture co-design). My contributions include but not limited to: Qwen-Image, Visual ChatGPT, NUWA-series (LayoutNUWA/StrokeNUWA), and OpenBA-series (OpenBA-V1/OpenBA-V2).

I have been also conducting research on algorithm optimization (model architecture, post-training) and infrastructure building (evaluation, training platforms) for scalable Long Sequence Modeling and Aligning at LCM-Lab, with 9 publications (currently) in this field. See my project homepage for a complete list.

Before that, I obtained my Bachelor's degree in Software Engineering from Soochow University (ranked 1st / 380). My formal name is Zecheng Tang (汤泽成), and you may call me Zecheng (/zə'tʃɛŋ/), or simply Tang.


Work Experience

2025.02 - 2026.01
Alibaba Cloud (Tongyi-Qwen)
Research Intern
Rep. Work: Qwen-Image (tech report)
(Foundation Generative Model & AI-Infra & Post-training)
2024.10 - 2025.01
Shanghai AI Lab (OpenDataLab)
Academic Collaboration
Rep. Work: CDT (ICLR 2026)
(Long-context Modeling and Optimization)
2024.05 - 2024.09
StepFun AI (Sora Team)
Research Intern
Rep. Work: Step-Video-T2V (tech report)
(Video Generation & Evaluation)
2023.01 - 2024.02
Microsoft Research Asia (NLC Group)
Research Intern
Rep. Work: Visual ChatGPT (Github Star 34K), LayoutNUWA (ICLR2024), StrokeNUWA (ICML2025)
(Symbolic Representation in Multimodal LLMs)

Selected Conference & Journal

* indicates equal contribution; Full list: Google Scholar; Rank system: CCF (China Computer Federation)

[Click to view / hide selected list]

Long-context Modeling & Evaluation

  • Revisiting Long-Context Modeling from a Context Denoising Perspective
    Zecheng Tang, Baibei Ji, Juntao Li, Lijun Wu, Haijia Gui, Min Zhang.
    ICLR 2026, CCF A.
    [Paper] [Code] [Media]
  • L-CiteEval: Do Long-Context Models Truly Leverage Context for Responding?
    Zecheng Tang, Keyan Zhou, Juntao Li, Baibei Ji, Jianye Hou, Min Zhang.
    ACL 2025 (Long Main), CCF A.
    [Paper] [Code] [Media]
  • LOGO -- Long cOntext aliGnment via efficient preference Optimization
    Zecheng Tang, Zechen Sun, Juntao Li, Qiaoming Zhu, Min Zhang.
    ICML 2025, CCF A.
    [Paper] [Code] [Media]
  • Open-ended long text generation via masked language modeling
    Xiaobo Liang*, Zecheng Tang*, Juntao Li, Min Zhang.
    ACL 2023 (Long Main), CCF A.
    [Paper] [Code]

Generative Model

  • StrokeNUWA: Tokenizing Strokes for Vector Graphic Synthesis
    Zecheng Tang, Chenfei Wu, Zekai Zhang, Mingheng Ni, Juntao Li, Nan Duan, et al.
    ICML 2024, CCF A.
    [Paper] [Code] [Media]
  • LayoutNUWA: Revealing the hidden layout expertise of large language models
    Zecheng Tang, Chenfei Wu, Juntao Li, Nan Duan.
    ICLR 2024, CCF A.
    [Paper] [Code] [Media]

Foundation Model (& Optimization)

  • OpenBA: An Open-sourced 15B Bilingual Asymmetric seq2seq Model
    Juntao Li*, Zecheng Tang*, Yuyang Ding*, Pinzheng Wang*, Min Zhang, et al.
    SCIS 2025, CCF A.
    [Paper] [Code] [Media]
  • CMD: a framework for Context-aware Model self-Detoxification
    Zecheng Tang, Keyan Zhou, Juntao Li, Yuyang Ding, Pinzheng Wang, Min Zhang, et al.
    EMNLP 2024 (Long Main), CCF B.
    [Paper] [Code] [Media]
  • Improving temporal generalization of pre-trained language models with lexical semantic change
    Zhaochen Su*, Zecheng Tang*, Xinyan Guan, Juntao Li, Lijun Wu, Min Zhang.
    EMNLP 2022 (Long Main), CCF B.
    [Paper] [Code]

Selected Preprint

* indicates equal contribution; Full list: Google Scholar.

[Click to view / hide selected list]

Technical Report

  • Qwen-Image Technical Report
    Core Contributor*
    [Aug 2025] [Arxiv] [Code] [Media]
  • Step-Video-T2V Technical Report
    Contributor*
    [Feb 2025] [Arxiv] [Code] [Media]
  • Visual chatgpt: Talking, drawing and editing with visual foundation models
    Chenfei Wu, Shengming Yin, Weizhen Qi, Xiaodong Wang, Zecheng Tang, Nan Duan.
    [Mar 2023] [Paper] [Code] [Media]

Long-context Modeling & Evaluation

  • Elastic Attention: Test-time Adaptive Sparsity Ratios for Efficient Transformers
    Zecheng Tang, Zecheng Tang, Quantong Qiu, Yi Yang, Zhiyi Hong, Haiya Xiang, Juntao Li, Min Zhang, et al.
    [Jan 2026] [Arxiv] [Code] [Media]
  • MemoryRewardBench: Benchmarking Reward Models for Long-Term Memory Management in Large Language Models
    Zecheng Tang, Baibei Ji, Ruoxi Sun, Haitian Wang, Juntao Li, Min Zhang, et al.
    [Jan 2026] [Arxiv] [Code] [Media]
  • MMLongCite: A Benchmark for Evaluating Fidelity of Long-Context Vision-Language Models
    Keyan Zhou, Zecheng Tang, Libo Qin, Juntao Li, Min Zhang, et al.
    [Oct 2025] [Arxiv] [Code]
  • LongRM: Revealing and Unlocking the Context Boundary of Reward Modeling
    Zecheng Tang, Baibei Ji, Quantong Qiu, Haitian Wang, Xiaobo Liang, Juntao Li, Min Zhang.
    [Oct 2025] [Arxiv] [Code] [Media]
  • LOOM-Scope: LOng-cOntext Model evaluation framework
    Zecheng Tang, Haitian Wang, Quantong Qiu, Baibei Ji, Ruoxi Sun, Keyan Zhou, Juntao Li, Min Zhang
    [Jul 2025] [Arxiv] [Code] [Media]

Honor & Award

Key terminology translation reference: Chinese Fund Translation.


Talk


Service

I have been serving as a reviewer for the following conferences:
Homepage last updated: 2026-01 | CV last updated: 2025-12 | Template adapted from dpkingma.com