Siqi Fan
Logo Researcher @ AIR, Tsinghua University

I'm Siqi Fan (范嗣祺), a researcher at Institute for AI Industry Research, Tsinghua University (AIR, THU). Previously, I received my M.S. degree from Institute of Automation, Chinese Academy of Sciences (CASIA) in 2022, and my B.E. degree from Shanghai Jiao Tong University (SJTU) in 2019.

I am broadly interested in Representation Learning in Complex Systems, from macro physical world to micro biological world, aiming to create AI agents that perceive like or beyond human. Driven by the dedication to innovation, I aspire to advance the fields of autonomous driving and biomedical discovery, pushing the boundaries of technology to create impactful products.

I love music and visual arts.


Education
  • Shanghai Jiao Tong University (SJTU)
    Shanghai Jiao Tong University (SJTU)
    School of Electronic Information and Electrical Engineering
    B.E. in Automation
    Sep. 2015 - Jul. 2019
  • University of Chinese Academy of Sciences (UCAS)
    University of Chinese Academy of Sciences (UCAS)
    Institute of Automation
    M.S. in Automation
    Sep. 2019 - Jul. 2022
Research Experience
  • Autonomous System Group, Intel Labs China (ILC)
    Autonomous System Group, Intel Labs China (ILC)
    Research Intern
    Aug. 2020 - Dec. 2021
  • Institute of Automation, Chinese Academy of Sciences (CASIA)
    Institute of Automation, Chinese Academy of Sciences (CASIA)
    Student Researcher
    Sep. 2019 - Jul. 2022
  • Institute for AI Industry Research, Tsinghua University (AIR, THU)
    Institute for AI Industry Research, Tsinghua University (AIR, THU)
    Researcher
    Jul. 2022 - Present
Academic Service
Honors & Awards
  • National Scholarship
    2021
  • Pan Deng First-class Scholarship, CAS
    2022
  • Excellent Scholarship, SJTU
    2018
  • China Industrial Intelligence Challenge, State-level Outstanding Award, CAA
    2018
News
2025
Release the OpenBioMed toolkit and an agent platform for biomedicine and life science built on that. News
Mar 07
2024
Our workshop proposal Multi-Agent Embodied Intelligent Systems Meet Generative-AI Era: Opportunities, Challenges and Futures is accepted as a full day workshop @ CVPR'25. Call for Papers
Dec 21
The AI-agent system project PharmAID is launched @ FUSON PHARMA. News
Oct 23
Serve as Area Chair for 1st Workshop on Cooperative Intelligence for Embodied AI @ ECCV'24 News
Oct 12
2023
Release the 1st real-world large-scale dataset for roadside cooperative perception RCooper News
Dec 25
Our ChatDD-FM-100B ranks 1st in all four medical disciplines in C-Eval Benchmark and is the only model with an average score of more than 90. News
Sep 21
Release the 1st commercial-friendly multimodal biomedical foundation model BioMedGPT-10B. News
Aug 18
2022
Give a talk on Traffic Scenes Understanding and Simulation Testing @ ITSC'22.
Sep 18
Research Roadmap (view publications )
Autonomous Driving (AD)
  • Onboard System (Intelligent Vehicle)

    My exploration on vehicle-side environment perception starts from drivable area detection (ITSC’20), and a series perception algorithms are proposed, including a RGB 2D object detection approach (FII-CenterNet, T-VT’21), a semi-supervised learning approach for RGB 2D segmentation (CPCL, T-IP’22), a RGB-T segmentation approach for challenging lighting conditions (SpiderMesh, TechReport’23), and a 3D segmentation approach for large-scale points cloud (SCF-Net, CVPR’21).

  • Roadside System (Intelligent Infrastructure)

    Compared with the well-studied vehicle-side perception, roadside perception has several specific challenges, and the exploration is hindered due to the lack of data. A calibration-free BEV representation network is proposed to address calibration noises caused by inevitable natural factors (CBR, IROS’23). A semantic-geometry decoupled contrastive learning framework is introduced to improve roadside perception performance by leveraging vehicle-side data (IROAM, ICRA’25), and the first real-world large-scale dataset for roadside cooperative perception is released with benchmarks to bloom the research on practical I2I perception (RCooper, CVPR’24).

  • Cooperative Autonomous Driving System (V2X)

    Cooperative perception can effectively enhance individual perception performance by providing additional viewpoint and expanding the sensing field. A scene-level feature cooperative perception approach is proposed (EMIFF, ICRA’24). To enable interpretable instance-level flexible feature interaction, the concept of query cooperation is proposed, and a cooperative perception framework is introduced, which let query stream flow among agents (QUEST, ICRA’24). Besides, motion forecasting can also benefit from learning cooperative trajectory representation ( NeurIPS'24). In addition to focusing on improving individual modules, a pioneering end-to-end cooperative autonomous driving framework is introduced (UniV2X, AAAI’25).

Biomedical Discovery (BD)
  • Biomedical Agent System

    Compared with representation learning in physical world, that for biological modality is more complicated.

  • Human-Agent Interaction System

    Recent advances in LLMs have shed light on the development of knowledgeable and versatile AI research assistants in various scientific domains. Multimodal large language models bridge the semantic gap between natural language and other modalities, including molecule, protein, and vision. A multimodal large language model is proposed for assisting biomedical research (BioMedGPT, J-BHI’24), and optical chemical structure understanding task is introduced and explored for molecule-centric scientific discovery (OCSU, TechReport’25).

  • Multi-Agent Cooperation System

    Multi-agent cooperation is a potential approach to solve complicated scientific research tasks in an autopilot manner. To facilitate the exploration, an agent platform for biomedicine and life science is presented and open-sourced ( OpenBioMed)