Skip to content
View xuyang-liu16's full-sized avatar
πŸ§‘β€πŸ’»
Focusing
πŸ§‘β€πŸ’»
Focusing

Block or report xuyang-liu16

Block user

Prevent this user from interacting with your repositories and sending you notifications. Learn more about blocking users.

You must be logged in to block users.

Maximum 250 characters. Please don't include any personal information such as legal names or email addresses. Markdown supported. This note will be visible to only you.
Report abuse

Contact GitHub support about this user’s behavior. Learn more about reporting abuse.

Report abuse
xuyang-liu16/README.md

🌈 I am Xuyang Liu (εˆ˜ζ—­ζ΄‹), a third-year Master's student at Sichuan University. I am also working as a research intern at OPPO Research Institute, supervised by Prof. Lei Zhang (PolyU, IEEE Fellow). Previously, I have interned at Ant Group focusing on GUI Agent, and Taobao & Tmall Group working on Efficient VLMs. I've also spent half a year visiting MiLAB at Westlake University, supervised by Prof. Donglin Wang. I am fortunate to work closely with Dr. Siteng Huang from DAMO Academy and Prof. Linfeng Zhang from SJTU.

πŸ“Œ My research centers on efficient Large Vision-Language Models (LVLMs), including:

  • πŸ–ΌοΈ Image-Text LVLMs: high-resolution understanding via context compression and fast decoding.
  • 🎬 Video Understanding: long/audio-video, and streaming reasoning via efficient encoding and compression.
  • βš™οΈ Efficiency Toolbox: efficient transfer/fine-tuning and benchmarking for downstream task adaptation.

πŸ“’ If you find these directions interesting, feel free to reach out via email: [email protected].

Pinned Loading

  1. Awesome-Generation-Acceleration Awesome-Generation-Acceleration Public

    πŸ“š Collection of awesome generation acceleration resources.

    388 13

  2. Awesome-Token-level-Model-Compression Awesome-Token-level-Model-Compression Public

    πŸ“š Collection of token-level model compression resources.

    190 8

  3. VidCom2 VidCom2 Public

    [EMNLP 2025 Main] Video Compression Commander: Plug-and-Play Inference Acceleration for Video Large Language Models

    Python 62 3

  4. GlobalCom2 GlobalCom2 Public

    [AAAI 2026] Global Compression Commander: Plug-and-Play Inference Acceleration for High-Resolution Large Vision-Language Models

    Python 38 1

  5. Shenyi-Z/ToCa Shenyi-Z/ToCa Public

    [ICLR2025] Accelerating Diffusion Transformers with Token-wise Feature Caching

    Python 210 9

  6. MixKV MixKV Public

    [ICLR 2026] Mixing Importance with Diversity: Joint Optimization for KV Cache Compression in Large Vision-Language Models

    Python 21 3