Skip to content
View Tomorrowdawn's full-sized avatar

Block or report Tomorrowdawn

Block user

Prevent this user from interacting with your repositories and sending you notifications. Learn more about blocking users.

You must be logged in to block users.

Please don't include any personal information such as legal names or email addresses. Maximum 250 characters, markdown supported. This note will be visible to only you.
Report abuse

Contact GitHub support about this user’s behavior. Learn more about reporting abuse.

Report abuse
Tomorrowdawn/README.md

My GitHub stats

GitHub User's stars

Hi there 👋

Euripides

Torrowdawn, or Chenxia Tang here. You may find my resume (in chinese) here.

I was born in 2003, and I'm currently a Master student in USTC. My research focuses on post-training and inference optimization for large language models. I agree with Euripides: The language of truth is simple.

My favorite Top-2 English words are Meditation and Philosophy. I treat meditation as the best medicine for life, and I also deeply love wisdom.

💻 Skills

I am highly proficient in

  • Python Python
  • C++ C++

which you can verify from my repos. I also wrote verilog, and planning to learn dart to develop some android apps.


🔬 Research Focus

Primary Research Area:

  • Post-training & Inference Optimization for Large Language Models
  • Token Sampling Methods and Generation Quality
  • High-speed Inference and Performance Optimization

Previous Experience:

  • Reinforcement Learning, Recommendation Systems (brief exploration)

My research on large language models is particularly in-depth. I have extensive experience with RLHF (GPT-2 scale), fine-tuning (LLaMA2, 3), and have measured execution times of almost all operators in LLaMA. I am quite familiar with the bottlenecks of LLaMA, especially for single-GPU execution on the A6000.


🎓 Education

  • Bachelor's Degree

University of Science and Technology of China (USTC)
Special Class for the Gifted Young (enrolled at age 16)
2019 -2023

  • Master's Degree (Current)

University of Science and Technology of China (USTC)
School of Computer Science and Technology
2023 -

I served as a teaching assistant(TA)for Algebra,in 2023 fall.

📄 Papers

1. Top-nσ: Eliminating Noise in Logit Space for Robust Token Sampling of LLM

Authors: Chenxia Tang, Jianchun Liu, Hongli Xu, Liusheng Huang
Conference: ACL 2025 (Long Papers) | Pages: 10758-10774 | Venue: Vienna, Austria

TL;DR: Proposed top-nσ, a novel sampling method that eliminates noise directly in logit space. Key findings: (1) pre-softmax logits show clear separation between informative tokens and noise, (2) proved mathematical equivalence of min-p and top-(1-p). Achieves temperature-invariant token selection while preserving diversity, outperforming existing methods especially at high temperatures.

Links: ACL Anthology | PDF

2. Heterogeneous Learning Rate Scheduling for Neural Architecture Search on Long-Tailed Datasets

Authors: Chenxia Tang | Preprint: arXiv:2406.07028 | Submitted: June 2024

TL;DR: Proposed adaptive learning rate scheduling for DARTS on long-tailed datasets. Traditional re-sampling/re-weighting techniques cause performance degradation with DARTS. Our method specifically optimizes architecture parameters for imbalanced class distributions.

Links: arXiv | PDF

🚀 Projects

1. GITCGSimulator - Genius Invokation TCG Simulator

Repository: GitHub | Language: Python/C++

TL;DR: A comprehensive simulator for Genshin Impact's Genius Invokation TCG, designed specifically for AI training and reinforcement learning research.

Key Contributions:

  • Game Engine Development: Built a complete TCG simulation engine with event-driven architecture supporting complex card interactions, elemental reactions, and turn-based gameplay mechanics
  • AI Framework: Implemented MCTS (Monte Carlo Tree Search) algorithms and AlphaBeta search for intelligent gameplay, achieving competitive performance against human players
  • Modular Design: Created extensible card system with string-based reflection, supporting custom card implementations and easy game state serialization

2. VLMWalk - Vision-Language Model Research

Repository: GitHub | Language: Python

TL;DR: Self-study project focused on advancing vision-language model capabilities with modern frameworks.

3. seqattn - Sequence Attention Mechanisms

Repository: GitHub | Language: Python

TL;DR: A light-weight FlashInfer wrapper that simplifies memory management and enables easy implementation of custom sparse attention operations.

Key Contributions:

  • Memory Management: Developed a streamlined wrapper around FlashInfer that significantly simplifies memory allocation and deallocation for attention operations
  • Custom Sparse Attention: Created an intuitive interface for implementing custom sparse attention patterns and operations
  • Performance Optimization: Leveraged FlashInfer's optimized kernels while providing a more accessible API for research and development
  • Research Framework: Built a flexible foundation for experimenting with various attention mechanisms and sparse patterns

🌟 Open Source Contributions

My research on Top-nσ sampling has been adopted by several open-source projects:

Popular repositories Loading

  1. top_nsigma top_nsigma Public

    The official code repo and data hub of top_nsigma sampling strategy for LLMs.

    Python 26 2

  2. GITCGSimulator GITCGSimulator Public archive

    Simulator for GeniusInvokation TCG, tailor-made for ai training.

    Jupyter Notebook 15 1

  3. 2021USTC_COD 2021USTC_COD Public archive

    2021年中科大(USTC)计算机组成原理(COD)综合实验,带三级浮点流水的五级RISCV流水线

    VHDL 5

  4. GenshinTCGBuilder GenshinTCGBuilder Public

    Genshin, Genius Invokation TCG builder powered by GA. TCG卡组配卡器,使用遗传算法.

    Jupyter Notebook 4

  5. DND5EBOT DND5EBOT Public

    DND5E mainly bot, based on MiraiCP.

    C++ 3 2

  6. CatsTailTavern CatsTailTavern Public

    Genius Invokation TCG Simulator, with engine in C++ and gym-style python interface

    C++ 3