Haoyu He (何灏宇)

prof_pic.jpg

Tübingen AI Center

Tübingen, Germany

Hi! I am a third-year PhD student in the Autonomous Vision Group (AVG) supervised by Prof. Andreas Geiger at Univeristy of Tübingen and Tübingen AI Center.

My research interests are generally driven by my research vision - to build computationally efficient intelligent agents that liberate human labor. Currently, I am doing research on diffusion language models motivated by my reflections on current LLMs: (1) the redundant long reasoning process of current LLMs could be caused by the noise accumulation of autoregressive generation, (2) memory-bound Diffusion LMs have the potential to be much faster on inference than compute-bound autoregressive models given sufficient GPUs. Therefore, let’s make diffusion language models great!

Besides, I am also interested in shifting the paradigm under the linearization assumption where everything in the inputs is flattened, to models that can use ubiquitous hierarchies effectively. For example, concept models, hierarchical models…

My life is somehow sports-centric. I am a crazy enthusiast in cycling 🚴🚵. I play tennis 🎾 and ski ⛷️ as well.

My CV is here

news

Oct 31, 2024 Our paper “NN4SysBench: Characterizing Neural Network Verification for Computer Systems” is accepted to NeurIPS 2024!
Jul 10, 2024 Our paper “HDT: Hierarchical Document Transformer” is accepted to COLM 2024!

selected publications

  1. HDT: Hierarchical Document Transformer
    Haoyu He, Markus Flicke, Jan Buchmann, and 2 more authors
    In First Conference on Language Modeling, 2024
  2. NN4SysBench: Characterizing Neural Network Verification for Computer Systems
    Shuyi Lin, Haoyu He, Tianhao Wei, and 5 more authors
    In The Thirty-eight Conference on Neural Information Processing Systems Datasets and Benchmarks Track, 2024
  3. Distiller: A Systematic Study of Model Distillation Methods in Natural Language Processing
    Haoyu He, Xingjian Shi, Jonas Mueller, and 3 more authors
    In Proceedings of the Second Workshop on Simple and Efficient Natural Language Processing, Nov 2021