Sitemap

A list of all the posts and pages found on the site. For you robots out there is an XML version available for digesting as well.

Pages

Posts

Future Blog Post

less than 1 minute read

Published:

This post will show up by default. To disable scheduling of future posts, edit config.yml and set future: false.

Blog Post number 4

less than 1 minute read

Published:

This is a sample blog post. Lorem ipsum I can’t remember the rest of lorem ipsum and don’t have an internet connection right now. Testing testing testing this blog post. Blog posts are cool.

Blog Post number 3

less than 1 minute read

Published:

This is a sample blog post. Lorem ipsum I can’t remember the rest of lorem ipsum and don’t have an internet connection right now. Testing testing testing this blog post. Blog posts are cool.

Blog Post number 2

less than 1 minute read

Published:

This is a sample blog post. Lorem ipsum I can’t remember the rest of lorem ipsum and don’t have an internet connection right now. Testing testing testing this blog post. Blog posts are cool.

Blog Post number 1

less than 1 minute read

Published:

This is a sample blog post. Lorem ipsum I can’t remember the rest of lorem ipsum and don’t have an internet connection right now. Testing testing testing this blog post. Blog posts are cool.

portfolio

publications

Rethinking Learning Rate Tuning in the Era of Large Language Models

Published in 2023 IEEE 5th International Conference on Cognitive Machine Intelligence (CogMI), 2023

This paper explores the challenges of learning rate tuning for Large Language Models (LLMs) and introduces LRBench++ for benchmarking.

Recommended citation: Jin, H., Wei, W., Wang, X., Zhang, W., & Wu, Y. (2023). "Rethinking Learning Rate Tuning in the Era of Large Language Models." 2023 IEEE 5th International Conference on Cognitive Machine Intelligence (CogMI), 112-121.
Download Paper

DA-MoE: Towards Dynamic Expert Allocation for Mixture-of-Experts Models

Published in arXiv Preprint, 2024

This paper proposes DA-MoE, a novel dynamic router mechanism for Mixture-of-Experts (MoE) models, enabling efficient expert allocation based on token importance.

Recommended citation: Aghdam, M. A., Jin, H., & Wu, Y. (2024). "DA-MoE: Towards Dynamic Expert Allocation for Mixture-of-Experts Models." arXiv Preprint. arXiv:2409.06669.
Download Paper

talks

teaching

Teaching experience 1

Undergraduate course, University 1, Department, 2014

This is a description of a teaching experience. You can use markdown like any other post.

Teaching experience 2

Workshop, University 1, Department, 2015

This is a description of a teaching experience. You can use markdown like any other post.