I like machine learning and systems. Above all, I like being useful.
I'm a masters student in ECE at UIUC, learning everything I can to make GPUs go brrr.
Previously, I worked for Coupang, training video recommendation models and building artwork generators.
[Full Resume]
Projects
- minimal-flash-attention is Flash Attention in ~100 lines of CUDA.
- cuda-1brc is my CUDA solution to the One Billion Row Challenge. Related to this is my blog post.
- mixed-precision-from-scratch shows all the details of mixed precision training applied to a simple 2-layer MLP. Read more here.
- paged-attention-minimal a simple cache manager for PagedAttention, on top of Llama 3.
Posts
Mixed Precision Training from Scratch
How to set up Nsight Compute Locally to profile Remote GPUs
The One Billion Row Challenge in CUDA: from 17m to 17s
Growing up in six different countries
A Quick Summary of "A Thorough Examination of the CNN/Daily Mail Reading Comprehension Task"
The Introspectiveness of Neural Networks
Patents
- Taeksang Kim. Personalized video DJ with text-to-video, Korean Pat. No. 10-0123761, 2023.
- Taeksang Kim. Video keyword tagging with large language models, Korean Pat. No. 10-0098584, 2023.
- Jina Hwang and Taeksang Kim. Sports awareness, mobile feed video player, Korean Pat. No. 10-0071581, 2023.
- Taeksang Kim. Next watch prediction using GPT, Korean Pat. No. 10-0036443, 2023.
- Taeksang Kim. First hero optimization through reinforcement learning, Korean Pat. No. 10-0184572, 2022.
Miscellaneous
- I designed, and was interviewed for, the first introductory ML course for employees at Buzzvil.
- I wrote a tech blog on how to optimize Item-to-item collaborative filtering with sparse vectors and Ray.
- I made an Instagram automation tool that went viral.