Ongoing Research
Efficient LLM Serving Systems
Current Collaborators: ETRI, Microsoft Research Redmond, Samsung Research, KAIST, Samsung DS
- SLO-aware, adaptive KV cache management
- Generalized KV cache reuse
- Accuracy-preserving long-context pruning
- SLO-aware failure recovery for LLMs
- Systems for agentic AI
Continual and On-Device Learning
Current Collaborators: Palantir, SNU, UIUC
- Low-latency, high-accuracy on-device learning
- Adaptive, resource-efficient continual learning
Large-Scale Distributed Training
Current Collaborators: Samsung Research, KAIST, UC Merced, USC
- Learning-based planning for heterogeneous, geo-distributed training
- Fast distributed training on heterogeneous accelerators
Fast and Scalable Big Data Analytics
Current Collaborators: Amazon, Samsung Electronics, SNU
- Efficient caching for iterative analytics
- Data preprocessing for scalable ML pipelines
