Haoran Qiu | Microsoft AzRS
Home
Publications
Experiences
Awards
Contact
Esha Choukse
Latest
Medha: Efficiently Serving Multi-Million Context Length LLM Inference Requests Without Approximations
ModServe: Scalable and Resource-Efficient Large Multimodal Model Serving
Towards Efficient Large Multimodal Model Serving
TAPAS: Thermal- and Power-Aware Scheduling for LLM Inference in Cloud Platforms
SmartOClock: Workload- and Risk-Aware Overclocking in the Cloud
Cite
×