Machine Learning

Citadel: Protecting Data Privacy and Model Confidentiality for Collaborative Learning

To be updated.

Towards Efficient and Secure Large-Scale Systems for Distributed Machine Learning Training

Please refer to the enclosed pdf.

BatchCrypt: Efficient Homomorphic Encryption for Cross-Silo Federated Learning

Cross-silo federated learning (FL) enables organizations (e.g., financial or medical) to collaboratively train a machine learning model by aggregating local gradient updates from each client without sharing privacy-sensitive data. To ensure no update …

Enabling Cost-Effective, SLO-Aware Machine Learning Inference Serving on Public Cloud

The remarkable advances of Machine Learning (ML) have spurred an increasing demand for ML-as-a-Service on public cloud - developers train and publish ML models as online services to provide low-latency inference for dynamic queries. The primary …

Not All Explorations Are Equal: Harnessing Heterogeneous Profiling Cost for Efficient MLaaS Training

Machine-Learning-as-a-Service (MLaaS) enables practitioners and AI service providers to train and deploy ML models in the cloud using diverse and scalable compute resources. A common problem for MLaaS users is to choose from a variety of training …

MArk: Exploiting Cloud Services for Cost-Effective, SLO-Aware Machine Learning Inference Serving

The advances of Machine Learning (ML) have sparked a growing demand of ML-as-a-Service, developers train ML models and publish them in the cloud as online services to provide low-latency inference at scale. The key challenge of ML model serving is to …

Stay Fresh: Speculative Synchronization for Fast Distributed Machine Learning

Large machine learning models are typically trained in parallel and distributed environments. The model parameters are iteratively refined by multiple worker nodes in parallel, each processing a subset of the training data. In practice, the training …