We can't find the internet
Attempting to reconnect
Something went wrong!
Hang in there while we get back on track
Matthew Rocklin - Dask in Production | SciPy 2024
Matthew Rocklin shares practical insights on optimizing Dask for production: from reducing cloud costs with ARM instances to avoiding common infrastructure pitfalls & deployment tips.
- 
    
Cloud computing costs can be significantly reduced by:
- Using ARM instances instead of Intel (5% faster, cheaper)
 - Leveraging spot instances when available
 - Turning off resources when not actively in use
 - Running workloads close to where data is stored
 
 - 
    
Running Dask in production revealed:
- The Global Interpreter Lock (GIL) is usually not a bottleneck (only ~25% contention)
 - Most workloads can process 1TB of data in ~5 hours for ~10 cents
 - Scaling is underutilized because people think it’s more expensive than it is
 - Raw cloud architecture (basic EC2 + networking) often works better than complex Kubernetes setups
 
 - 
    
Common cloud infrastructure challenges:
- Docker wasn’t designed for rapid development cycles
 - Serverless functions (Lambda) are 4x more expensive than regular instances
 - Users often leave large VMs running 24/7 unnecessarily
 - Moving data between regions/services is extremely costly
 
 - 
    
Success factors for cloud deployments:
- Making cloud environments match local development environments
 - Collecting detailed metrics on usage patterns
 - Supporting hardware flexibility across regions/instance types
 - Enabling rapid environment synchronization
 
 - 
    
The scientific Python ecosystem is increasingly ARM-compatible:
- 90-95% of workloads can run on ARM
 - Only specific cases (like MKL-dependent code) require Intel
 - Community should move towards ARM as the default