Advertisement

SigOpt at O'Reilly - Best Practices for Scaling Modeling Platforms

SigOpt
May. 15, 2019
Advertisement

More Related Content

Slideshows for you(20)

Similar to SigOpt at O'Reilly - Best Practices for Scaling Modeling Platforms(20)

Advertisement
Advertisement

SigOpt at O'Reilly - Best Practices for Scaling Modeling Platforms

  1. Best Practices for Scaling Modeling Platforms Matt Greenwood, CIO, Two Sigma Scott Clark, CEO, SigOpt O’Reilly AI Conference, New York AI Business Summit, Case Studies
  2. Research & the Scientific MethodInvestment ManagementTechnology & Data
  3. *Current SigOpt trading customers represent over $300B in assets under management $300B+Assets Under Management*
  4. Hardware: Scalable, Efficient Compute Solutions: Evolving | Innovation: Existential | Implication: Build Data Management Pre-Processing Pipelines Feature Eng Solutions: Evolving __________ Innovation: Incremental __________ Implication: Mixed Experimentation, Training, Evaluation Modeling: Notebooks, Libraries, Frameworks Solutions: Evolving | Innovation: Existential | Implication: Build Experimentation: Tracking, Analysis, Optimization, Resource Mgmt Solutions: Standard | Innovation: Incremental | Implication: Buy Serving Monitoring Solutions: Evolving __________ Innovation: Existential __________ Implication: Build Model Productionalization
  5. Hardware Environment Transformation Labeling Pre-Processing Pipeline Dev. Feature Eng. Feature Stores Data Preparation Experimentation, Training, Evaluation Notebook, Library, Framework Experimentation & Model Optimization On-Premise Hybrid Multi-Cloud Insights, Tracking, Collaboration Model Search, Hyperparameter Tuning Resource Scheduler, Management Validation Serving Deploying Monitoring Managing Inference Online Testing Model Productionalization
  6. SigOpt. Confidential. Solution: Rapidly prototype models with SigOpt Experiment InsightsOptimization Engine Track, analyze and reproduce any model to improve the productivity of your modeling teams Enterprise Platform Automate parameter tuning to maximize the impact of your models with our optimization ensemble Standardize experimentation across any combination of library, infrastructure, model, or task On-Premise Hybrid/Multi 8
  7. SigOpt. Confidential.9
  8. SigOpt. Confidential. Hyperparameter Optimization Model Tuning Grid Search Random Search Bayesian Optimization Training & Tuning Evolutionary Algorithms Deep Learning Architecture Search Hyperparameter Search
  9. SigOpt. Confidential. Pro Con Manual Search Leverages expertise Not scalable, inconsistent Grid Search Simple to implement Not scalable, often infeasible Random Search Scalable Inefficient Evolutionary Algorithms Effective at architecture search Very resource intensive Bayesian Optimization Efficient, effective Can be tough to parallelize
  10. SigOpt. Confidential. Pro Con Manual Search Leverages expertise Not scalable, inconsistent Grid Search Simple to implement Not scalable, often infeasible Random Search Scalable Inefficient Evolutionary Algorithms Effective at architecture search Very resource intensive Bayesian Optimization Efficient, effective Can be tough to parallelize
  11. SigOpt. Confidential. Pro Con Manual Search Leverages expertise Not scalable, inconsistent Grid Search Simple to implement Not scalable, often infeasible Random Search Scalable Inefficient Evolutionary Algorithms Effective at architecture search Very resource intensive Bayesian Optimization Efficient, effective Can be tough to parallelize
  12. SigOpt. Confidential. Pro Con Manual Search Leverages expertise Not scalable, inconsistent Grid Search Simple to implement Not scalable, often infeasible Random Search Scalable Inefficient Evolutionary Algorithms Effective at architecture search Very resource intensive Bayesian Optimization Efficient, effective Can be tough to parallelize
  13. SigOpt. Confidential. Pro Con Manual Search Leverages expertise Not scalable, inconsistent Grid Search Simple to implement Not scalable, often infeasible Random Search Scalable Inefficient Evolutionary Algorithms Effective at architecture search Very resource intensive Bayesian Optimization Efficient, effective Can be tough to parallelize
  14. Data and models stay private Iterative, automated optimization Built specifically for scalable enterprise use cases Training Data AI/ML Model Model Evaluation Testing Data New Configurations Objective Metric Better Results EXPERIMENT INSIGHTS Organize and introspect experiments OPTIMIZATION ENSEMBLE Explore and exploit with a variety of techniques ENTERPRISE PLATFORM Built to scale with your models in production RESTAPI
  15. Build or Buy It there a proprietary advantage to DIY? Does it benefit from domain expertise? Is the process the same for each model? Is the open source well maintained? Is the open source reliable? Are there more performant alternatives? Is there a low maintenance burden to buy? Can the product scale with our needs? Can the product evolve with our needs?
  16. Benefits Of implementing SigOpt Realize Performance Gains Address entirely new problems Maximize Resource Utilization
  17. 90% fewer training runs to optimize https://devblogs.nvidia.com/sigopt-deep-learning-h yperparameter-optimization/ 400x faster time to optimize https://aws.amazon.com/blogs/machine-learning/fast-c nn-tuning-with-aws-gpu-instances-and-sigopt/ 20x the cost efficiency to optimize https://devblogs.nvidia.com/optimizing-end-to-end- memory-networks-using-sigopt-gpus/ Benefit: Performance Gains
  18. Better Results, 8x Faster “We’ve integrated SigOpt’s optimization service and are now able to get better results faster and cheaper than any solution we’ve seen before.” Matt Adereth Managing Director Two Sigma Benefit: Performance Gains
  19. Benefit: Address New Problems Failed observations Constraints on the model Noise in the data Competing metrics Lengthy training cycles Distributed training
  20. Benefit: Address New Problems Balancing Competing Metrics
  21. Agnostic Any Infrastructure Any Model Any Library Any Language Benefit: Maximize Resource Utilization Reliable Requires no data access 99.9% SLA 24/7 on-call support >1M API events per hour Complete Experiment reproducibility Modeling insights Differentiated algorithms Easy to scale in parallel
  22. Asynchronous parallelization Is critical for resource utilization Benefit: Maximize Resource Utilization
  23. SigOpt Impact Modeling Scale Exploration Identify valuable modeling use cases Scale Build repeatable process for scaling use cases Performance Fine-tune repeatable process to maximize gain
  24. Thank you https://www.twosigma.com/careers/ https://www.twosigma.com/insights/ https://www.twosigma.com/news/ https://opensource.twosigma.com/ https://sigopt.com/company/careers/ https://sigopt.com/blog/ https://sigopt.com/research/ https://sigopt.com/try-it
Advertisement