1 of 21

Scalable Hyper-parameter Optimization using RAPIDS and AWS

AKSHIT ARORA

@_AkshitArora

SRISHTI YADAV

@_srishtiyadav

2 of 21

SPEAKER: AKSHIT ARORA

  • Deep learning solutions architect at NVIDIA
  • Optimize and deploy deep learning and machine learning pipelines in production
  • Previously, built deep learning models for tasks in domains such as โ€“ Education, Virtual Reality, Natural Language Processing and Weather Prediction

๐Ÿ”— aroraakshit.github.io

\Uhk_Shith Aurora\

2

3 of 21

SPEAKER: SRISHTI YADAV

  • Deep learning researcher at Simon Fraser University, Canada
  • Research, Create and Optimize deep learning and machine learning algorithm
  • Also, building deep-learning algorithm for sensors at International Space Station (Intern: Urthecast)

๐Ÿ”— srishti.dev

3

3

4 of 21

WHY THIS TALK?

4

5 of 21

WE ALL LOVE TO SPEED UP THINGS

BUT ARE WE ๐Ÿƒโ€โ™€๏ธ๐Ÿƒโ€โ™‚๏ธ THINGS FASTER ?

5

6 of 21

COMPANIES LOVE TO STORE AND USE DATA

BUT ARE THEY REALLY CASHING IT ?

6

7 of 21

WE KNOW CLOUDS ARE A THING

BUT WE DONโ€™T KNOW IF ITโ€™S WITHIN REACH ๐Ÿคฃ

7

8 of 21

HOW ARE WE HELPING?

8

9 of 21

โšก๐Ÿƒโ€โ™€๏ธ๐Ÿƒโ€โ™‚๏ธ : Papermill

๐Ÿ’ฐ๐Ÿ’ฐ๐Ÿ’ฐ : GPU

โ˜๏ธ๐Ÿ–ฅ๏ธโ˜๏ธ ๐Ÿ–ฅ๏ธ : AWS

9

10 of 21

IT CAN BE AN OCEAN OF DETAILS

BUT BEFORE WE IN

10

11 of 21

Hyper-Parameter Optimization (HPO)

For 5 parameters, each with 4 desired values, we will have 45 possible combinations

  • Combinations can increase exponentially. And your training time with it. ๐Ÿ˜ฏ
  • Manually trying each of them ๐Ÿคฎ
  • Solution: Automating Hyper-parameter optimization (HPO) ๐ŸŽ‰

11

12 of 21

WHAT ARE WE GOING TO DO?

12

13 of 21

CPU Based Scalability

Learn to parameterize notebooks from a set of parameters

You donโ€™t want to do all of them manually! ๐Ÿคฆ

Automate parallel computation of parameterized notebooks.

Since we use CPU here, you can play with it at home too, without any cost.

13

14 of 21

Papermill lets you:

  • Parameterize notebooks
  • Execute notebooks

Helpful for running HPO tasks where the same model needs to be trained again and again with different sets of parameters.

14

15 of 21

DEMO TIME

15

16 of 21

GPU Based Scalability

Learn to speed up the computation using GPU

Make efficient use of GPU using RAPIDS

Scale up the computation and have better metric visualization using AWS

Learn to do hyper parameter optimization to find best parameters from given set

You donโ€™t want to iterate over all of them manually, trust us!

16

17 of 21

  • Amazon SageMaker is a fully managed machine learning service
  • Allows the users to build, train, analyze and deploy models in production ready environment

17

18 of 21

  • A collection of open-source libraries for end-to-end GPU Accelerated Data Science.
  • Involves little to no code change

18

19 of 21

  • A collection of open-source libraries for end-to-end GPU Accelerated Data Science.
  • Involves little to no code change

19

20 of 21

DEMO TIME

20

21 of 21

Thank you!

Please find our code, slides and other resources at our github page:

https://github.com/copperwiring/scalable-hpo-pybay

21