filmov
tv
How can we use AWS lambda and Tensorflow to serve deep learning models // Alexey Grigorev

Показать описание
Deep learning achieves the best performance for many computer vision, natural language processing, and recommendation tasks and thus it’s becoming increasingly more popular.
However, it’s quite difficult to use deep learning in production as it requires a lot of effort to develop proper infrastructure for serving deep learning models.
Platforms for serverless computing, such as AWS Lambda, provide a good alternative: they take care of scaling up and down and offer attractive pricing based only on actual usage. These platforms, unfortunately, have other limitations that make it problematic.
In this talk, we show how to come around these limitations and be able to use AWS lambda and TensorFlow to serve deep learning models. We’ll also cover the limitations of AWS lambda, compare it with “serverful” solutions, and suggest workloads for which serverless is not the best option.
However, it’s quite difficult to use deep learning in production as it requires a lot of effort to develop proper infrastructure for serving deep learning models.
Platforms for serverless computing, such as AWS Lambda, provide a good alternative: they take care of scaling up and down and offer attractive pricing based only on actual usage. These platforms, unfortunately, have other limitations that make it problematic.
In this talk, we show how to come around these limitations and be able to use AWS lambda and TensorFlow to serve deep learning models. We’ll also cover the limitations of AWS lambda, compare it with “serverful” solutions, and suggest workloads for which serverless is not the best option.