Robust Data Processing Pipeline with Elixir and Flow - Laszlo Bacsi - ElixirConf EU 2018

preview_player
Показать описание
It doesn't matter if you have "big data" or "small data" if you need to import and process it in near realtime you want to have a system that is robust and maintainable. This is where the failure tolerance and scalability of Erlang/OTP, the expressiveness of Elixir, and the flexibility of Flow and GenStage are all great assets. This is a story of how we built a data pipeline to move and process billions of rows from MySQL and CSV files into Redshift and what we learned along the way.
Рекомендации по теме
Комментарии
Автор

Good talk. Note that the introductory part is over at 18:40. You may want to just skip over it.

JXSMS
Автор

Thanks for the talk. I have a bit of a side-topic question: how are you streaming objects to and from S3 buckets? From what I can tell, ex_aws_s3 doesn't support streaming, but it is possible to stream files using HTTP calls if they are made available on a URL i.e. either publicly or using a pre-signed URL. Is there any other way I'm missing?

yawaramin
Автор

1:42 - Mic problem is taken care of and talk starts.

zesky