This post gives a walkthrough of how to use airflow to schedule spark jobs triggered by downloading reddit data from s3. apache airflow is an incubating project developed by airbnb used for scheduling tasks and dependencies between tasks. scheduling a tas
Total 2 post Blog.insightdatascience.com. There may be jobs you're interested in.
Total 2 post Blog.insightdatascience.com . There may be jobs you're interested in..
Getting started with spark & batch processing frameworks
https://blog.insightdatascience.com/getting-starte...