Spark is shaping up as the leading alternative to Map/Reduce for several reasons including the wide adoption by the different Hadoop distributions, combining both batch and streaming on a single platform and a growing library of machine-learning integration (both in terms of included algorithms and the integration with machine learning languages namely R and Python). At AppsFlyer, we’ve been using
![The bleeding edge: Spark, Parquet and S3](https://cdn-ak-scissors.b.st-hatena.com/image/square/dd3fad2a79b2a7634ceb024633e9548313057a3c/height=288;version=1;width=512/https%3A%2F%2Fwww.appsflyer.com%2Fwp-content%2Fuploads%2F2021%2F05%2Fspark-parquet-s3-og.jpg)