Amazon Web Services today announced the public beta of its newest addition to its web services portfolio; Amazon Elastic MapReduce.
Elastic MapReduce is a web service that allows researchers, analysts and businesses to process and analyze data using a hosted Apache Hadoop “data processing engine” that runs on Amazon S3 and Amazon EC2.
Much like the Mathematica and Mathworks EC2 computational solutions, Amazon Elastic MapReduce is a metered service that is designed to be used for applications like data mining, machine learning, financial analysis or scientific research. Amazon’s entry into the market will provide a bit of competition (to say the least) to Cloudera for users interested in running Hadoop in the Enterprise.
Elastic MapReduce will initially only be available in the USA and users can choose standard Amazon EC2 Instances or High CPU EC2 Instances to run their applications. Pricing for Elastic MapReduce ranges from 10 cents per hour for a small standard EC2 instance to 80 cents per hour for an extra large High CPU instance. Elastic MapReduce supports Ruby, Python, PHP, Java, and Perl.
Elastic MapReduce will be appealing to users that need to scale up quickly as the High CPU plans allow users to run up to hundreds of instances running Hadoop and instances can be run concurrently. Amazon EC2 and S3 have to date been very stable and the fact that EC2 and S3 are integrated with Elastic MapReduce and that instances can be managed via the AWS Console or command line are key selling points.