How Spark is complementing Hadoop

How Spark is complementing Hadoop

Spark is in memory analytics tool designed for complementing Map reduce jobs for low latency and iterative jobs.

Spark is next generation distributed processing engine for Bigdata .

Spark never been competition to Hadoop jobs in fact it will increase scope of Bigdata project for different type of job which are difficult to do in Hadoop

Spark became good tool for online near realtime analytics with straming support.

Spark is very fast and follows the top down approach for streaming and analytics.

Spark is independent of file systesm and can integrate with any Hadoop file system or no SQL DB or Cloud platform.

Spark has very good SQL tool called Spark Sql which can read and write to almost any DB.

Spark supports Machine learning with Mlib and supports R with SparkR.

Spark integrates with all Hadoop ecosystem tools very well.

要查看或添加评论,请登录

janardhan reddy的更多文章

  • Disconnecting in the Connected world

    Disconnecting in the Connected world

    Welcome to the world of Social media and connected world . I am a Data and Social Engineer working in world since my…

  • Preparing pitch for Spark

    Preparing pitch for Spark

    What is needed to learn Apache Spark Apache Spark is open source technoly for inmemory data analytics and complex data…

  • Blockchain unleashed

    Blockchain unleashed

    A blockchain is decentralized transactional system for internet based generation ,It manages series of transactions as…

    3 条评论
  • Lets play with Hadoop Elephant : Part 0

    Lets play with Hadoop Elephant : Part 0

    What is Hadoop Hadoop is distributed processing and storage framework Hadoop is data processing tool to handle big and…

  • Anybody can become Hadooper

    Anybody can become Hadooper

    Bigdata /Hadoop is new terminology we are listening from last few years What It means for a fresh graduate engineer is…

    1 条评论

社区洞察

其他会员也浏览了