Skip to main content

Learning Spark Streaming #1

I have been doing a lot of Spark in the past few months, and of late, have taken a keen interest in Spark Streaming. In a series of posts, I intend to cover a lot of details about Spark streaming and even other stream processing systems in general, either presenting technical arguments/critiques, with any micro benchmarks as needed.

Some high level description of Spark Streaming (as of 1.4),  most of which you can find in the programming guide.  At a high level, Spark streaming is simply a spark job run on very small increments of input data (i.e micro batch), every 't' seconds, where t can be as low as 1 second.

As with any stream processing system, there are three big aspects to the framework itself.

  1. Ingesting the data streams : This is accomplished via DStreams, which you can think of effectively as a thin wrapper around an input source such as Kafka/HDFS which knows how to read the next N entries from the input.
    • The receiver based approach is a little complicated IMHO, and my recommendation is to avoid that route completely, since for the most part, it should be possible to refetch the data from the input source such as Kafka/HDFS

  2. Computing the result : Once new input is available, typically you execute the Spark DAG (which is basically your streaming job), computing your final result and optionally outputting that to a serving store such as a NoSQL datastore periodically. 
  3. Updating the internal state : Along the way during DAG execution, you would frequently need access to the results from previous computations, that you want to 'update'  based on new input. For e.g.: if you were tracking the page views in the last hour, then you need to increment some counters based on the last 't' second of data. Spark Streaming currently implements this via the StateDStream/StateRDD, which is basically like any other normal Spark RDD. 

Some good things about this model : (the paper is still a better source)

  1. Spark Streaming computes the entire computation once, for each micro batch. Thus, there are very clear semantics of what the result means, whereas in a per-event model (such as Storm/Samza) different input/output/intermediate streams can be in different points in time. 
  2. Once you read data off your input stream, you are no longer bound by the number of partitions in say your Kafka stream. You can repartition your data as you like, and let Spark & in-memory glory take over. 

So, what could be bad? (sticking to design level problems only)
  • Given the tight coupling between different computations in the DAG, if a single stage in the DAG were to be slow, you could end up with back pressure issues. (Good news inside bad news is, since next batches cannot be scheduled, it self throttles)
  • RDDs are built for immutability, reusing the same constructs for internal state maintenance, can be very inefficient
  • The common criticism is that its not pure streaming, but simply "micro batch" alluding to increased latency, when you compare to something like Storm. (My take on this is, if you want sub-second latency, then your service must be talking to a database i.e state store directly to begin with). But its a fair point and remains to be seen in what shape/form this truly impacts practical applications.

Let's take baby steps, towards understanding how Spark Streaming works under the hoods, by writing our very own CountingDStream, which does one simple thing, keeps counting. We will evolve this as we go to understand aspects like Checkpointing, State management, Recovery, Caching and so on.

Its actually very straightforward. First we implement the InputDStream trait/interface. The meat is in the compute() method, which is tasked to return the next set of inputs as an RDD every micro batch.

Now, the driver program that runs this and just prints out the counters. (Exercise: Increase the number of partitions to parallelize(..) from 1, to observe the effect on the counts printed out)

Next, lets add checkpoints to this DStream, so we will be able to kill and resume our counting..


  1. How does Spark Streaming know to fetch the entire record from the stream? Is it possible for micro batch to have partial record (the last record can be incomplete)?

    1. IEEE Final Year projects Project Center in Chennai are consistently sought after. Final Year Students Projects take a shot at them to improve their aptitudes. Final Year Project Domains for IT

      JavaScript Training in Chennai

      JavaScript Training in Chennai

      The Angular Training covers a wide range of topics including Components, project projects for cse. Angular Training

  2. Nice info regarding spark streaming My sincere thanks for sharing this post Please Continue to share this post
    Hadoop Training in Chennai

  3. nice blog has been shared by you. before i read this blog i didn't have any knowledge about this but now i got some knowledge. so keep on sharing such kind of an interesting blogs.
    hadoop training in chennai

  4. Excellent blog thanks for sharing Looking for the best place in Chennai to get your cosmetics at wholesale? The Pixies Beauty Shop is the premium wholesale cosmetics shop in Chennai that has all the international brands your salon deserves.
    beauty Shop in Chennai

  5. Thank you for sharing this blog, it is very useful for understanding the java programming.
    Even we also provide some tutorial related to this topic. For more information visit trishana technology.

  6. I just loved your article on the beginners guide to starting a blog.If somebody take this blog article seriously in their life, he/she can earn his living by doing blogging.thank you for this article.

    angular js training in chennai

    angular js training in velachery

    full stack training in chennai

    full stack training in velachery

    php training in chennai

    php training in velachery

    photoshop training in chennai

    photoshop training in velachery

  7. I am truly getting a charge out of perusing your elegantly composed articles. It would seem that you burn through a ton of energy and time on your blog. I have bookmarked it and I am anticipating perusing new articles. Keep doing awesome.
    data science coaching in hyderabad

  8. This post is so interactive and informative.keep update more information…
    Hadoop Training in Anna Nagar
    Big data training in chennai

  9. I simply stumbled upon your weblog and desired to say that I have really enjoyed surfing your blog articles.
    Positive site, where did u come up with the info on this uploading?
    yorkie puppies for sale
    yorkies for sale
    yorkie for sale
    yorkshire terrier for sale
    yorkie puppy for sale
    teacup yorkies for sale
    teacup yorkie for sale
    yorkie teacup for sale





  12. Sony Vegas Pro Crack is a powerful software for video and audio editing as well as for DVD and HD or Blu-ray disc burning. Cracks In Sony Firmware

  13. I have only one sorrow in my life, for I wish I had met you sooner.If only I knew how happy I would be with you; if only I knew then how wonderful you are, I would have roamed the streets till I found you. Wife Birthday Wishes

  14. PC Reviver License Key is the most recent several applications. It is used to generate a thorough image resolution of the whole pc program. Pc Reviver License Key Generator

  15. Overall, this piece was extremely fascinating. I like reading this because I was looking for information of this nature. Post more often. It is appropriate to express gratitude for sharing.
    Best CA Foundation Coaching in Hyderabad


Post a Comment

Popular posts from this blog

Setting up Hadoop/YARN/Spark/Hive on Mac OSX El Capitan

If you are like me, who loves to have everything you are developing against working locally in a mini-integration environment, read on Here, we attempt to get some pretty heavy-weight stuff working locally on your mac, namely Hadoop (Hadoop2/HDFS) YARN (So you can submit MR jobs) Spark (We will illustrate with Spark Shell, but should work on YARN mode as well) Hive (So we can create some tables and play with it)  We will use the latest stable Cloudera distribution, and work off the jars. Most of the methodology is borrowed from here , we just link the four pieces together nicely in this blog.  Download Stuff First off all, make sure you have Java 7/8 installed, with JAVA_HOME variable setup to point to the correct location. You have to download the CDH tarballs for Hadoop, Zookeeper, Hive from the tarball page (CDH 5.4.x page ) and untar them under a folder (refered to as CDH_HOME going forward) as hadoop, zookeeper $ ls $HOME /bin/cdh/5.4.7 hadoop

HDFS Client Configs for talking to HA Hadoop NameNodes

One more simple thing, that had relatively scarce documentation out on the Internet. As you might know, Hadoop NameNodes finally became HA in 2.0 . The HDFS client configuration, which is already a little bit tedious, became more complicated. Traditionally, there were two ways to configure a HDFS client (lets stick to Java) Copy over the entire Hadoop config directory with all the xml files, place it somewhere in the classpath of your app or construct a Hadoop Configuration object by manually adding in those files. Simply provide the HDFS NameNode URI and let the client do the rest.          Configuration conf = new Configuration(false);         conf.set("", "hdfs://localhost:8020"); // this is deprecated now         conf.set("fs.defaultFS", "hdfs://localhost:8020");         FileSystem fs = FileSystem.get(conf); Most people prefer 2, unless you need way more configs from the actual xml config files, at which po