Skip to content

Latest commit

 

History

History
37 lines (27 loc) · 2.23 KB

File metadata and controls

37 lines (27 loc) · 2.23 KB

Architecture-for-real-time-video-streaming-analytics

This is an architecture to process video in real time using Kafka as a messaging service and Spark to process frames in real time, the architecture can be summarized in the image below. img

Start Zookeeper and kafka

To simplify the tedious task of building a kafka server, we will use Docker with Kafka and Zookeeper out of the box: Go to the kafka-docker folder

$ cd kafka-docker 

Start Kafka and Zookeeper with Docker Compose

$ docker-compose -f docker-compose_local.yml up 

In the docker-compose_local.yml configuration file, the topics are defined, the url and the port through which our publisher and subscribers connected.

To stop Kafka uses

$ docker-compose stop 

for more information Apache Kafka: Docker Container refer to this post

Send frames in real time with Kafka:

We will create a subscriber that takes the webcam frames, transforms them into bytes and sends them to a topic named "distributed-video1"

python my_pub.py 

Get frames in real time with Kafka:

We create a subsriber that pulls the messages we store in the topic "distributed-video1". Spark Streaming is responsible for converting them into a DStream type that can be processed by Spark. As an example we will take the DStream (RDDs stacked over time), we will extract the frames, we will detect contained faces and we will identify the emotions of the face (check my_sub.py).

python my_sub.py 

If you have come this far you should be able to see something like this: alt text

finally we send the result to another kafka topic "final_result_topic" and we consume the result from another subscriber.

python my_sub_end.py 

References