Skip to content

patriknw/kafka-with-akka-streams-kafka-streams-tutorial

 
 

Repository files navigation

Building Kafka-based Microservices with Akka Streams and Kafka Streams - a Tutorial

Join the chat at https://gitter.im/kafka-with-akka-streams-kafka-streams-tutorial

NOTE: This tutorial is a work in progress. The "1.0" version will be available in time for the Software Architecture Conf tutorial, February 26.

Boris Lublinsky and Dean Wampler, Lightbend

This tutorial provides an introduction to streaming data microservices using Kafka with Akka Streams and Kafka Streams. Hence, the tutorial helps you compare and contrast these streaming libraries for your own use.

The core "use case" implemented is a stream processing application that also ingests updated parameters for a machine learning model and then uses the model to score the data. Several implementations of this use case are provided. They not only compare Akka Streams vs. Kafka Streams, but they also show how to support a few other common production requirements, such as managing the in-memory state of the application.

First, we will describe how to build and run the applications. Then we will discuss their designs. For reference materials and more information, see the end of this README.

Tutorial Setup

Note: If you are attending this tutorial at a conference, please follow the setup steps ahead of time. If you encounter problems, ask for help on the project's Gitter room.

Install the Required Tools

The Java JDK v8 is required. If not already installed, see the instructions here.

SBT, the de facto build tool for Scala is used to build the code, both the Scala and Java implementations. The SBT build files are configured to download all the required dependencies.

We recommend using IntelliJ IDEA for managing and building the code, which can drive SBT. The free Community Edition is sufficient. However, using IntelliJ isn't required; any favorite IDE or editor environment will do.

If you wish to use SBT in a terminal (e.g., in conjunction with your text editor), follow the SBT installation instructions here.

If you use IntelliJ IDEA or another IDE environment, also install available Scala and SBT plugins for the IDE.

Build the Code

Building the code before the tutorial session will ensure that everything works.

To compile the code with SBT in a terminal outside your IDE/Editor environment, use this command:

sbt compile

It's convenient to use SBT's interactive mode if you intend to run more than one command. Type sbt and you'll see a prompt sbt:akkaKafkaTutorial> (akkaKafkaTutorial is the name of the top-level SBT project). Now you can run tasks like tasks (list the most common tasks - add the -V flag to see all of them), clean, compile, help, etc.

The interactive mode is also convenient if you want to work in one of the nested projects, rather than the top-level project. At the SBT prompt, enter projects to list the projects (which we'll discuss below), then use project <name> to switch to that subproject. Now use commands like clean, compile, etc.

If you want to use IntelliJ to manage the project and run the SBT build:

  1. Use the Import Project feature to import the top-level directory of this repo as a project
  2. Select SBT as the project type
  3. In the settings dialog, check Use sbt shell for build and import
  4. After the project loads, open the SBT console, and click the green arrow to start the SBT interactive mode
  5. After it presents the prompt sbt:akkaKafkaTutorial>, type compile.

Alternatively instead of steps 4 and 5 you can use IntelliJ native compile and build capabilities.

Note If you use IntelliJ native compile and build capabilities, keep in mind that the project includes ScalaPB plugin, which is not always invoked by IntelliJ build. We recommend to do at least initial build using SBT to avoid erroneous compilation errors

To use the SBT build with other IDEs and editors, consult their documentation on using SBT and Scala. If your editor doesn't offer Scala and SBT support, just load the project as a Java project and use the editor to browse and edit the files. Use a terminal window to run SBT.

If you successfully built the code, you should be ready for the tutorial. We discuss running the services below.

About the Project

Let's explore the tutorial code in more detail.

Kafka

The examples rely on Kafka version 1.0 and leverage embedded Kafka servers. Using embedded Kafka servers, rather than standalone Kafka services, simplifies the setup and execution of the tutorial, but doesn't materially change how the examples work.

The implementations use two Kafka topics:

  • models_data - for sending data to the applications
  • models_models - for sending machine learning model updates, used for scoring the data

The model and data "provider" application, the client nested SBT project, will create an embedded Kafka server and the required topics. As a result, it has to be started before running any of the other services.

Project Overview

The tutorial is organized as several SBT nested projects in subdirectories. Other subdirectories provide supporting functions, like the data directory:

Each of the following directories is used to create a service executable. Each is also is defined as a nested SBT project:

  • client - Data and model loader used to run either the Akka Streams or Kafka Streams applications. This process has to be started first to ensure that the Kafka embedded server is running and the topics are created.
  • akkaStreamsCustomStage - Akka Streams implementation of model serving using a custom stage.
  • akkaActorsPersistent - Akka Actors and Akka Streams implementation of model serving that also provides state persistence to disk. This supports a production requirement to allow a failed service to be restarted from where it left off and it supports distributed scalability when using Akka cluster.
  • kafkaStreamsModelServerInMemoryStore - Kafka Streams implementation of model serving using the Processor Topology API and in-memory storage. You'll see the package name "naive", because there is no durability if the state is stored only in memory.
  • kafkaStreamsModelServerKVStore - Kafka Streams implementation of model serving using a key-value store provided by Kafka Streams and the Kafka Streams DSL.
  • kafkaStreamsModelServerCustomStore - Kafka Streams implementation of model serving using a custom store and the Kafka Streams DSL.

The following directories provide shared code. They are also nested SBT projects:

You can see the list of SBT projects using the sbt projects command. Note that the top level project, in the root directory, is called akkaKafkaTutorial and the SBT prompt is defined to be sbt:akkaKafkaTutorial> .

Finally, these directories provide other resources:

  • data - Some data files for running the applications.
  • images - Diagrams used for this document.

We provide Java and Scala implementations for all the examples, although we prefer Scala for its expressiveness and concision. Because the same names are often used in both the Scala and Java implementations, we use the package naming convention com.lightbend.scala for (almost) all Scala code and com.lightbend.java for (almost) all Java code. An exception is the code generated in the protobufs directory, where no scala or java package names are used. If you're new to Scala, this naming scheme is our convention; it's not commonly used in projects.

Application Architecture

A high level view of the overall model serving architecture is as follows:

Overall architecture of model serving

It is similar to this dynamically controlled stream, as described for Apache Flink.

This architecture assumes there are two data streams - one containing data that needs to be scored, and one containing the model updates. The streaming engine contains the current model used for the actual scoring in memory. The results of scoring can be either delivered downstream or used by the streaming engine internally as a new stream, i.e., as input for additional calculations.

If there is no model currently defined, the input data is dropped. When the new model is received, it is instantiated in memory. Once instantiation is complete, scoring is switched to the new model.

The model stream can either contain the binary blob of the data itself ("pass by value") or the reference to the model data stored externally ("pass by reference") in a database or a file system, like HDFS or S3. Only pass by value is implemented in the code.

This approach effectively implements model scoring as a new type of functional transformation, which can be used by any other stream transformations.

Although the overall architecture shown above is represented as a single model, a single streaming engine could score with multiple models simultaneously. This enhancement is left as an exercise for the student ;)

Queryable State

Kafka Streams recently introduced queryable state, which is a nice approach for execution monitoring and extracting the state of a stream. This feature allows you to treat the stream processing layer as a lightweight embedded database and, more concretely, to directly query the latest state of your stream processing application, without needing to materialize that state to external databases or external storage first.

Queriable state

Both the Akka Streams and Kafka Streams implementations support queryable state. Here, we use Akka HTTP for Web based access to both the Akka Streams and Kafka Streams queryable APIs.

Running the Application

We listed the SBT projects that have service executables above. Now let's explore these executables and how to run them.

First, if you using SBT from a terminal, to run a particular executable, use the following convention that specifies the SBT project, the client in this example, then select the executable to run from the list presented:

sbt:akkaKafkaTutorial> client/run

Multiple main classes detected, select one to run:

 [1] com.lightbend.scala.kafka.client.DataProvider
 [2] com.lightbend.scala.kafka.client.DataReader

Enter number: 1
...

If you know the fully qualified name of a specific executable, for example the DataProvider just shown, you can run it using runMain:

sbt:akkaKafkaTutorial> client/runMain com.lightbend.scala.kafka.client.DataProvider
...

Note: You will need one terminal for each service executed concurrently.

Alternatively you can run an executable directly from IntelliJ and other IDEs. Just right click on the source file containing the Scala or Java main, then in the pop-up menu, click run (or debug).

Now let's look at the specific executables, starting with the client, since it has to be executed first to set up the embedded Kafka instance. Then we'll list the rest of the projects alphabetically. We'll focus on the classes (Scala and Java types) that provide the executable main routines. We won't discuss the other classes here. We'll discuss them verbally during the tutorial session.

Client

The client service that starts the embedded Kafka server, creates the topics, and publishes data to them. Hence, the client corresponds to the Data Source and Model Source in the first architecture diagram above.

SBT Project client
Directory client
Data Provider com.lightbend.scala.kafka.client.DataProvider (Scala only) Provider service
Data Reader com.lightbend.scala.kafka.client.DataReader Verifies that messages are published correctly

In this case, we only provide Scala implementations, as this service isn't the focus of this tutorial.

The DataProvider is a data and model publisher. This executable contains an embedded Kafka server and has to be started first to ensure that the embedded Kafka server is running and the necessary topics are created.

Start the DataProvider using SBT as shown above:

sbt:akkaKafkaTutorial> client/run

Multiple main classes detected, select one to run:

 [1] com.lightbend.scala.kafka.client.DataProvider
 [2] com.lightbend.scala.kafka.client.DataReader

Enter number: 1
...

Or use this command to run DataProvider without the prompt:

sbt:akkaKafkaTutorial> client/runMain com.lightbend.scala.kafka.client.DataProvider
...

Alternatively you can run it from within IntelliJ by clicking the com.lightbend.scala.kafka.client.DataProvider source and invoking from the pop-up menu, Run DataProvider.

While DataProvider runs, you'll see output to the terminal like this:

...
printed 280 records
printed 290 records
Published Model generated from SparkML
printed 300 records
printed 310 records
printed 320 records
printed 330 records
...

Note the occasional messages about updates from SparkML!

You can run the DataReader, if desired, the same way. It reads the Kafka topics to which DataProvider is writing, to verify that data is being successfully written to the topics.

The DataReader outputs messages like the following:

...
Retrieved message #12, key = null, value = [B@53c536c6
Retrieved message #13, key = null, value = [B@71862844
Retrieved message #14, key = null, value = [B@4cfa1bf6
Retrieved message #15, key = null, value = [B@6fca5ad6
Retrieved message #16, key = null, value = [B@154322ef
Retrieved message #17, key = null, value = [B@55c7fc0b
...

Akka Streams Custom Stage Model Server

There are two versions of model serving based on Akka Streams (with both Scala and Java implementations):

  • An implementation based on the usage of a custom "stage", which is a fully type-safe way to encapsulate the required functionality (akkaStreamsCustomStage - this one)
  • An implementation based on Akka actors and the router pattern (akkaActorsPersistent - the next one)

The custom stage implementation uses the stream processor functionality from the overall architecture diagram shown above. The overall implementation looks as follows:

Akka streams custom stage model serving

SBT Project akkaStreamsCustomStage
Directory akkaStreamsCustomStage
Scala Version com.lightbend.scala.modelServer.modelServer.AkkaModelServer
Java Version com.lightbend.java.modelserver.modelserver.AkkaModelServer

(Note that modelserver is used for the Java version and modelServer for the Scala version...)

Start one of the applications by running sbt akkaStreamsCustomStage/run in a terminal window, then select the executable to run. Or use runMain, e.g., for the Scala version:

sbt:akkaKafkaTutorial> akkaStreamsCustomStage/runMain com.lightbend.scala.modelServer.modelServer.AkkaModelServer

For the Java version:

sbt:akkaKafkaTutorial> akkaStreamsCustomStage/runMain com.lightbend.java.modelserver.modelserver.AkkaModelServer

Alternatively you can run it from within IntelliJ by clicking at com.lightbend.scala.modelServer.modelServer.AkkaModelServer (Scala) or com.lightbend.java.modelserver.modelserver.AkkaModelServer (java) and pick Run AkkaModelServer

You'll see output like the following, where at first it skips processing, because it hasn't received a model yet, and then it receives a "model to serve":

...
No model available - skipping
New model - ModelToServe(winequalityDesisionTreeRegression,generated from SparkML,PMML,[B@471244ec,wine)
Calculated quality - 5.67910447761194 calculated in 76 ms
New model - ModelToServe(winequalityMultilayerPerceptron,generated from SparkML,PMML,[B@72ea5ccb,wine)
Calculated quality - 6.0 calculated in 48 ms
New model - ModelToServe(winequalityGeneralizedLinearRegressionGamma,generated from SparkML,PMML,[B@48d97260,wine)
Calculated quality - 6.1631297415506 calculated in 16 ms
New model - ModelToServe(winequalityRandonForrestClassification,generated from SparkML,PMML,[B@51ddd2e1,wine)
Calculated quality - 6.0 calculated in 34 ms
Calculated quality - 6.0 calculated in 1 ms
Calculated quality - 6.0 calculated in 1 ms
Calculated quality - 5.0 calculated in 1 ms
Calculated quality - 6.0 calculated in 2 ms
...

Once either server is running, go to http://localhost:5500/state to obtain the current state of execution, which in this case is the current model used for scoring.

Note: Only run one of the Akka model servers or the Kafka model servers at the same time, since they share the same ports to serve information.

Akka Streams Server with Akka Persistence

This is the implementation based on Akka actors and the router pattern (akkaActorsPersistent)

The stream processor functionality is implemented by an individual model-serving actor. A model-serving manager serves as a router, passing execution for a given data type to a specific actor.

Additionally this implementation uses a simple "local file" persistence, using Akka Persistence, which insures that the application can be restarted if it crashes. Hence, this feature helps support the resiliency required for production systems.

The overall implementation looks as follows:

Akka streams router model serving

SBT Project akkaActorsPersistent
Directory akkaActorsPersistent
Scala Version com.lightbend.scala.modelserver.actor.modelserver.AkkaModelServer
Java Version com.lightbend.java.modelserver.actor.modelserver.AkkaModelServer

Start one of the applications by running sbt akkaActorsPersistent/run in a terminal window, then select the executable to run. Or use runMain, e.g., for the Scala version:

sbt:akkaKafkaTutorial> akkaActorsPersistent/runMain com.lightbend.scala.modelServer.actor.modelServer.AkkaModelServer
...

Alternatively you can run it from within IntelliJ by clicking at com.lightbend.scala.modelserver.actor.modelserver.AkkaModelServer (Scala) or com.lightbend.java.modelserver.actor.modelServer.AkkaModelServer (java) and pick Run AkkaModelServer

Once running, visit these links:

Kafka Streams Model Server

Three versions of model serving are provided using Kafka Streams, with implementations in both Scala and Java:

  • "Naive" - an implementation that "persists" state in internal memory (which is "naive") and uses the Processor Topology API (kafkaStreamsModelServerInMemoryStore - two projects below)
  • Key-value store - an implementation based on the Kafka Streams built-in key-value Store and the Kafka Streams DSL (kafkaStreamsModelServerKVStore - the next project)
  • A custom store - an implementation based on a custom store (files) and the Kafka Streams DSL (kafkaStreamsModelServerCustomStore - this project)

See this blog post for a discussion of the differences between the Kafka Streams Processor Topology and the DSL.

For the Scala implementations based on Kafka Stream DSL, in addition to leveraging Kafka Java APIs we are also showing implementations based on the new Scala API for Kafka Streams that Lightbend recently open sourced. (It will be contributed to the Kafka project once the API details have settled.) It provides an idiomatic Scala version of the fluent Java Kafka Streams API. See also the Kafka Streams Query API, which provides a convenient way to query stream state across a cluster. (not used in this tutorial)

The model serving implementation is illustrated here:

Kafka streams model serving

Kafka Streams Model Server with an In-memory ("Naive") Store

This Kafka Streams implementation of model serving "persists" state in memory, which is "naive". Both Scala and Java implementations are provided for model serving that provide identical functionality.

SBT Project kafkaStreamsModelServerInMemoryStore
Directory kafkaStreamsModelServerInMemoryStore
Scala Version com.lightbend.scala.naive.modelserver.ModelServer
Java Version com.lightbend.java.naive.modelserver.ModelServer

Start one of the applications by running sbt kafkaStreamsModelServerInMemoryStore/run in a terminal window, then select the executable to run. Or use runMain, e.g., for the Scala version:

sbt:akkaKafkaTutorial> kafkaStreamsModelServerInMemoryStore/runMain com.lightbend.scala.naive.modelserver.ModelServer
...

Alternatively you can run it from within IntelliJ by clicking at com.lightbend.scala.naive.modelserver.ModelServer (Scala) or com.lightbend.java.naive.modelserver.ModelServer (java) and pick Run ModelServer

Once running, visit http://localhost:8888/state/value for the current state of execution for a given model.

Kafka Streams Model Server with a Key-Value Store

The second Kafka Streams example of model serving uses the provided key-value store in Kafka Streams. Two Scala and Java implementations are provided for model serving that provide identical functionality. An additional Scala implementation is provided that uses a "fluent" API.

SBT Project kafkaStreamsModelServerKVStore
Directory kafkaStreamsModelServerKVStore
Scala Version com.lightbend.scala.standard.modelserver.ModelServer
Java Version com.lightbend.java.standard.modelserver.ModelServer
Scala "Fluent" Version com.lightbend.scala.standard.modelserver.ModelServerFluent

Start one of the applications by running sbt kafkaStreamsModelServerKVStore/run in a terminal window, then select the executable to run. Or use runMain, e.g., for the Scala version:

sbt:akkaKafkaTutorial> kafkaStreamsModelServerKVStore/runMain com.lightbend.scala.standard.modelserver.ModelServer
...

Alternatively you can run it from within IntelliJ by clicking at com.lightbend.scala.standard.modelserver.scala.ModelServer (Scala, based on Java APIs) or com.lightbend.scala.standard.modelserver.scala.ModelServerFluent (Scala, based on Scala APIs) or com.lightbend.java.standard.modelserver.ModelServer (java) and pick Run ModelServer or Run ModelServerFluent

Once running, visit these links:

Kafka Streams Model Server with a Custom Store

The final Kafka Streams example of model serving demonstrates implementation of a custom store in Kafka Streams. Two Scala and Java implementations are provided for model serving that provide identical functionality. An additional Scala implementation is provided that uses a "fluent" API.

SBT Project kafkaStreamsModelServerCustomStore
Directory kafkaStreamsModelServerCustomStore
Scala Version com.lightbend.scala.custom.modelserver.ModelServer
Java Version com.lightbend.java.custom.modelserver.ModelServer
Scala "Fluent" Version com.lightbend.scala.custom.modelserver.ModelServerFluent

Start one of the applications by running sbt kafkaStreamsModelServerCustomStore/run in a terminal window, then select the executable to run. Or use runMain, e.g., for the Scala version:

sbt:akkaKafkaTutorial> kafkaStreamsModelServerCustomStore/runMain com.lightbend.scala.custom.modelserver.ModelServer
...

Alternatively you can run it from within IntelliJ by clicking at com.lightbend.scala.custom.modelserver.ModelServer (Scala, based on Java APIs) or com.lightbend.scala.custom.modelserver.ModelServerFluent (Scala, based on Scala APIs) or com.lightbend.java.standard.modelserver.ModelServer (java) and pick Run ModelServer or Run ModelServerFluent

Once running, visit these links:

Production Considerations

The examples illustrate how to implement a model-serving microservice using Akka Streams and Kafka Streams in various ways. A few illustrate other production considerations, such as persisting state so that restarts can pick up where processing left off.

Here we consider a few other production considerations.

Scaling

Since both Akka and Kafka Streams implementations are in JVM implementations, they can be deployed in a cluster environment in the usual ways, as long as the source of the streams is Kafka.

The following diagram shows a Kafka Streams cluster with several server instances, Kafka topic partitions, etc. An Akka Streams implementation can be scaled a similar way.

scaling

TODO - Expand

References

Kafka

Akka and Akka Streams

Kafka Streams

For More Information

Interested in an integrated and commercially supported distribution of Akka Streams, Kafka Streams, and Kafka, plus other tools like Spark and HDFS? See the https://www.lightbend.com/products/fast-data-platform for more information about the Lightbend Fast Data Platform.

About

Code samples for the Lightbend tutorial on writing microservices with Akka Streams, Kafka Streams, and Kafka

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Java 51.3%
  • Scala 48.7%