+doc Add more imports to the stream quick start guides for Java and Scala (#20797)
* Add more imports to the stream quick start guides. This makes it easier for people to execute the code samples while they read through the guide. * Change line endings to be consistent with other files For some reason these 2 files had CR+LF line endings.
This commit is contained in:
parent
a9cb8ab4b5
commit
f246c56087
4 changed files with 692 additions and 677 deletions
|
|
@ -3,23 +3,27 @@
|
|||
*/
|
||||
package docs.stream;
|
||||
|
||||
//#stream-imports
|
||||
import akka.stream.*;
|
||||
import akka.stream.javadsl.*;
|
||||
//#stream-imports
|
||||
|
||||
//#other-imports
|
||||
import akka.Done;
|
||||
import akka.NotUsed;
|
||||
import akka.actor.ActorSystem;
|
||||
import akka.util.ByteString;
|
||||
|
||||
import java.nio.file.Paths;
|
||||
import java.math.BigInteger;
|
||||
import java.util.concurrent.CompletionStage;
|
||||
import java.util.concurrent.ExecutionException;
|
||||
import java.util.concurrent.TimeUnit;
|
||||
|
||||
import org.junit.*;
|
||||
|
||||
import akka.Done;
|
||||
import akka.NotUsed;
|
||||
import akka.actor.ActorSystem;
|
||||
//#imports
|
||||
import akka.stream.*;
|
||||
import akka.stream.javadsl.*;
|
||||
//#imports
|
||||
import akka.util.ByteString;
|
||||
import scala.concurrent.duration.Duration;
|
||||
//#other-imports
|
||||
|
||||
import org.junit.*;
|
||||
|
||||
/**
|
||||
* This class is not meant to be run as a test in the test suite, but it
|
||||
|
|
|
|||
|
|
@ -1,333 +1,337 @@
|
|||
.. _stream-quickstart-java:
|
||||
|
||||
Quick Start Guide
|
||||
=================
|
||||
|
||||
A stream usually begins at a source, so this is also how we start an Akka
|
||||
Stream. Before we create one, we import the full complement of streaming tools:
|
||||
|
||||
.. includecode:: ../code/docs/stream/QuickStartDocTest.java#imports
|
||||
|
||||
Now we will start with a rather simple source, emitting the integers 1 to 100:
|
||||
|
||||
.. includecode:: ../code/docs/stream/QuickStartDocTest.java#create-source
|
||||
|
||||
The :class:`Source` type is parameterized with two types: the first one is the
|
||||
type of element that this source emits and the second one may signal that
|
||||
running the source produces some auxiliary value (e.g. a network source may
|
||||
provide information about the bound port or the peer’s address). Where no
|
||||
auxiliary information is produced, the type ``akka.NotUsed`` is used—and a
|
||||
simple range of integers surely falls into this category.
|
||||
|
||||
Having created this source means that we have a description of how to emit the
|
||||
first 100 natural numbers, but this source is not yet active. In order to get
|
||||
those numbers out we have to run it:
|
||||
|
||||
.. includecode:: ../code/docs/stream/QuickStartDocTest.java#run-source
|
||||
|
||||
This line will complement the source with a consumer function—in this example
|
||||
we simply print out the numbers to the console—and pass this little stream
|
||||
setup to an Actor that runs it. This activation is signaled by having “run” be
|
||||
part of the method name; there are other methods that run Akka Streams, and
|
||||
they all follow this pattern.
|
||||
|
||||
You may wonder where the Actor gets created that runs the stream, and you are
|
||||
probably also asking yourself what this ``materializer`` means. In order to get
|
||||
this value we first need to create an Actor system:
|
||||
|
||||
.. includecode:: ../code/docs/stream/QuickStartDocTest.java#create-materializer
|
||||
|
||||
There are other ways to create a materializer, e.g. from an
|
||||
:class:`ActorContext` when using streams from within Actors. The
|
||||
:class:`Materializer` is a factory for stream execution engines, it is the
|
||||
thing that makes streams run—you don’t need to worry about any of the details
|
||||
just now apart from that you need one for calling any of the ``run`` methods on
|
||||
a :class:`Source`.
|
||||
|
||||
The nice thing about Akka Streams is that the :class:`Source` is just a
|
||||
description of what you want to run, and like an architect’s blueprint it can
|
||||
be reused, incorporated into a larger design. We may choose to transform the
|
||||
source of integers and write it to a file instead:
|
||||
|
||||
.. includecode:: ../code/docs/stream/QuickStartDocTest.java#transform-source
|
||||
|
||||
First we use the ``scan`` combinator to run a computation over the whole
|
||||
stream: starting with the number 1 (``BigInteger.ONE``) we multiple by each of
|
||||
the incoming numbers, one after the other; the scan operationemits the initial
|
||||
value and then every calculation result. This yields the series of factorial
|
||||
numbers which we stash away as a :class:`Source` for later reuse—it is
|
||||
important to keep in mind that nothing is actually computed yet, this is just a
|
||||
description of what we want to have computed once we run the stream. Then we
|
||||
convert the resulting series of numbers into a stream of :class:`ByteString`
|
||||
objects describing lines in a text file. This stream is then run by attaching a
|
||||
file as the receiver of the data. In the terminology of Akka Streams this is
|
||||
called a :class:`Sink`. :class:`IOResult` is a type that IO operations return
|
||||
in Akka Streams in order to tell you how many bytes or elements were consumed
|
||||
and whether the stream terminated normally or exceptionally.
|
||||
|
||||
Reusable Pieces
|
||||
---------------
|
||||
|
||||
One of the nice parts of Akka Streams—and something that other stream libraries
|
||||
do not offer—is that not only sources can be reused like blueprints, all other
|
||||
elements can be as well. We can take the file-writing :class:`Sink`, prepend
|
||||
the processing steps necessary to get the :class:`ByteString` elements from
|
||||
incoming strings and package that up as a reusable piece as well. Since the
|
||||
language for writing these streams always flows from left to right (just like
|
||||
plain English), we need a starting point that is like a source but with an
|
||||
“open” input. In Akka Streams this is called a :class:`Flow`:
|
||||
|
||||
.. includecode:: ../code/docs/stream/QuickStartDocTest.java#transform-sink
|
||||
|
||||
Starting from a flow of strings we convert each to :class:`ByteString` and then
|
||||
feed to the already known file-writing :class:`Sink`. The resulting blueprint
|
||||
is a :class:`Sink<String, CompletionStage<IOResult>>`, which means that it
|
||||
accepts strings as its input and when materialized it will create auxiliary
|
||||
information of type ``CompletionStage<IOResult>`` (when chaining operations on
|
||||
a :class:`Source` or :class:`Flow` the type of the auxiliary information—called
|
||||
the “materialized value”—is given by the leftmost starting point; since we want
|
||||
to retain what the ``FileIO.toFile`` sink has to offer, we need to say
|
||||
``Keep.right()``).
|
||||
|
||||
We can use the new and shiny :class:`Sink` we just created by
|
||||
attaching it to our ``factorials`` source—after a small adaptation to turn the
|
||||
numbers into strings:
|
||||
|
||||
.. includecode:: ../code/docs/stream/QuickStartDocTest.java#use-transformed-sink
|
||||
|
||||
Time-Based Processing
|
||||
---------------------
|
||||
|
||||
Before we start looking at a more involved example we explore the streaming
|
||||
nature of what Akka Streams can do. Starting from the ``factorials`` source
|
||||
we transform the stream by zipping it together with another stream,
|
||||
represented by a :class:`Source` that emits the number 0 to 100: the first
|
||||
number emitted by the ``factorials`` source is the factorial of zero, the
|
||||
second is the factorial of one, and so on. We combine these two by forming
|
||||
strings like ``"3! = 6"``.
|
||||
|
||||
.. includecode:: ../code/docs/stream/QuickStartDocTest.java#add-streams
|
||||
|
||||
All operations so far have been time-independent and could have been performed
|
||||
in the same fashion on strict collections of elements. The next line
|
||||
demonstrates that we are in fact dealing with streams that can flow at a
|
||||
certain speed: we use the ``throttle`` combinator to slow down the stream to 1
|
||||
element per second (the second ``1`` in the argument list is the maximum size
|
||||
of a burst that we want to allow—passing ``1`` means that the first element
|
||||
gets through immediately and the second then has to wait for one second and so
|
||||
on).
|
||||
|
||||
If you run this program you will see one line printed per second. One aspect
|
||||
that is not immediately visible deserves mention, though: if you try and set
|
||||
the streams to produce a billion numbers each then you will notice that your
|
||||
JVM does not crash with an OutOfMemoryError, even though you will also notice
|
||||
that running the streams happens in the background, asynchronously (this is the
|
||||
reason for the auxiliary information to be provided as a
|
||||
:class:`CompletionStage`, in the future). The secret that makes this work is
|
||||
that Akka Streams implicitly implement pervasive flow control, all combinators
|
||||
respect back-pressure. This allows the throttle combinator to signal to all its
|
||||
upstream sources of data that it can only accept elements at a certain
|
||||
rate—when the incoming rate is higher than one per second the throttle
|
||||
combinator will assert *back-pressure* upstream.
|
||||
|
||||
This is basically all there is to Akka Streams in a nutshell—glossing over the
|
||||
fact that there are dozens of sources and sinks and many more stream
|
||||
transformation combinators to choose from, see also :ref:`stages-overview_java`.
|
||||
|
||||
Reactive Tweets
|
||||
===============
|
||||
|
||||
A typical use case for stream processing is consuming a live stream of data that we want to extract or aggregate some
|
||||
other data from. In this example we'll consider consuming a stream of tweets and extracting information concerning Akka from them.
|
||||
|
||||
We will also consider the problem inherent to all non-blocking streaming
|
||||
solutions: *"What if the subscriber is too slow to consume the live stream of
|
||||
data?"*. Traditionally the solution is often to buffer the elements, but this
|
||||
can—and usually will—cause eventual buffer overflows and instability of such
|
||||
systems. Instead Akka Streams depend on internal backpressure signals that
|
||||
allow to control what should happen in such scenarios.
|
||||
|
||||
Here's the data model we'll be working with throughout the quickstart examples:
|
||||
|
||||
.. includecode:: ../code/docs/stream/TwitterStreamQuickstartDocTest.java#model
|
||||
|
||||
|
||||
.. note::
|
||||
If you would like to get an overview of the used vocabulary first instead of diving head-first
|
||||
into an actual example you can have a look at the :ref:`core-concepts-java` and :ref:`defining-and-running-streams-java`
|
||||
sections of the docs, and then come back to this quickstart to see it all pieced together into a simple example application.
|
||||
|
||||
Transforming and consuming simple streams
|
||||
-----------------------------------------
|
||||
The example application we will be looking at is a simple Twitter feed stream from which we'll want to extract certain information,
|
||||
like for example finding all twitter handles of users who tweet about ``#akka``.
|
||||
|
||||
In order to prepare our environment by creating an :class:`ActorSystem` and :class:`ActorMaterializer`,
|
||||
which will be responsible for materializing and running the streams we are about to create:
|
||||
|
||||
.. includecode:: ../code/docs/stream/TwitterStreamQuickstartDocTest.java#materializer-setup
|
||||
|
||||
The :class:`ActorMaterializer` can optionally take :class:`ActorMaterializerSettings` which can be used to define
|
||||
materialization properties, such as default buffer sizes (see also :ref:`async-stream-buffers-java`), the dispatcher to
|
||||
be used by the pipeline etc. These can be overridden with ``withAttributes`` on :class:`Flow`, :class:`Source`, :class:`Sink` and :class:`Graph`.
|
||||
|
||||
Let's assume we have a stream of tweets readily available. In Akka this is expressed as a :class:`Source<Out, M>`:
|
||||
|
||||
.. includecode:: ../code/docs/stream/TwitterStreamQuickstartDocTest.java#tweet-source
|
||||
|
||||
Streams always start flowing from a ``Source<Out,M1>`` then can continue through ``Flow<In,Out,M2>`` elements or
|
||||
more advanced graph elements to finally be consumed by a ``Sink<In,M3>``.
|
||||
|
||||
The first type parameter—:class:`Tweet` in this case—designates the kind of elements produced
|
||||
by the source while the ``M`` type parameters describe the object that is created during
|
||||
materialization (:ref:`see below <materialized-values-quick-java>`)—:class:`BoxedUnit` (from the ``scala.runtime``
|
||||
package) means that no value is produced, it is the generic equivalent of ``void``.
|
||||
|
||||
The operations should look familiar to anyone who has used the Scala Collections library,
|
||||
however they operate on streams and not collections of data (which is a very important distinction, as some operations
|
||||
only make sense in streaming and vice versa):
|
||||
|
||||
.. includecode:: ../code/docs/stream/TwitterStreamQuickstartDocTest.java#authors-filter-map
|
||||
|
||||
Finally in order to :ref:`materialize <stream-materialization-java>` and run the stream computation we need to attach
|
||||
the Flow to a ``Sink<T, M>`` that will get the Flow running. The simplest way to do this is to call
|
||||
``runWith(sink)`` on a ``Source<Out, M>``. For convenience a number of common Sinks are predefined and collected as static methods on
|
||||
the `Sink class <http://doc.akka.io/japi/akka-stream-and-http-experimental/@version@/akka/stream/javadsl/Sink.html>`_.
|
||||
For now let's simply print each author:
|
||||
|
||||
.. includecode:: ../code/docs/stream/TwitterStreamQuickstartDocTest.java#authors-foreachsink-println
|
||||
|
||||
or by using the shorthand version (which are defined only for the most popular Sinks such as :class:`Sink.fold` and :class:`Sink.foreach`):
|
||||
|
||||
.. includecode:: ../code/docs/stream/TwitterStreamQuickstartDocTest.java#authors-foreach-println
|
||||
|
||||
Materializing and running a stream always requires a :class:`Materializer` to be passed in explicitly,
|
||||
like this: ``.run(mat)``.
|
||||
|
||||
The complete snippet looks like this:
|
||||
|
||||
.. includecode:: ../code/docs/stream/TwitterStreamQuickstartDocTest.java#first-sample
|
||||
|
||||
Flattening sequences in streams
|
||||
-------------------------------
|
||||
In the previous section we were working on 1:1 relationships of elements which is the most common case, but sometimes
|
||||
we might want to map from one element to a number of elements and receive a "flattened" stream, similarly like ``flatMap``
|
||||
works on Scala Collections. In order to get a flattened stream of hashtags from our stream of tweets we can use the ``mapConcat``
|
||||
combinator:
|
||||
|
||||
.. includecode:: ../code/docs/stream/TwitterStreamQuickstartDocTest.java#hashtags-mapConcat
|
||||
|
||||
.. note::
|
||||
The name ``flatMap`` was consciously avoided due to its proximity with for-comprehensions and monadic composition.
|
||||
It is problematic for two reasons: firstly, flattening by concatenation is often undesirable in bounded stream processing
|
||||
due to the risk of deadlock (with merge being the preferred strategy), and secondly, the monad laws would not hold for
|
||||
our implementation of flatMap (due to the liveness issues).
|
||||
|
||||
Please note that the ``mapConcat`` requires the supplied function to return a strict collection (``Out f -> java.util.List<T>``),
|
||||
whereas ``flatMap`` would have to operate on streams all the way through.
|
||||
|
||||
|
||||
Broadcasting a stream
|
||||
---------------------
|
||||
Now let's say we want to persist all hashtags, as well as all author names from this one live stream.
|
||||
For example we'd like to write all author handles into one file, and all hashtags into another file on disk.
|
||||
This means we have to split the source stream into two streams which will handle the writing to these different files.
|
||||
|
||||
Elements that can be used to form such "fan-out" (or "fan-in") structures are referred to as "junctions" in Akka Streams.
|
||||
One of these that we'll be using in this example is called :class:`Broadcast`, and it simply emits elements from its
|
||||
input port to all of its output ports.
|
||||
|
||||
Akka Streams intentionally separate the linear stream structures (Flows) from the non-linear, branching ones (Graphs)
|
||||
in order to offer the most convenient API for both of these cases. Graphs can express arbitrarily complex stream setups
|
||||
at the expense of not reading as familiarly as collection transformations.
|
||||
|
||||
Graphs are constructed using :class:`GraphDSL` like this:
|
||||
|
||||
.. includecode:: ../code/docs/stream/TwitterStreamQuickstartDocTest.java#graph-dsl-broadcast
|
||||
|
||||
As you can see, we use graph builder ``b`` to construct the graph using ``UniformFanOutShape`` and ``Flow`` s.
|
||||
|
||||
``GraphDSL.create`` returns a :class:`Graph`, in this example a ``Graph<ClosedShape,Unit>`` where
|
||||
:class:`ClosedShape` means that it is *a fully connected graph* or "closed" - there are no unconnected inputs or outputs.
|
||||
Since it is closed it is possible to transform the graph into a :class:`RunnableGraph` using ``RunnableGraph.fromGraph``.
|
||||
The runnable graph can then be ``run()`` to materialize a stream out of it.
|
||||
|
||||
Both :class:`Graph` and :class:`RunnableGraph` are *immutable, thread-safe, and freely shareable*.
|
||||
|
||||
A graph can also have one of several other shapes, with one or more unconnected ports. Having unconnected ports
|
||||
expresses a graph that is a *partial graph*. Concepts around composing and nesting graphs in large structures are
|
||||
explained in detail in :ref:`composition-java`. It is also possible to wrap complex computation graphs
|
||||
as Flows, Sinks or Sources, which will be explained in detail in :ref:`partial-graph-dsl-java`.
|
||||
|
||||
|
||||
Back-pressure in action
|
||||
-----------------------
|
||||
|
||||
One of the main advantages of Akka Streams is that they *always* propagate back-pressure information from stream Sinks
|
||||
(Subscribers) to their Sources (Publishers). It is not an optional feature, and is enabled at all times. To learn more
|
||||
about the back-pressure protocol used by Akka Streams and all other Reactive Streams compatible implementations read
|
||||
:ref:`back-pressure-explained-java`.
|
||||
|
||||
A typical problem applications (not using Akka Streams) like this often face is that they are unable to process the incoming data fast enough,
|
||||
either temporarily or by design, and will start buffering incoming data until there's no more space to buffer, resulting
|
||||
in either ``OutOfMemoryError`` s or other severe degradations of service responsiveness. With Akka Streams buffering can
|
||||
and must be handled explicitly. For example, if we are only interested in the "*most recent tweets, with a buffer of 10
|
||||
elements*" this can be expressed using the ``buffer`` element:
|
||||
|
||||
.. includecode:: ../code/docs/stream/TwitterStreamQuickstartDocTest.java#tweets-slow-consumption-dropHead
|
||||
|
||||
The ``buffer`` element takes an explicit and required ``OverflowStrategy``, which defines how the buffer should react
|
||||
when it receives another element while it is full. Strategies provided include dropping the oldest element (``dropHead``),
|
||||
dropping the entire buffer, signalling failures etc. Be sure to pick and choose the strategy that fits your use case best.
|
||||
|
||||
.. _materialized-values-quick-java:
|
||||
|
||||
Materialized values
|
||||
-------------------
|
||||
So far we've been only processing data using Flows and consuming it into some kind of external Sink - be it by printing
|
||||
values or storing them in some external system. However sometimes we may be interested in some value that can be
|
||||
obtained from the materialized processing pipeline. For example, we want to know how many tweets we have processed.
|
||||
While this question is not as obvious to give an answer to in case of an infinite stream of tweets (one way to answer
|
||||
this question in a streaming setting would be to create a stream of counts described as "*up until now*, we've processed N tweets"),
|
||||
but in general it is possible to deal with finite streams and come up with a nice result such as a total count of elements.
|
||||
|
||||
First, let's write such an element counter using ``Flow.of(Class)`` and ``Sink.fold`` to see how the types look like:
|
||||
|
||||
.. includecode:: ../code/docs/stream/TwitterStreamQuickstartDocTest.java#tweets-fold-count
|
||||
|
||||
First we prepare a reusable ``Flow`` that will change each incoming tweet into an integer of value ``1``. We'll use this in
|
||||
order to combine those with a ``Sink.fold`` that will sum all ``Integer`` elements of the stream and make its result available as
|
||||
a ``CompletionStage<Integer>``. Next we connect the ``tweets`` stream to ``count`` with ``via``. Finally we connect the Flow to the previously
|
||||
prepared Sink using ``toMat``.
|
||||
|
||||
Remember those mysterious ``Mat`` type parameters on ``Source<Out, Mat>``, ``Flow<In, Out, Mat>`` and ``Sink<In, Mat>``?
|
||||
They represent the type of values these processing parts return when materialized. When you chain these together,
|
||||
you can explicitly combine their materialized values: in our example we used the ``Keep.right`` predefined function,
|
||||
which tells the implementation to only care about the materialized type of the stage currently appended to the right.
|
||||
The materialized type of ``sumSink`` is ``CompletionStage<Integer>`` and because of using ``Keep.right``, the resulting :class:`RunnableGraph`
|
||||
has also a type parameter of ``CompletionStage<Integer>``.
|
||||
|
||||
This step does *not* yet materialize the
|
||||
processing pipeline, it merely prepares the description of the Flow, which is now connected to a Sink, and therefore can
|
||||
be ``run()``, as indicated by its type: ``RunnableGraph<CompletionStage<Integer>>``. Next we call ``run()`` which uses the :class:`ActorMaterializer`
|
||||
to materialize and run the Flow. The value returned by calling ``run()`` on a ``RunnableGraph<T>`` is of type ``T``.
|
||||
In our case this type is ``CompletionStage<Integer>`` which, when completed, will contain the total length of our tweets stream.
|
||||
In case of the stream failing, this future would complete with a Failure.
|
||||
|
||||
A :class:`RunnableGraph` may be reused
|
||||
and materialized multiple times, because it is just the "blueprint" of the stream. This means that if we materialize a stream,
|
||||
for example one that consumes a live stream of tweets within a minute, the materialized values for those two materializations
|
||||
will be different, as illustrated by this example:
|
||||
|
||||
.. includecode:: ../code/docs/stream/TwitterStreamQuickstartDocTest.java#tweets-runnable-flow-materialized-twice
|
||||
|
||||
Many elements in Akka Streams provide materialized values which can be used for obtaining either results of computation or
|
||||
steering these elements which will be discussed in detail in :ref:`stream-materialization-java`. Summing up this section, now we know
|
||||
what happens behind the scenes when we run this one-liner, which is equivalent to the multi line version above:
|
||||
|
||||
.. includecode:: ../code/docs/stream/TwitterStreamQuickstartDocTest.java#tweets-fold-count-oneline
|
||||
|
||||
.. note::
|
||||
``runWith()`` is a convenience method that automatically ignores the materialized value of any other stages except
|
||||
those appended by the ``runWith()`` itself. In the above example it translates to using ``Keep.right`` as the combiner
|
||||
for materialized values.
|
||||
.. _stream-quickstart-java:
|
||||
|
||||
Quick Start Guide
|
||||
=================
|
||||
|
||||
A stream usually begins at a source, so this is also how we start an Akka
|
||||
Stream. Before we create one, we import the full complement of streaming tools:
|
||||
|
||||
.. includecode:: ../code/docs/stream/QuickStartDocTest.java#stream-imports
|
||||
|
||||
If you want to execute the code samples while you read through the quick start guide, you will also need the following imports:
|
||||
|
||||
.. includecode:: ../code/docs/stream/QuickStartDocTest.java#other-imports
|
||||
|
||||
Now we will start with a rather simple source, emitting the integers 1 to 100:
|
||||
|
||||
.. includecode:: ../code/docs/stream/QuickStartDocTest.java#create-source
|
||||
|
||||
The :class:`Source` type is parameterized with two types: the first one is the
|
||||
type of element that this source emits and the second one may signal that
|
||||
running the source produces some auxiliary value (e.g. a network source may
|
||||
provide information about the bound port or the peer’s address). Where no
|
||||
auxiliary information is produced, the type ``akka.NotUsed`` is used—and a
|
||||
simple range of integers surely falls into this category.
|
||||
|
||||
Having created this source means that we have a description of how to emit the
|
||||
first 100 natural numbers, but this source is not yet active. In order to get
|
||||
those numbers out we have to run it:
|
||||
|
||||
.. includecode:: ../code/docs/stream/QuickStartDocTest.java#run-source
|
||||
|
||||
This line will complement the source with a consumer function—in this example
|
||||
we simply print out the numbers to the console—and pass this little stream
|
||||
setup to an Actor that runs it. This activation is signaled by having “run” be
|
||||
part of the method name; there are other methods that run Akka Streams, and
|
||||
they all follow this pattern.
|
||||
|
||||
You may wonder where the Actor gets created that runs the stream, and you are
|
||||
probably also asking yourself what this ``materializer`` means. In order to get
|
||||
this value we first need to create an Actor system:
|
||||
|
||||
.. includecode:: ../code/docs/stream/QuickStartDocTest.java#create-materializer
|
||||
|
||||
There are other ways to create a materializer, e.g. from an
|
||||
:class:`ActorContext` when using streams from within Actors. The
|
||||
:class:`Materializer` is a factory for stream execution engines, it is the
|
||||
thing that makes streams run—you don’t need to worry about any of the details
|
||||
just now apart from that you need one for calling any of the ``run`` methods on
|
||||
a :class:`Source`.
|
||||
|
||||
The nice thing about Akka Streams is that the :class:`Source` is just a
|
||||
description of what you want to run, and like an architect’s blueprint it can
|
||||
be reused, incorporated into a larger design. We may choose to transform the
|
||||
source of integers and write it to a file instead:
|
||||
|
||||
.. includecode:: ../code/docs/stream/QuickStartDocTest.java#transform-source
|
||||
|
||||
First we use the ``scan`` combinator to run a computation over the whole
|
||||
stream: starting with the number 1 (``BigInteger.ONE``) we multiple by each of
|
||||
the incoming numbers, one after the other; the scan operationemits the initial
|
||||
value and then every calculation result. This yields the series of factorial
|
||||
numbers which we stash away as a :class:`Source` for later reuse—it is
|
||||
important to keep in mind that nothing is actually computed yet, this is just a
|
||||
description of what we want to have computed once we run the stream. Then we
|
||||
convert the resulting series of numbers into a stream of :class:`ByteString`
|
||||
objects describing lines in a text file. This stream is then run by attaching a
|
||||
file as the receiver of the data. In the terminology of Akka Streams this is
|
||||
called a :class:`Sink`. :class:`IOResult` is a type that IO operations return
|
||||
in Akka Streams in order to tell you how many bytes or elements were consumed
|
||||
and whether the stream terminated normally or exceptionally.
|
||||
|
||||
Reusable Pieces
|
||||
---------------
|
||||
|
||||
One of the nice parts of Akka Streams—and something that other stream libraries
|
||||
do not offer—is that not only sources can be reused like blueprints, all other
|
||||
elements can be as well. We can take the file-writing :class:`Sink`, prepend
|
||||
the processing steps necessary to get the :class:`ByteString` elements from
|
||||
incoming strings and package that up as a reusable piece as well. Since the
|
||||
language for writing these streams always flows from left to right (just like
|
||||
plain English), we need a starting point that is like a source but with an
|
||||
“open” input. In Akka Streams this is called a :class:`Flow`:
|
||||
|
||||
.. includecode:: ../code/docs/stream/QuickStartDocTest.java#transform-sink
|
||||
|
||||
Starting from a flow of strings we convert each to :class:`ByteString` and then
|
||||
feed to the already known file-writing :class:`Sink`. The resulting blueprint
|
||||
is a :class:`Sink<String, CompletionStage<IOResult>>`, which means that it
|
||||
accepts strings as its input and when materialized it will create auxiliary
|
||||
information of type ``CompletionStage<IOResult>`` (when chaining operations on
|
||||
a :class:`Source` or :class:`Flow` the type of the auxiliary information—called
|
||||
the “materialized value”—is given by the leftmost starting point; since we want
|
||||
to retain what the ``FileIO.toFile`` sink has to offer, we need to say
|
||||
``Keep.right()``).
|
||||
|
||||
We can use the new and shiny :class:`Sink` we just created by
|
||||
attaching it to our ``factorials`` source—after a small adaptation to turn the
|
||||
numbers into strings:
|
||||
|
||||
.. includecode:: ../code/docs/stream/QuickStartDocTest.java#use-transformed-sink
|
||||
|
||||
Time-Based Processing
|
||||
---------------------
|
||||
|
||||
Before we start looking at a more involved example we explore the streaming
|
||||
nature of what Akka Streams can do. Starting from the ``factorials`` source
|
||||
we transform the stream by zipping it together with another stream,
|
||||
represented by a :class:`Source` that emits the number 0 to 100: the first
|
||||
number emitted by the ``factorials`` source is the factorial of zero, the
|
||||
second is the factorial of one, and so on. We combine these two by forming
|
||||
strings like ``"3! = 6"``.
|
||||
|
||||
.. includecode:: ../code/docs/stream/QuickStartDocTest.java#add-streams
|
||||
|
||||
All operations so far have been time-independent and could have been performed
|
||||
in the same fashion on strict collections of elements. The next line
|
||||
demonstrates that we are in fact dealing with streams that can flow at a
|
||||
certain speed: we use the ``throttle`` combinator to slow down the stream to 1
|
||||
element per second (the second ``1`` in the argument list is the maximum size
|
||||
of a burst that we want to allow—passing ``1`` means that the first element
|
||||
gets through immediately and the second then has to wait for one second and so
|
||||
on).
|
||||
|
||||
If you run this program you will see one line printed per second. One aspect
|
||||
that is not immediately visible deserves mention, though: if you try and set
|
||||
the streams to produce a billion numbers each then you will notice that your
|
||||
JVM does not crash with an OutOfMemoryError, even though you will also notice
|
||||
that running the streams happens in the background, asynchronously (this is the
|
||||
reason for the auxiliary information to be provided as a
|
||||
:class:`CompletionStage`, in the future). The secret that makes this work is
|
||||
that Akka Streams implicitly implement pervasive flow control, all combinators
|
||||
respect back-pressure. This allows the throttle combinator to signal to all its
|
||||
upstream sources of data that it can only accept elements at a certain
|
||||
rate—when the incoming rate is higher than one per second the throttle
|
||||
combinator will assert *back-pressure* upstream.
|
||||
|
||||
This is basically all there is to Akka Streams in a nutshell—glossing over the
|
||||
fact that there are dozens of sources and sinks and many more stream
|
||||
transformation combinators to choose from, see also :ref:`stages-overview_java`.
|
||||
|
||||
Reactive Tweets
|
||||
===============
|
||||
|
||||
A typical use case for stream processing is consuming a live stream of data that we want to extract or aggregate some
|
||||
other data from. In this example we'll consider consuming a stream of tweets and extracting information concerning Akka from them.
|
||||
|
||||
We will also consider the problem inherent to all non-blocking streaming
|
||||
solutions: *"What if the subscriber is too slow to consume the live stream of
|
||||
data?"*. Traditionally the solution is often to buffer the elements, but this
|
||||
can—and usually will—cause eventual buffer overflows and instability of such
|
||||
systems. Instead Akka Streams depend on internal backpressure signals that
|
||||
allow to control what should happen in such scenarios.
|
||||
|
||||
Here's the data model we'll be working with throughout the quickstart examples:
|
||||
|
||||
.. includecode:: ../code/docs/stream/TwitterStreamQuickstartDocTest.java#model
|
||||
|
||||
|
||||
.. note::
|
||||
If you would like to get an overview of the used vocabulary first instead of diving head-first
|
||||
into an actual example you can have a look at the :ref:`core-concepts-java` and :ref:`defining-and-running-streams-java`
|
||||
sections of the docs, and then come back to this quickstart to see it all pieced together into a simple example application.
|
||||
|
||||
Transforming and consuming simple streams
|
||||
-----------------------------------------
|
||||
The example application we will be looking at is a simple Twitter feed stream from which we'll want to extract certain information,
|
||||
like for example finding all twitter handles of users who tweet about ``#akka``.
|
||||
|
||||
In order to prepare our environment by creating an :class:`ActorSystem` and :class:`ActorMaterializer`,
|
||||
which will be responsible for materializing and running the streams we are about to create:
|
||||
|
||||
.. includecode:: ../code/docs/stream/TwitterStreamQuickstartDocTest.java#materializer-setup
|
||||
|
||||
The :class:`ActorMaterializer` can optionally take :class:`ActorMaterializerSettings` which can be used to define
|
||||
materialization properties, such as default buffer sizes (see also :ref:`async-stream-buffers-java`), the dispatcher to
|
||||
be used by the pipeline etc. These can be overridden with ``withAttributes`` on :class:`Flow`, :class:`Source`, :class:`Sink` and :class:`Graph`.
|
||||
|
||||
Let's assume we have a stream of tweets readily available. In Akka this is expressed as a :class:`Source<Out, M>`:
|
||||
|
||||
.. includecode:: ../code/docs/stream/TwitterStreamQuickstartDocTest.java#tweet-source
|
||||
|
||||
Streams always start flowing from a ``Source<Out,M1>`` then can continue through ``Flow<In,Out,M2>`` elements or
|
||||
more advanced graph elements to finally be consumed by a ``Sink<In,M3>``.
|
||||
|
||||
The first type parameter—:class:`Tweet` in this case—designates the kind of elements produced
|
||||
by the source while the ``M`` type parameters describe the object that is created during
|
||||
materialization (:ref:`see below <materialized-values-quick-java>`)—:class:`BoxedUnit` (from the ``scala.runtime``
|
||||
package) means that no value is produced, it is the generic equivalent of ``void``.
|
||||
|
||||
The operations should look familiar to anyone who has used the Scala Collections library,
|
||||
however they operate on streams and not collections of data (which is a very important distinction, as some operations
|
||||
only make sense in streaming and vice versa):
|
||||
|
||||
.. includecode:: ../code/docs/stream/TwitterStreamQuickstartDocTest.java#authors-filter-map
|
||||
|
||||
Finally in order to :ref:`materialize <stream-materialization-java>` and run the stream computation we need to attach
|
||||
the Flow to a ``Sink<T, M>`` that will get the Flow running. The simplest way to do this is to call
|
||||
``runWith(sink)`` on a ``Source<Out, M>``. For convenience a number of common Sinks are predefined and collected as static methods on
|
||||
the `Sink class <http://doc.akka.io/japi/akka-stream-and-http-experimental/@version@/akka/stream/javadsl/Sink.html>`_.
|
||||
For now let's simply print each author:
|
||||
|
||||
.. includecode:: ../code/docs/stream/TwitterStreamQuickstartDocTest.java#authors-foreachsink-println
|
||||
|
||||
or by using the shorthand version (which are defined only for the most popular Sinks such as :class:`Sink.fold` and :class:`Sink.foreach`):
|
||||
|
||||
.. includecode:: ../code/docs/stream/TwitterStreamQuickstartDocTest.java#authors-foreach-println
|
||||
|
||||
Materializing and running a stream always requires a :class:`Materializer` to be passed in explicitly,
|
||||
like this: ``.run(mat)``.
|
||||
|
||||
The complete snippet looks like this:
|
||||
|
||||
.. includecode:: ../code/docs/stream/TwitterStreamQuickstartDocTest.java#first-sample
|
||||
|
||||
Flattening sequences in streams
|
||||
-------------------------------
|
||||
In the previous section we were working on 1:1 relationships of elements which is the most common case, but sometimes
|
||||
we might want to map from one element to a number of elements and receive a "flattened" stream, similarly like ``flatMap``
|
||||
works on Scala Collections. In order to get a flattened stream of hashtags from our stream of tweets we can use the ``mapConcat``
|
||||
combinator:
|
||||
|
||||
.. includecode:: ../code/docs/stream/TwitterStreamQuickstartDocTest.java#hashtags-mapConcat
|
||||
|
||||
.. note::
|
||||
The name ``flatMap`` was consciously avoided due to its proximity with for-comprehensions and monadic composition.
|
||||
It is problematic for two reasons: firstly, flattening by concatenation is often undesirable in bounded stream processing
|
||||
due to the risk of deadlock (with merge being the preferred strategy), and secondly, the monad laws would not hold for
|
||||
our implementation of flatMap (due to the liveness issues).
|
||||
|
||||
Please note that the ``mapConcat`` requires the supplied function to return a strict collection (``Out f -> java.util.List<T>``),
|
||||
whereas ``flatMap`` would have to operate on streams all the way through.
|
||||
|
||||
|
||||
Broadcasting a stream
|
||||
---------------------
|
||||
Now let's say we want to persist all hashtags, as well as all author names from this one live stream.
|
||||
For example we'd like to write all author handles into one file, and all hashtags into another file on disk.
|
||||
This means we have to split the source stream into two streams which will handle the writing to these different files.
|
||||
|
||||
Elements that can be used to form such "fan-out" (or "fan-in") structures are referred to as "junctions" in Akka Streams.
|
||||
One of these that we'll be using in this example is called :class:`Broadcast`, and it simply emits elements from its
|
||||
input port to all of its output ports.
|
||||
|
||||
Akka Streams intentionally separate the linear stream structures (Flows) from the non-linear, branching ones (Graphs)
|
||||
in order to offer the most convenient API for both of these cases. Graphs can express arbitrarily complex stream setups
|
||||
at the expense of not reading as familiarly as collection transformations.
|
||||
|
||||
Graphs are constructed using :class:`GraphDSL` like this:
|
||||
|
||||
.. includecode:: ../code/docs/stream/TwitterStreamQuickstartDocTest.java#graph-dsl-broadcast
|
||||
|
||||
As you can see, we use graph builder ``b`` to construct the graph using ``UniformFanOutShape`` and ``Flow`` s.
|
||||
|
||||
``GraphDSL.create`` returns a :class:`Graph`, in this example a ``Graph<ClosedShape,Unit>`` where
|
||||
:class:`ClosedShape` means that it is *a fully connected graph* or "closed" - there are no unconnected inputs or outputs.
|
||||
Since it is closed it is possible to transform the graph into a :class:`RunnableGraph` using ``RunnableGraph.fromGraph``.
|
||||
The runnable graph can then be ``run()`` to materialize a stream out of it.
|
||||
|
||||
Both :class:`Graph` and :class:`RunnableGraph` are *immutable, thread-safe, and freely shareable*.
|
||||
|
||||
A graph can also have one of several other shapes, with one or more unconnected ports. Having unconnected ports
|
||||
expresses a graph that is a *partial graph*. Concepts around composing and nesting graphs in large structures are
|
||||
explained in detail in :ref:`composition-java`. It is also possible to wrap complex computation graphs
|
||||
as Flows, Sinks or Sources, which will be explained in detail in :ref:`partial-graph-dsl-java`.
|
||||
|
||||
|
||||
Back-pressure in action
|
||||
-----------------------
|
||||
|
||||
One of the main advantages of Akka Streams is that they *always* propagate back-pressure information from stream Sinks
|
||||
(Subscribers) to their Sources (Publishers). It is not an optional feature, and is enabled at all times. To learn more
|
||||
about the back-pressure protocol used by Akka Streams and all other Reactive Streams compatible implementations read
|
||||
:ref:`back-pressure-explained-java`.
|
||||
|
||||
A typical problem applications (not using Akka Streams) like this often face is that they are unable to process the incoming data fast enough,
|
||||
either temporarily or by design, and will start buffering incoming data until there's no more space to buffer, resulting
|
||||
in either ``OutOfMemoryError`` s or other severe degradations of service responsiveness. With Akka Streams buffering can
|
||||
and must be handled explicitly. For example, if we are only interested in the "*most recent tweets, with a buffer of 10
|
||||
elements*" this can be expressed using the ``buffer`` element:
|
||||
|
||||
.. includecode:: ../code/docs/stream/TwitterStreamQuickstartDocTest.java#tweets-slow-consumption-dropHead
|
||||
|
||||
The ``buffer`` element takes an explicit and required ``OverflowStrategy``, which defines how the buffer should react
|
||||
when it receives another element while it is full. Strategies provided include dropping the oldest element (``dropHead``),
|
||||
dropping the entire buffer, signalling failures etc. Be sure to pick and choose the strategy that fits your use case best.
|
||||
|
||||
.. _materialized-values-quick-java:
|
||||
|
||||
Materialized values
|
||||
-------------------
|
||||
So far we've been only processing data using Flows and consuming it into some kind of external Sink - be it by printing
|
||||
values or storing them in some external system. However sometimes we may be interested in some value that can be
|
||||
obtained from the materialized processing pipeline. For example, we want to know how many tweets we have processed.
|
||||
While this question is not as obvious to give an answer to in case of an infinite stream of tweets (one way to answer
|
||||
this question in a streaming setting would be to create a stream of counts described as "*up until now*, we've processed N tweets"),
|
||||
but in general it is possible to deal with finite streams and come up with a nice result such as a total count of elements.
|
||||
|
||||
First, let's write such an element counter using ``Flow.of(Class)`` and ``Sink.fold`` to see how the types look like:
|
||||
|
||||
.. includecode:: ../code/docs/stream/TwitterStreamQuickstartDocTest.java#tweets-fold-count
|
||||
|
||||
First we prepare a reusable ``Flow`` that will change each incoming tweet into an integer of value ``1``. We'll use this in
|
||||
order to combine those with a ``Sink.fold`` that will sum all ``Integer`` elements of the stream and make its result available as
|
||||
a ``CompletionStage<Integer>``. Next we connect the ``tweets`` stream to ``count`` with ``via``. Finally we connect the Flow to the previously
|
||||
prepared Sink using ``toMat``.
|
||||
|
||||
Remember those mysterious ``Mat`` type parameters on ``Source<Out, Mat>``, ``Flow<In, Out, Mat>`` and ``Sink<In, Mat>``?
|
||||
They represent the type of values these processing parts return when materialized. When you chain these together,
|
||||
you can explicitly combine their materialized values: in our example we used the ``Keep.right`` predefined function,
|
||||
which tells the implementation to only care about the materialized type of the stage currently appended to the right.
|
||||
The materialized type of ``sumSink`` is ``CompletionStage<Integer>`` and because of using ``Keep.right``, the resulting :class:`RunnableGraph`
|
||||
has also a type parameter of ``CompletionStage<Integer>``.
|
||||
|
||||
This step does *not* yet materialize the
|
||||
processing pipeline, it merely prepares the description of the Flow, which is now connected to a Sink, and therefore can
|
||||
be ``run()``, as indicated by its type: ``RunnableGraph<CompletionStage<Integer>>``. Next we call ``run()`` which uses the :class:`ActorMaterializer`
|
||||
to materialize and run the Flow. The value returned by calling ``run()`` on a ``RunnableGraph<T>`` is of type ``T``.
|
||||
In our case this type is ``CompletionStage<Integer>`` which, when completed, will contain the total length of our tweets stream.
|
||||
In case of the stream failing, this future would complete with a Failure.
|
||||
|
||||
A :class:`RunnableGraph` may be reused
|
||||
and materialized multiple times, because it is just the "blueprint" of the stream. This means that if we materialize a stream,
|
||||
for example one that consumes a live stream of tweets within a minute, the materialized values for those two materializations
|
||||
will be different, as illustrated by this example:
|
||||
|
||||
.. includecode:: ../code/docs/stream/TwitterStreamQuickstartDocTest.java#tweets-runnable-flow-materialized-twice
|
||||
|
||||
Many elements in Akka Streams provide materialized values which can be used for obtaining either results of computation or
|
||||
steering these elements which will be discussed in detail in :ref:`stream-materialization-java`. Summing up this section, now we know
|
||||
what happens behind the scenes when we run this one-liner, which is equivalent to the multi line version above:
|
||||
|
||||
.. includecode:: ../code/docs/stream/TwitterStreamQuickstartDocTest.java#tweets-fold-count-oneline
|
||||
|
||||
.. note::
|
||||
``runWith()`` is a convenience method that automatically ignores the materialized value of any other stages except
|
||||
those appended by the ``runWith()`` itself. In the above example it translates to using ``Keep.right`` as the combiner
|
||||
for materialized values.
|
||||
|
|
|
|||
Loading…
Add table
Add a link
Reference in a new issue