diff --git a/akka-actor-tests/src/test/scala/akka/pattern/AskSpec.scala b/akka-actor-tests/src/test/scala/akka/pattern/AskSpec.scala index 3e2b2ac2a8..67d7207d6a 100644 --- a/akka-actor-tests/src/test/scala/akka/pattern/AskSpec.scala +++ b/akka-actor-tests/src/test/scala/akka/pattern/AskSpec.scala @@ -12,6 +12,8 @@ import scala.concurrent.Await import scala.concurrent.duration._ import scala.util.Failure +import language.postfixOps + class AskSpec extends AkkaSpec with ScalaFutures { "The “ask” pattern" must { diff --git a/akka-actor/src/main/java/akka/japi/pf/DeciderBuilder.java b/akka-actor/src/main/java/akka/japi/pf/DeciderBuilder.java index 4c79703157..855a8c3c68 100644 --- a/akka-actor/src/main/java/akka/japi/pf/DeciderBuilder.java +++ b/akka-actor/src/main/java/akka/japi/pf/DeciderBuilder.java @@ -10,18 +10,19 @@ import static akka.actor.SupervisorStrategy.Directive; * Used for building a partial function for {@link akka.actor.Actor#supervisorStrategy() Actor.supervisorStrategy()}. * * * Inside an actor you can use it like this with Java 8 to define your supervisorStrategy. - *
+ ** Example: + *
*
- * @Override
+ * @Override
* private static SupervisorStrategy strategy =
* new OneForOneStrategy(10, Duration.create("1 minute"), DeciderBuilder.
- * match(ArithmeticException.class, e -> resume()).
- * match(NullPointerException.class, e -> restart()).
- * match(IllegalArgumentException.class, e -> stop()).
- * matchAny(o -> escalate()).build());
+ * match(ArithmeticException.class, e -> resume()).
+ * match(NullPointerException.class, e -> restart()).
+ * match(IllegalArgumentException.class, e -> stop()).
+ * matchAny(o -> escalate()).build());
*
- * @Override
+ * @Override
* public SupervisorStrategy supervisorStrategy() {
* return strategy;
* }
diff --git a/akka-actor/src/main/java/akka/japi/pf/Match.java b/akka-actor/src/main/java/akka/japi/pf/Match.java
index 09c5e9ff30..2ac8b2d3c8 100644
--- a/akka-actor/src/main/java/akka/japi/pf/Match.java
+++ b/akka-actor/src/main/java/akka/japi/pf/Match.java
@@ -90,10 +90,11 @@ public class Match extends AbstractMatch {
/**
* Convenience function to make the Java code more readable.
- *
+ *
+ *
*
* Matcher<X, Y> matcher = Matcher.create(...);
- *
+ *
* Y someY = matcher.match(obj);
*
*
diff --git a/akka-actor/src/main/java/akka/japi/pf/ReceiveBuilder.java b/akka-actor/src/main/java/akka/japi/pf/ReceiveBuilder.java
index 890b07f139..590bc3d35d 100644
--- a/akka-actor/src/main/java/akka/japi/pf/ReceiveBuilder.java
+++ b/akka-actor/src/main/java/akka/japi/pf/ReceiveBuilder.java
@@ -10,19 +10,20 @@ package akka.japi.pf;
* There is both a match on type only, and a match on type and predicate.
*
* Inside an actor you can use it like this with Java 8 to define your receive method.
- *
+ *
* Example:
+ *
*
- * @Override
+ * @Override
* public Actor() {
* receive(ReceiveBuilder.
- * match(Double.class, d -> {
+ * match(Double.class, d -> {
* sender().tell(d.isNaN() ? 0 : d, self());
* }).
- * match(Integer.class, i -> {
+ * match(Integer.class, i -> {
* sender().tell(i * 10, self());
* }).
- * match(String.class, s -> s.startsWith("foo"), s -> {
+ * match(String.class, s -> s.startsWith("foo"), s -> {
* sender().tell(s.toUpperCase(), self());
* }).build()
* );
diff --git a/akka-actor/src/main/java/akka/japi/pf/UnitMatch.java b/akka-actor/src/main/java/akka/japi/pf/UnitMatch.java
index 1ab252a039..bb8ec77634 100644
--- a/akka-actor/src/main/java/akka/japi/pf/UnitMatch.java
+++ b/akka-actor/src/main/java/akka/japi/pf/UnitMatch.java
@@ -110,7 +110,7 @@ public class UnitMatch extends AbstractMatch {
*
*
* UnitMatcher<X> matcher = UnitMatcher.create(...);
- *
+ *
* matcher.match(obj);
*
*
diff --git a/akka-actor/src/main/scala/akka/AkkaException.scala b/akka-actor/src/main/scala/akka/AkkaException.scala
index 9de9247980..f2929166c1 100644
--- a/akka-actor/src/main/scala/akka/AkkaException.scala
+++ b/akka-actor/src/main/scala/akka/AkkaException.scala
@@ -5,11 +5,7 @@
package akka
/**
- * Akka base Exception. Each Exception gets:
- *
* akka.actor.mailbox.requirements {
diff --git a/akka-actor/src/main/scala/akka/actor/ActorPath.scala b/akka-actor/src/main/scala/akka/actor/ActorPath.scala
index 6ed9997404..f9bf5c4dcd 100644
--- a/akka-actor/src/main/scala/akka/actor/ActorPath.scala
+++ b/akka-actor/src/main/scala/akka/actor/ActorPath.scala
@@ -27,7 +27,7 @@ object ActorPath {
/**
* Validates the given actor path element and throws an [[InvalidActorNameException]] if invalid.
- * See [[isValidPathElement()]] for a non-throwing version.
+ * See [[#isValidPathElement]] for a non-throwing version.
*
* @param element actor path element to be validated
*/
@@ -35,7 +35,7 @@ object ActorPath {
/**
* Validates the given actor path element and throws an [[InvalidActorNameException]] if invalid.
- * See [[isValidPathElement()]] for a non-throwing version.
+ * See [[#isValidPathElement]] for a non-throwing version.
*
* @param element actor path element to be validated
* @param fullPath optional fullPath element that may be included for better error messages; null if not given
@@ -60,7 +60,7 @@ object ActorPath {
/**
* This method is used to validate a path element (Actor Name).
* Since Actors form a tree, it is addressable using an URL, therefore an Actor Name has to conform to:
- * [[http://www.ietf.org/rfc/rfc2396.txt RFC-2396]].
+ * RFC-2396.
*
* User defined Actor names may not start from a `$` sign - these are reserved for system names.
*/
@@ -97,7 +97,7 @@ object ActorPath {
* ActorPath defines a natural ordering (so that ActorRefs can be put into
* collections with this requirement); this ordering is intended to be as fast
* as possible, which owing to the bottom-up recursive nature of ActorPath
- * is sorted by path elements FROM RIGHT TO LEFT, where RootActorPath >
+ * is sorted by path elements FROM RIGHT TO LEFT, where RootActorPath >
* ChildActorPath in case the number of elements is different.
*
* Two actor paths are compared equal when they have the same name and parent
diff --git a/akka-actor/src/main/scala/akka/actor/ActorRef.scala b/akka-actor/src/main/scala/akka/actor/ActorRef.scala
index d71a95ed65..719218ce1c 100644
--- a/akka-actor/src/main/scala/akka/actor/ActorRef.scala
+++ b/akka-actor/src/main/scala/akka/actor/ActorRef.scala
@@ -118,7 +118,7 @@ abstract class ActorRef extends java.lang.Comparable[ActorRef] with Serializable
* Sends the specified message to this ActorRef, i.e. fire-and-forget
* semantics, including the sender reference if possible.
*
- * Pass [[akka.actor.ActorRef$.noSender]] or `null` as sender if there is nobody to reply to
+ * Pass [[akka.actor.ActorRef]] `noSender` or `null` as sender if there is nobody to reply to
*/
final def tell(msg: Any, sender: ActorRef): Unit = this.!(msg)(sender)
@@ -157,7 +157,7 @@ abstract class ActorRef extends java.lang.Comparable[ActorRef] with Serializable
/**
* This trait represents the Scala Actor API
* There are implicit conversions in ../actor/Implicits.scala
- * from ActorRef -> ScalaActorRef and back
+ * from ActorRef -> ScalaActorRef and back
*/
trait ScalaActorRef { ref: ActorRef ⇒
diff --git a/akka-actor/src/main/scala/akka/actor/ActorRefProvider.scala b/akka-actor/src/main/scala/akka/actor/ActorRefProvider.scala
index af3fdc6e22..24818061e0 100644
--- a/akka-actor/src/main/scala/akka/actor/ActorRefProvider.scala
+++ b/akka-actor/src/main/scala/akka/actor/ActorRefProvider.scala
@@ -358,9 +358,9 @@ private[akka] object SystemGuardian {
/**
* For the purpose of orderly shutdown it's possible
* to register interest in the termination of systemGuardian
- * and receive a notification [[akka.actor.Guardian.TerminationHook]]
+ * and receive a notification `TerminationHook`
* before systemGuardian is stopped. The registered hook is supposed
- * to reply with [[akka.actor.Guardian.TerminationHookDone]] and the
+ * to reply with `TerminationHookDone` and the
* systemGuardian will not stop until all registered hooks have replied.
*/
case object RegisterTerminationHook
diff --git a/akka-actor/src/main/scala/akka/actor/ActorSelection.scala b/akka-actor/src/main/scala/akka/actor/ActorSelection.scala
index d7e202bf40..67051a53c1 100644
--- a/akka-actor/src/main/scala/akka/actor/ActorSelection.scala
+++ b/akka-actor/src/main/scala/akka/actor/ActorSelection.scala
@@ -101,7 +101,7 @@ abstract class ActorSelection extends Serializable {
/**
* String representation of the actor selection suitable for storage and recreation.
- * The output is similar to the URI fragment returned by [[akka.actor.ActorPath.toSerializationFormat]].
+ * The output is similar to the URI fragment returned by [[akka.actor.ActorPath#toSerializationFormat]].
* @return URI fragment
*/
def toSerializationFormat: String = {
diff --git a/akka-actor/src/main/scala/akka/actor/ActorSystem.scala b/akka-actor/src/main/scala/akka/actor/ActorSystem.scala
index 225c9b2f51..29ff241fc1 100644
--- a/akka-actor/src/main/scala/akka/actor/ActorSystem.scala
+++ b/akka-actor/src/main/scala/akka/actor/ActorSystem.scala
@@ -291,7 +291,7 @@ abstract class ActorSystem extends ActorRefFactory {
def logConfiguration(): Unit
/**
- * Construct a path below the application guardian to be used with [[ActorSystem.actorSelection]].
+ * Construct a path below the application guardian to be used with [[ActorSystem#actorSelection]].
*/
def /(name: String): ActorPath
@@ -301,7 +301,7 @@ abstract class ActorSystem extends ActorRefFactory {
def child(child: String): ActorPath = /(child)
/**
- * Construct a path below the application guardian to be used with [[ActorSystem.actorSelection]].
+ * Construct a path below the application guardian to be used with [[ActorSystem#actorSelection]].
*/
def /(name: Iterable[String]): ActorPath
@@ -326,7 +326,7 @@ abstract class ActorSystem extends ActorRefFactory {
def eventStream: EventStream
/**
- * Convenient logging adapter for logging to the [[ActorSystem.eventStream]].
+ * Convenient logging adapter for logging to the [[ActorSystem#eventStream]].
*/
def log: LoggingAdapter
@@ -369,7 +369,7 @@ abstract class ActorSystem extends ActorRefFactory {
* The callbacks will be run sequentially in reverse order of registration, i.e.
* last registration is run first.
*
- * @throws a RejectedExecutionException if the System has already shut down or if shutdown has been initiated.
+ * Throws a RejectedExecutionException if the System has already shut down or if shutdown has been initiated.
*
* Scala API
*/
@@ -382,7 +382,7 @@ abstract class ActorSystem extends ActorRefFactory {
* The callbacks will be run sequentially in reverse order of registration, i.e.
* last registration is run first.
*
- * @throws a RejectedExecutionException if the System has already shut down or if shutdown has been initiated.
+ * Throws a RejectedExecutionException if the System has already shut down or if shutdown has been initiated.
*/
def registerOnTermination(code: Runnable): Unit
@@ -391,7 +391,7 @@ abstract class ActorSystem extends ActorRefFactory {
* timeout has elapsed. This will block until after all on termination
* callbacks have been run.
*
- * @throws TimeoutException in case of timeout
+ * Throws TimeoutException in case of timeout.
*/
@deprecated("Use Await.result(whenTerminated, timeout) instead", "2.4")
def awaitTermination(timeout: Duration): Unit
@@ -407,7 +407,7 @@ abstract class ActorSystem extends ActorRefFactory {
* Stop this actor system. This will stop the guardian actor, which in turn
* will recursively stop all its child actors, then the system guardian
* (below which the logging actors reside) and the execute all registered
- * termination handlers (see [[ActorSystem.registerOnTermination]]).
+ * termination handlers (see [[ActorSystem#registerOnTermination]]).
*/
@deprecated("Use the terminate() method instead", "2.4")
def shutdown(): Unit
@@ -426,7 +426,7 @@ abstract class ActorSystem extends ActorRefFactory {
* Terminates this actor system. This will stop the guardian actor, which in turn
* will recursively stop all its child actors, then the system guardian
* (below which the logging actors reside) and the execute all registered
- * termination handlers (see [[ActorSystem.registerOnTermination]]).
+ * termination handlers (see [[ActorSystem#registerOnTermination]]).
*/
def terminate(): Future[Terminated]
@@ -835,7 +835,7 @@ private[akka] class ActorSystemImpl(
* Adds a Runnable that will be executed on ActorSystem termination.
* Note that callbacks are executed in reverse order of insertion.
* @param r The callback to be executed on ActorSystem termination
- * @throws RejectedExecutionException if called after ActorSystem has been terminated
+ * Throws RejectedExecutionException if called after ActorSystem has been terminated.
*/
final def add(r: Runnable): Unit = {
@tailrec def addRec(r: Runnable, p: Promise[T]): Unit = ref.get match {
diff --git a/akka-actor/src/main/scala/akka/actor/Address.scala b/akka-actor/src/main/scala/akka/actor/Address.scala
index 6bd03c2787..955f127522 100644
--- a/akka-actor/src/main/scala/akka/actor/Address.scala
+++ b/akka-actor/src/main/scala/akka/actor/Address.scala
@@ -46,7 +46,7 @@ final case class Address private (protocol: String, system: String, host: Option
/**
* Returns the canonical String representation of this Address formatted as:
*
- * ://@:
+ * `protocol://system@host:port`
*/
@transient
override lazy val toString: String = {
@@ -61,7 +61,7 @@ final case class Address private (protocol: String, system: String, host: Option
/**
* Returns a String representation formatted as:
*
- * @:
+ * `system@host:port`
*/
def hostPort: String = toString.substring(protocol.length + 3)
}
diff --git a/akka-actor/src/main/scala/akka/actor/Deployer.scala b/akka-actor/src/main/scala/akka/actor/Deployer.scala
index 2dbebab9fc..fde0247ecb 100644
--- a/akka-actor/src/main/scala/akka/actor/Deployer.scala
+++ b/akka-actor/src/main/scala/akka/actor/Deployer.scala
@@ -60,7 +60,7 @@ final case class Deploy(
/**
* Do a merge between this and the other Deploy, where values from “this” take
* precedence. The “path” of the other Deploy is not taken into account. All
- * other members are merged using ``.withFallback(other.)``.
+ * other members are merged using `X.withFallback(other.X)`.
*/
def withFallback(other: Deploy): Deploy = {
Deploy(
diff --git a/akka-actor/src/main/scala/akka/actor/FSM.scala b/akka-actor/src/main/scala/akka/actor/FSM.scala
index 2234e84b4a..09e3597ff1 100644
--- a/akka-actor/src/main/scala/akka/actor/FSM.scala
+++ b/akka-actor/src/main/scala/akka/actor/FSM.scala
@@ -70,7 +70,7 @@ object FSM {
/**
* Signifies that the [[akka.actor.FSM]] is shutting itself down because of
* an error, e.g. if the state to transition into does not exist. You can use
- * this to communicate a more precise cause to the [[akka.actor.FSM.onTermination]] block.
+ * this to communicate a more precise cause to the `onTermination` block.
*/
final case class Failure(cause: Any) extends Reason
@@ -374,7 +374,7 @@ trait FSM[S, D] extends Actor with Listeners with ActorLogging {
* Return this from a state function when no state change is to be effected.
*
* No transition event will be triggered by [[#stay]].
- * If you want to trigger an event like `S -> S` for [[#onTransition]] to handle use [[#goto]] instead.
+ * If you want to trigger an event like `S -> S` for `onTransition` to handle use `goto` instead.
*
* @return descriptor for staying in current state
*/
@@ -410,7 +410,6 @@ trait FSM[S, D] extends Actor with Listeners with ActorLogging {
* @param msg message to be delivered
* @param timeout delay of first message delivery and between subsequent messages
* @param repeat send once if false, scheduleAtFixedRate if true
- * @return current state descriptor
*/
final def setTimer(name: String, msg: Any, timeout: FiniteDuration, repeat: Boolean = false): Unit = {
if (debugEvent)
@@ -1160,7 +1159,6 @@ abstract class AbstractFSM[S, D] extends FSM[S, D] {
* @param name identifier to be used with cancelTimer()
* @param msg message to be delivered
* @param timeout delay of first message delivery and between subsequent messages
- * @return current state descriptor
*/
final def setTimer(name: String, msg: Any, timeout: FiniteDuration): Unit =
setTimer(name, msg, timeout, false)
diff --git a/akka-actor/src/main/scala/akka/actor/FaultHandling.scala b/akka-actor/src/main/scala/akka/actor/FaultHandling.scala
index a8ae39865d..40d4f51db6 100644
--- a/akka-actor/src/main/scala/akka/actor/FaultHandling.scala
+++ b/akka-actor/src/main/scala/akka/actor/FaultHandling.scala
@@ -145,7 +145,7 @@ object SupervisorStrategy extends SupervisorStrategyLowPriorityImplicits {
/**
* When supervisorStrategy is not specified for an actor this
- * [[Decider]] is used by default in the supervisor strategy.
+ * `Decider` is used by default in the supervisor strategy.
* The child will be stopped when [[akka.actor.ActorInitializationException]],
* [[akka.actor.ActorKilledException]], or [[akka.actor.DeathPactException]] is
* thrown. It will be restarted for other `Exception` types.
diff --git a/akka-actor/src/main/scala/akka/actor/Scheduler.scala b/akka-actor/src/main/scala/akka/actor/Scheduler.scala
index 92ad6e4ac1..0ea99a562a 100644
--- a/akka-actor/src/main/scala/akka/actor/Scheduler.scala
+++ b/akka-actor/src/main/scala/akka/actor/Scheduler.scala
@@ -187,7 +187,7 @@ trait Cancellable {
* when scheduling single-shot tasks, instead it always rounds up the task
* delay to a full multiple of the TickDuration. This means that tasks are
* scheduled possibly one tick later than they could be (if checking that
- * “now() + delay <= nextTick” were done).
+ * “now() + delay <= nextTick” were done).
*/
class LightArrayRevolverScheduler(config: Config,
log: LoggingAdapter,
diff --git a/akka-actor/src/main/scala/akka/actor/TypedActor.scala b/akka-actor/src/main/scala/akka/actor/TypedActor.scala
index e597b16c5e..50198a61aa 100644
--- a/akka-actor/src/main/scala/akka/actor/TypedActor.scala
+++ b/akka-actor/src/main/scala/akka/actor/TypedActor.scala
@@ -138,7 +138,7 @@ object TypedActor extends ExtensionId[TypedActorExtension] with ExtensionIdProvi
/**
* Invokes the Method on the supplied instance
*
- * @throws the underlying exception if there's an InvocationTargetException thrown on the invocation
+ * Throws the underlying exception if there's an InvocationTargetException thrown on the invocation.
*/
def apply(instance: AnyRef): AnyRef = try {
parameters match {
@@ -217,8 +217,9 @@ object TypedActor extends ExtensionId[TypedActorExtension] with ExtensionIdProvi
*
* NEVER EXPOSE "this" to someone else, always use "self[TypeOfInterface(s)]"
*
- * @throws IllegalStateException if called outside of the scope of a method on this TypedActor
- * @throws ClassCastException if the supplied type T isn't the type of the proxy associated with this TypedActor
+ * Throws IllegalStateException if called outside of the scope of a method on this TypedActor.
+ *
+ * Throws ClassCastException if the supplied type T isn't the type of the proxy associated with this TypedActor.
*/
def self[T <: AnyRef] = selfReference.get.asInstanceOf[T] match {
case null ⇒ throw new IllegalStateException("Calling TypedActor.self outside of a TypedActor implementation method!")
diff --git a/akka-actor/src/main/scala/akka/actor/UntypedActorWithStash.scala b/akka-actor/src/main/scala/akka/actor/UntypedActorWithStash.scala
index 111a85f4cd..ae5450aee8 100644
--- a/akka-actor/src/main/scala/akka/actor/UntypedActorWithStash.scala
+++ b/akka-actor/src/main/scala/akka/actor/UntypedActorWithStash.scala
@@ -30,7 +30,7 @@ package akka.actor
* }
*
* Note that the subclasses of `UntypedActorWithStash` by default request a Deque based mailbox since this class
- * implements the `RequiresMessageQueue
* akka.actor.mailbox.requirements {
diff --git a/akka-actor/src/main/scala/akka/actor/dsl/Inbox.scala b/akka-actor/src/main/scala/akka/actor/dsl/Inbox.scala
index 6e7f4d773b..dc36cae14b 100644
--- a/akka-actor/src/main/scala/akka/actor/dsl/Inbox.scala
+++ b/akka-actor/src/main/scala/akka/actor/dsl/Inbox.scala
@@ -157,8 +157,8 @@ trait Inbox { this: ActorDSL.type ⇒
/**
* Create a new actor which will internally queue up messages it gets so that
- * they can be interrogated with the [[akka.actor.dsl.Inbox!.Inbox!.receive]]
- * and [[akka.actor.dsl.Inbox!.Inbox!.select]] methods. It will be created as
+ * they can be interrogated with the `akka.actor.dsl.Inbox!.Inbox!.receive`
+ * and `akka.actor.dsl.Inbox!.Inbox!.select` methods. It will be created as
* a system actor in the ActorSystem which is implicitly (or explicitly)
* supplied.
*/
diff --git a/akka-actor/src/main/scala/akka/dispatch/Dispatchers.scala b/akka-actor/src/main/scala/akka/dispatch/Dispatchers.scala
index 0ea91952ea..cfe484c8ad 100644
--- a/akka-actor/src/main/scala/akka/dispatch/Dispatchers.scala
+++ b/akka-actor/src/main/scala/akka/dispatch/Dispatchers.scala
@@ -76,7 +76,7 @@ class Dispatchers(val settings: ActorSystem.Settings, val prerequisites: Dispatc
* Returns a dispatcher as specified in configuration. Please note that this
* method _may_ create and return a NEW dispatcher, _every_ call.
*
- * @throws ConfigurationException if the specified dispatcher cannot be found in the configuration
+ * Throws ConfigurationException if the specified dispatcher cannot be found in the configuration.
*/
def lookup(id: String): MessageDispatcher = lookupConfigurator(id).dispatcher()
diff --git a/akka-actor/src/main/scala/akka/dispatch/Future.scala b/akka-actor/src/main/scala/akka/dispatch/Future.scala
index 8618c104fd..6af5853aa1 100644
--- a/akka-actor/src/main/scala/akka/dispatch/Future.scala
+++ b/akka-actor/src/main/scala/akka/dispatch/Future.scala
@@ -41,7 +41,7 @@ object ExecutionContexts {
* Returns a new ExecutionContextExecutorService which will delegate execution to the underlying ExecutorService,
* and which will use the default error reporter.
*
- * @param executor the ExecutorService which will be used for the ExecutionContext
+ * @param executorService the ExecutorService which will be used for the ExecutionContext
* @return a new ExecutionContext
*/
def fromExecutorService(executorService: ExecutorService): ExecutionContextExecutorService =
@@ -51,7 +51,7 @@ object ExecutionContexts {
* Returns a new ExecutionContextExecutorService which will delegate execution to the underlying ExecutorService,
* and which will use the provided error reporter.
*
- * @param executor the ExecutorService which will be used for the ExecutionContext
+ * @param executorService the ExecutorService which will be used for the ExecutionContext
* @param errorReporter a Procedure that will log any exceptions passed to it
* @return a new ExecutionContext
*/
@@ -272,7 +272,7 @@ abstract class Recover[+T] extends japi.RecoverBridge[T] {
* becomes completed with a failure.
*
* @return a successful value for the passed in failure
- * @throws the passed in failure to propagate it.
+ * Throws the passed in failure to propagate it.
*
* Java API
*/
@@ -350,7 +350,7 @@ abstract class Mapper[-T, +R] extends scala.runtime.AbstractFunction1[T, R] {
/**
* Override this method if you need to throw checked exceptions
*
- * @throws UnsupportedOperation by default
+ * Throws UnsupportedOperation by default.
*/
@throws(classOf[Throwable])
def checkedApply(parameter: T): R = throw new UnsupportedOperationException("Mapper.checkedApply has not been implemented")
diff --git a/akka-actor/src/main/scala/akka/dispatch/sysmsg/SystemMessage.scala b/akka-actor/src/main/scala/akka/dispatch/sysmsg/SystemMessage.scala
index 0dff418daa..697dcb3edd 100644
--- a/akka-actor/src/main/scala/akka/dispatch/sysmsg/SystemMessage.scala
+++ b/akka-actor/src/main/scala/akka/dispatch/sysmsg/SystemMessage.scala
@@ -182,7 +182,7 @@ private[akka] class EarliestFirstSystemMessageList(val head: SystemMessage) exte
*
* INTERNAL API
*
- * ➡➡➡ NEVER SEND THE SAME SYSTEM MESSAGE OBJECT TO TWO ACTORS ⬅⬅⬅
+ * NEVER SEND THE SAME SYSTEM MESSAGE OBJECT TO TWO ACTORS
*/
private[akka] sealed trait SystemMessage extends PossiblyHarmful with Serializable {
// Next fields are only modifiable via the SystemMessageList value class
diff --git a/akka-actor/src/main/scala/akka/event/EventBusUnsubscribers.scala b/akka-actor/src/main/scala/akka/event/EventBusUnsubscribers.scala
index 53c146abf6..7bc3ee25c5 100644
--- a/akka-actor/src/main/scala/akka/event/EventBusUnsubscribers.scala
+++ b/akka-actor/src/main/scala/akka/event/EventBusUnsubscribers.scala
@@ -51,7 +51,7 @@ private[akka] class EventStreamUnsubscriber(eventStream: EventStream, debug: Boo
* INTERNAL API
*
* Provides factory for [[akka.event.EventStreamUnsubscriber]] actors with **unique names**.
- * This is needed if someone spins up more [[EventStream]]s using the same [[ActorSystem]],
+ * This is needed if someone spins up more [[EventStream]]s using the same [[akka.actor.ActorSystem]],
* each stream gets it's own unsubscriber.
*/
private[akka] object EventStreamUnsubscriber {
diff --git a/akka-actor/src/main/scala/akka/event/Logging.scala b/akka-actor/src/main/scala/akka/event/Logging.scala
index 447b2c3928..d77976a0b0 100644
--- a/akka-actor/src/main/scala/akka/event/Logging.scala
+++ b/akka-actor/src/main/scala/akka/event/Logging.scala
@@ -251,8 +251,8 @@ class DummyClassForStringSources
* In case an [[akka.actor.ActorSystem]] is provided, the following apply:
*
+ * {{{
* val log = Logging(<bus>, <source object>)
* ...
* log.info("hello world!")
- *
+ * }}}
*
* All log-level methods support simple interpolation templates with up to four
* arguments placed by using {} within the template (first string
* argument):
*
- *
+ * {{{
* log.error(exception, "Exception while processing {} in state {}", msg, state)
- *
+ * }}}
*/
trait LoggingAdapter {
@@ -1140,7 +1140,7 @@ trait DiagnosticLoggingAdapter extends LoggingAdapter {
/**
* Scala API:
* Mapped Diagnostic Context for application defined values
- * which can be used in PatternLayout when [[akka.event.slf4j.Slf4jLogger]] is configured.
+ * which can be used in PatternLayout when `akka.event.slf4j.Slf4jLogger` is configured.
* Visit Logback Docs: MDC for more information.
*
* @return A Map containing the MDC values added by the application, or empty Map if no value was added.
@@ -1150,7 +1150,7 @@ trait DiagnosticLoggingAdapter extends LoggingAdapter {
/**
* Scala API:
* Sets the values to be added to the MDC (Mapped Diagnostic Context) before the log is appended.
- * These values can be used in PatternLayout when [[akka.event.slf4j.Slf4jLogger]] is configured.
+ * These values can be used in PatternLayout when `akka.event.slf4j.Slf4jLogger` is configured.
* Visit Logback Docs: MDC for more information.
*/
def mdc(mdc: MDC): Unit = _mdc = if (mdc != null) mdc else emptyMDC
@@ -1158,17 +1158,18 @@ trait DiagnosticLoggingAdapter extends LoggingAdapter {
/**
* Java API:
* Mapped Diagnostic Context for application defined values
- * which can be used in PatternLayout when [[akka.event.slf4j.Slf4jLogger]] is configured.
+ * which can be used in PatternLayout when `akka.event.slf4j.Slf4jLogger` is configured.
* Visit Logback Docs: MDC for more information.
* Note tha it returns a COPY of the actual MDC values.
* You cannot modify any value by changing the returned Map.
* Code like the following won't have any effect unless you set back the modified Map.
- *
+ *
+ * {{{
* Map mdc = log.getMDC();
* mdc.put("key", value);
* // NEEDED
* log.setMDC(mdc);
- *
+ * }}}
*
* @return A copy of the actual MDC values
*/
@@ -1177,7 +1178,7 @@ trait DiagnosticLoggingAdapter extends LoggingAdapter {
/**
* Java API:
* Sets the values to be added to the MDC (Mapped Diagnostic Context) before the log is appended.
- * These values can be used in PatternLayout when [[akka.event.slf4j.Slf4jLogger]] is configured.
+ * These values can be used in PatternLayout when `akka.event.slf4j.Slf4jLogger` is configured.
* Visit Logback Docs: MDC for more information.
*/
def setMDC(jMdc: java.util.Map[String, Any]): Unit = mdc(if (jMdc != null) jMdc.asScala.toMap else emptyMDC)
diff --git a/akka-actor/src/main/scala/akka/io/Inet.scala b/akka-actor/src/main/scala/akka/io/Inet.scala
index 289722c601..105e00c519 100644
--- a/akka-actor/src/main/scala/akka/io/Inet.scala
+++ b/akka-actor/src/main/scala/akka/io/Inet.scala
@@ -78,7 +78,7 @@ object Inet {
/**
* Open and return new DatagramChannel.
*
- * [[scala.throws]] is needed because [[DatagramChannel.open]] method
+ * `throws` is needed because `DatagramChannel.open` method
* can throw an exception.
*/
@throws(classOf[Exception])
@@ -95,7 +95,7 @@ object Inet {
/**
* [[akka.io.Inet.SocketOption]] to set the SO_RCVBUF option
*
- * For more information see [[java.net.Socket.setReceiveBufferSize]]
+ * For more information see [[java.net.Socket#setReceiveBufferSize]]
*/
final case class ReceiveBufferSize(size: Int) extends SocketOption {
require(size > 0, "ReceiveBufferSize must be > 0")
@@ -109,7 +109,7 @@ object Inet {
/**
* [[akka.io.Inet.SocketOption]] to enable or disable SO_REUSEADDR
*
- * For more information see [[java.net.Socket.setReuseAddress]]
+ * For more information see [[java.net.Socket#setReuseAddress]]
*/
final case class ReuseAddress(on: Boolean) extends SocketOption {
override def beforeServerSocketBind(s: ServerSocket): Unit = s.setReuseAddress(on)
@@ -120,7 +120,7 @@ object Inet {
/**
* [[akka.io.Inet.SocketOption]] to set the SO_SNDBUF option.
*
- * For more information see [[java.net.Socket.setSendBufferSize]]
+ * For more information see [[java.net.Socket#setSendBufferSize]]
*/
final case class SendBufferSize(size: Int) extends SocketOption {
require(size > 0, "SendBufferSize must be > 0")
@@ -132,7 +132,7 @@ object Inet {
* type-of-service octet in the IP header for packets sent from this
* socket.
*
- * For more information see [[java.net.Socket.setTrafficClass]]
+ * For more information see [[java.net.Socket#setTrafficClass]]
*/
final case class TrafficClass(tc: Int) extends SocketOption {
require(0 <= tc && tc <= 255, "TrafficClass needs to be in the interval [0, 255]")
@@ -145,21 +145,21 @@ object Inet {
/**
* [[akka.io.Inet.SocketOption]] to set the SO_RCVBUF option
*
- * For more information see [[java.net.Socket.setReceiveBufferSize]]
+ * For more information see [[java.net.Socket#setReceiveBufferSize]]
*/
val ReceiveBufferSize = SO.ReceiveBufferSize
/**
* [[akka.io.Inet.SocketOption]] to enable or disable SO_REUSEADDR
*
- * For more information see [[java.net.Socket.setReuseAddress]]
+ * For more information see [[java.net.Socket#setReuseAddress]]
*/
val ReuseAddress = SO.ReuseAddress
/**
* [[akka.io.Inet.SocketOption]] to set the SO_SNDBUF option.
*
- * For more information see [[java.net.Socket.setSendBufferSize]]
+ * For more information see [[java.net.Socket#setSendBufferSize]]
*/
val SendBufferSize = SO.SendBufferSize
@@ -168,7 +168,7 @@ object Inet {
* type-of-service octet in the IP header for packets sent from this
* socket.
*
- * For more information see [[java.net.Socket.setTrafficClass]]
+ * For more information see [[java.net.Socket#setTrafficClass]]
*/
val TrafficClass = SO.TrafficClass
}
@@ -178,21 +178,21 @@ object Inet {
/**
* [[akka.io.Inet.SocketOption]] to set the SO_RCVBUF option
*
- * For more information see [[java.net.Socket.setReceiveBufferSize]]
+ * For more information see [[java.net.Socket#setReceiveBufferSize]]
*/
def receiveBufferSize(size: Int) = ReceiveBufferSize(size)
/**
* [[akka.io.Inet.SocketOption]] to enable or disable SO_REUSEADDR
*
- * For more information see [[java.net.Socket.setReuseAddress]]
+ * For more information see [[java.net.Socket#setReuseAddress]]
*/
def reuseAddress(on: Boolean) = ReuseAddress(on)
/**
* [[akka.io.Inet.SocketOption]] to set the SO_SNDBUF option.
*
- * For more information see [[java.net.Socket.setSendBufferSize]]
+ * For more information see [[java.net.Socket#setSendBufferSize]]
*/
def sendBufferSize(size: Int) = SendBufferSize(size)
@@ -201,7 +201,7 @@ object Inet {
* type-of-service octet in the IP header for packets sent from this
* socket.
*
- * For more information see [[java.net.Socket.setTrafficClass]]
+ * For more information see [[java.net.Socket#setTrafficClass]]
*/
def trafficClass(tc: Int) = TrafficClass(tc)
}
diff --git a/akka-actor/src/main/scala/akka/io/Tcp.scala b/akka-actor/src/main/scala/akka/io/Tcp.scala
index 9d77d53129..cb4f964a40 100644
--- a/akka-actor/src/main/scala/akka/io/Tcp.scala
+++ b/akka-actor/src/main/scala/akka/io/Tcp.scala
@@ -54,7 +54,7 @@ object Tcp extends ExtensionId[TcpExt] with ExtensionIdProvider {
/**
* [[akka.io.Inet.SocketOption]] to enable or disable SO_KEEPALIVE
*
- * For more information see [[java.net.Socket.setKeepAlive]]
+ * For more information see `java.net.Socket.setKeepAlive`
*/
final case class KeepAlive(on: Boolean) extends SocketOption {
override def afterConnect(s: Socket): Unit = s.setKeepAlive(on)
@@ -65,7 +65,7 @@ object Tcp extends ExtensionId[TcpExt] with ExtensionIdProvider {
* of TCP urgent data) By default, this option is disabled and TCP urgent
* data is silently discarded.
*
- * For more information see [[java.net.Socket.setOOBInline]]
+ * For more information see `java.net.Socket.setOOBInline`
*/
final case class OOBInline(on: Boolean) extends SocketOption {
override def afterConnect(s: Socket): Unit = s.setOOBInline(on)
@@ -79,7 +79,7 @@ object Tcp extends ExtensionId[TcpExt] with ExtensionIdProvider {
*
* Please note, that TCP_NODELAY is enabled by default.
*
- * For more information see [[java.net.Socket.setTcpNoDelay]]
+ * For more information see `java.net.Socket.setTcpNoDelay`
*/
final case class TcpNoDelay(on: Boolean) extends SocketOption {
override def afterConnect(s: Socket): Unit = s.setTcpNoDelay(on)
@@ -109,7 +109,7 @@ object Tcp extends ExtensionId[TcpExt] with ExtensionIdProvider {
*
* @param remoteAddress is the address to connect to
* @param localAddress optionally specifies a specific address to bind to
- * @param options Please refer to the [[SO]] object for a list of all supported options.
+ * @param options Please refer to the `Tcp.SO` object for a list of all supported options.
*/
final case class Connect(remoteAddress: InetSocketAddress,
localAddress: Option[InetSocketAddress] = None,
@@ -134,7 +134,7 @@ object Tcp extends ExtensionId[TcpExt] with ExtensionIdProvider {
* @param backlog This specifies the number of unaccepted connections the O/S
* kernel will hold for this port before refusing connections.
*
- * @param options Please refer to the [[SO]] object for a list of all supported options.
+ * @param options Please refer to the `Tcp.SO` object for a list of all supported options.
*/
final case class Bind(handler: ActorRef,
localAddress: InetSocketAddress,
@@ -153,11 +153,11 @@ object Tcp extends ExtensionId[TcpExt] with ExtensionIdProvider {
*
* @param keepOpenOnPeerClosed If this is set to true then the connection
* is not automatically closed when the peer closes its half,
- * requiring an explicit [[Closed]] from our side when finished.
+ * requiring an explicit [[CloseCommand]] from our side when finished.
*
* @param useResumeWriting If this is set to true then the connection actor
* will refuse all further writes after issuing a [[CommandFailed]]
- * notification until [[ResumeWriting]] is received. This can
+ * notification until `ResumeWriting` is received. This can
* be used to implement NACK-based write backpressure.
*/
final case class Register(handler: ActorRef, keepOpenOnPeerClosed: Boolean = false, useResumeWriting: Boolean = true) extends Command
@@ -183,7 +183,7 @@ object Tcp extends ExtensionId[TcpExt] with ExtensionIdProvider {
/**
* A normal close operation will first flush pending writes and then close the
* socket. The sender of this command and the registered handler for incoming
- * data will both be notified once the socket is closed using a [[Closed]]
+ * data will both be notified once the socket is closed using a `Closed`
* message.
*/
case object Close extends CloseCommand {
@@ -198,7 +198,7 @@ object Tcp extends ExtensionId[TcpExt] with ExtensionIdProvider {
* A confirmed close operation will flush pending writes and half-close the
* connection, waiting for the peer to close the other half. The sender of this
* command and the registered handler for incoming data will both be notified
- * once the socket is closed using a [[ConfirmedClosed]] message.
+ * once the socket is closed using a `ConfirmedClosed` message.
*/
case object ConfirmedClose extends CloseCommand {
/**
@@ -213,7 +213,7 @@ object Tcp extends ExtensionId[TcpExt] with ExtensionIdProvider {
* command to the O/S kernel which should result in a TCP_RST packet being sent
* to the peer. The sender of this command and the registered handler for
* incoming data will both be notified once the socket is closed using a
- * [[Aborted]] message.
+ * `Aborted` message.
*/
case object Abort extends CloseCommand {
/**
@@ -225,7 +225,7 @@ object Tcp extends ExtensionId[TcpExt] with ExtensionIdProvider {
/**
* Each [[WriteCommand]] can optionally request a positive acknowledgment to be sent
- * to the commanding actor. If such notification is not desired the [[WriteCommand#ack]]
+ * to the commanding actor. If such notification is not desired the [[SimpleWriteCommand#ack]]
* must be set to an instance of this class. The token contained within can be used
* to recognize which write failed when receiving a [[CommandFailed]] message.
*/
@@ -238,7 +238,7 @@ object Tcp extends ExtensionId[TcpExt] with ExtensionIdProvider {
object NoAck extends NoAck(null)
/**
- * Common interface for all write commands, currently [[Write]], [[WriteFile]] and [[CompoundWrite]].
+ * Common interface for all write commands.
*/
sealed abstract class WriteCommand extends Command {
/**
@@ -310,8 +310,8 @@ object Tcp extends ExtensionId[TcpExt] with ExtensionIdProvider {
/**
* Write data to the TCP connection. If no ack is needed use the special
* `NoAck` object. The connection actor will reply with a [[CommandFailed]]
- * message if the write could not be enqueued. If [[WriteCommand#wantsAck]]
- * returns true, the connection actor will reply with the supplied [[WriteCommand#ack]]
+ * message if the write could not be enqueued. If [[SimpleWriteCommand#wantsAck]]
+ * returns true, the connection actor will reply with the supplied [[SimpleWriteCommand#ack]]
* token once the write has been successfully enqueued to the O/S kernel.
* Note that this does not in any way guarantee that the data will be
* or have been sent! Unfortunately there is no way to determine whether
@@ -336,9 +336,9 @@ object Tcp extends ExtensionId[TcpExt] with ExtensionIdProvider {
/**
* Write `count` bytes starting at `position` from file at `filePath` to the connection.
- * The count must be > 0. The connection actor will reply with a [[CommandFailed]]
- * message if the write could not be enqueued. If [[WriteCommand#wantsAck]]
- * returns true, the connection actor will reply with the supplied [[WriteCommand#ack]]
+ * The count must be > 0. The connection actor will reply with a [[CommandFailed]]
+ * message if the write could not be enqueued. If [[SimpleWriteCommand#wantsAck]]
+ * returns true, the connection actor will reply with the supplied [[SimpleWriteCommand#ack]]
* token once the write has been successfully enqueued to the O/S kernel.
* Note that this does not in any way guarantee that the data will be
* or have been sent! Unfortunately there is no way to determine whether
@@ -385,12 +385,12 @@ object Tcp extends ExtensionId[TcpExt] with ExtensionIdProvider {
/**
* Sending this command to the connection actor will disable reading from the TCP
* socket. TCP flow-control will then propagate backpressure to the sender side
- * as buffers fill up on either end. To re-enable reading send [[ResumeReading]].
+ * as buffers fill up on either end. To re-enable reading send `ResumeReading`.
*/
case object SuspendReading extends Command
/**
- * This command needs to be sent to the connection actor after a [[SuspendReading]]
+ * This command needs to be sent to the connection actor after a `SuspendReading`
* command in order to resume reading from the socket.
*/
case object ResumeReading extends Command
@@ -430,7 +430,7 @@ object Tcp extends ExtensionId[TcpExt] with ExtensionIdProvider {
/**
* When `useResumeWriting` is in effect as indicated in the [[Register]] message,
- * the [[ResumeWriting]] command will be acknowledged by this message type, upon
+ * the `ResumeWriting` command will be acknowledged by this message type, upon
* which it is safe to send at least one write. This means that all writes preceding
* the first [[CommandFailed]] message have been enqueued to the O/S kernel at this
* point.
@@ -446,7 +446,7 @@ object Tcp extends ExtensionId[TcpExt] with ExtensionIdProvider {
final case class Bound(localAddress: InetSocketAddress) extends Event
/**
- * The sender of an [[Unbind]] command will receive confirmation through this
+ * The sender of an `Unbind` command will receive confirmation through this
* message once the listening socket has been closed.
*/
sealed trait Unbound extends Event
@@ -458,12 +458,12 @@ object Tcp extends ExtensionId[TcpExt] with ExtensionIdProvider {
*/
sealed trait ConnectionClosed extends Event with DeadLetterSuppression {
/**
- * `true` iff the connection has been closed in response to an [[Abort]] command.
+ * `true` iff the connection has been closed in response to an `Abort` command.
*/
def isAborted: Boolean = false
/**
* `true` iff the connection has been fully closed in response to a
- * [[ConfirmedClose]] command.
+ * `ConfirmedClose` command.
*/
def isConfirmed: Boolean = false
/**
@@ -483,18 +483,18 @@ object Tcp extends ExtensionId[TcpExt] with ExtensionIdProvider {
def getErrorCause: String = null
}
/**
- * The connection has been closed normally in response to a [[Close]] command.
+ * The connection has been closed normally in response to a `Close` command.
*/
case object Closed extends ConnectionClosed
/**
- * The connection has been aborted in response to an [[Abort]] command.
+ * The connection has been aborted in response to an `Abort` command.
*/
case object Aborted extends ConnectionClosed {
override def isAborted = true
}
/**
* The connection has been half-closed by us and then half-close by the peer
- * in response to a [[ConfirmedClose]] command.
+ * in response to a `ConfirmedClose` command.
*/
case object ConfirmedClosed extends ConnectionClosed {
override def isConfirmed = true
@@ -585,7 +585,7 @@ object TcpSO extends SoJavaFactories {
/**
* [[akka.io.Inet.SocketOption]] to enable or disable SO_KEEPALIVE
*
- * For more information see [[java.net.Socket.setKeepAlive]]
+ * For more information see `java.net.Socket.setKeepAlive`
*/
def keepAlive(on: Boolean) = KeepAlive(on)
@@ -594,7 +594,7 @@ object TcpSO extends SoJavaFactories {
* of TCP urgent data) By default, this option is disabled and TCP urgent
* data is silently discarded.
*
- * For more information see [[java.net.Socket.setOOBInline]]
+ * For more information see `java.net.Socket.setOOBInline`
*/
def oobInline(on: Boolean) = OOBInline(on)
@@ -604,7 +604,7 @@ object TcpSO extends SoJavaFactories {
*
* Please note, that TCP_NODELAY is enabled by default.
*
- * For more information see [[java.net.Socket.setTcpNoDelay]]
+ * For more information see `java.net.Socket.setTcpNoDelay`
*/
def tcpNoDelay(on: Boolean) = TcpNoDelay(on)
}
@@ -648,8 +648,8 @@ object TcpMessage {
* @param handler The actor which will receive all incoming connection requests
* in the form of [[Tcp.Connected]] messages.
*
- * @param localAddress The socket address to bind to; use port zero for
- * automatic assignment (i.e. an ephemeral port, see [[Bound]])
+ * @param endpoint The socket address to bind to; use port zero for
+ * automatic assignment (i.e. an ephemeral port, see [[Tcp.Bound]])
*
* @param backlog This specifies the number of unaccepted connections the O/S
* kernel will hold for this port before refusing connections.
@@ -682,11 +682,11 @@ object TcpMessage {
*
* @param keepOpenOnPeerClosed If this is set to true then the connection
* is not automatically closed when the peer closes its half,
- * requiring an explicit [[Tcp.Closed]] from our side when finished.
+ * requiring an explicit `Tcp.ConnectionClosed from our side when finished.
*
* @param useResumeWriting If this is set to true then the connection actor
* will refuse all further writes after issuing a [[Tcp.CommandFailed]]
- * notification until [[Tcp.ResumeWriting]] is received. This can
+ * notification until [[Tcp]] `ResumeWriting` is received. This can
* be used to implement NACK-based write backpressure.
*/
def register(handler: ActorRef, keepOpenOnPeerClosed: Boolean, useResumeWriting: Boolean): Command =
@@ -699,14 +699,14 @@ object TcpMessage {
/**
* In order to close down a listening socket, send this message to that socket’s
* actor (that is the actor which previously had sent the [[Tcp.Bound]] message). The
- * listener socket actor will reply with a [[Tcp.Unbound]] message.
+ * listener socket actor will reply with a `Tcp.Unbound` message.
*/
def unbind: Command = Unbind
/**
* A normal close operation will first flush pending writes and then close the
* socket. The sender of this command and the registered handler for incoming
- * data will both be notified once the socket is closed using a [[Tcp.Closed]]
+ * data will both be notified once the socket is closed using a `Tcp.Closed`
* message.
*/
def close: Command = Close
@@ -715,7 +715,7 @@ object TcpMessage {
* A confirmed close operation will flush pending writes and half-close the
* connection, waiting for the peer to close the other half. The sender of this
* command and the registered handler for incoming data will both be notified
- * once the socket is closed using a [[Tcp.ConfirmedClosed]] message.
+ * once the socket is closed using a `Tcp.ConfirmedClosed` message.
*/
def confirmedClose: Command = ConfirmedClose
@@ -724,13 +724,13 @@ object TcpMessage {
* command to the O/S kernel which should result in a TCP_RST packet being sent
* to the peer. The sender of this command and the registered handler for
* incoming data will both be notified once the socket is closed using a
- * [[Tcp.Aborted]] message.
+ * `Tcp.Aborted` message.
*/
def abort: Command = Abort
/**
* Each [[Tcp.WriteCommand]] can optionally request a positive acknowledgment to be sent
- * to the commanding actor. If such notification is not desired the [[Tcp.WriteCommand#ack]]
+ * to the commanding actor. If such notification is not desired the [[Tcp.SimpleWriteCommand#ack]]
* must be set to an instance of this class. The token contained within can be used
* to recognize which write failed when receiving a [[Tcp.CommandFailed]] message.
*/
@@ -744,8 +744,8 @@ object TcpMessage {
/**
* Write data to the TCP connection. If no ack is needed use the special
* `NoAck` object. The connection actor will reply with a [[Tcp.CommandFailed]]
- * message if the write could not be enqueued. If [[Tcp.WriteCommand#wantsAck]]
- * returns true, the connection actor will reply with the supplied [[Tcp.WriteCommand#ack]]
+ * message if the write could not be enqueued. If [[Tcp.SimpleWriteCommand#wantsAck]]
+ * returns true, the connection actor will reply with the supplied [[Tcp.SimpleWriteCommand#ack]]
* token once the write has been successfully enqueued to the O/S kernel.
* Note that this does not in any way guarantee that the data will be
* or have been sent! Unfortunately there is no way to determine whether
@@ -759,9 +759,9 @@ object TcpMessage {
/**
* Write `count` bytes starting at `position` from file at `filePath` to the connection.
- * The count must be > 0. The connection actor will reply with a [[Tcp.CommandFailed]]
- * message if the write could not be enqueued. If [[Tcp.WriteCommand#wantsAck]]
- * returns true, the connection actor will reply with the supplied [[Tcp.WriteCommand#ack]]
+ * The count must be > 0. The connection actor will reply with a [[Tcp.CommandFailed]]
+ * message if the write could not be enqueued. If [[Tcp.SimpleWriteCommand#wantsAck]]
+ * returns true, the connection actor will reply with the supplied [[Tcp.SimpleWriteCommand#ack]]
* token once the write has been successfully enqueued to the O/S kernel.
* Note that this does not in any way guarantee that the data will be
* or have been sent! Unfortunately there is no way to determine whether
@@ -782,12 +782,12 @@ object TcpMessage {
/**
* Sending this command to the connection actor will disable reading from the TCP
* socket. TCP flow-control will then propagate backpressure to the sender side
- * as buffers fill up on either end. To re-enable reading send [[Tcp.ResumeReading]].
+ * as buffers fill up on either end. To re-enable reading send `Tcp.ResumeReading`.
*/
def suspendReading: Command = SuspendReading
/**
- * This command needs to be sent to the connection actor after a [[Tcp.SuspendReading]]
+ * This command needs to be sent to the connection actor after a `Tcp.SuspendReading`
* command in order to resume reading from the socket.
*/
def resumeReading: Command = ResumeReading
diff --git a/akka-actor/src/main/scala/akka/io/Udp.scala b/akka-actor/src/main/scala/akka/io/Udp.scala
index 61bddd2608..53125c158b 100644
--- a/akka-actor/src/main/scala/akka/io/Udp.scala
+++ b/akka-actor/src/main/scala/akka/io/Udp.scala
@@ -120,13 +120,13 @@ object Udp extends ExtensionId[UdpExt] with ExtensionIdProvider {
* Send this message to a listener actor (which sent a [[Bound]] message) to
* have it stop reading datagrams from the network. If the O/S kernel’s receive
* buffer runs full then subsequent datagrams will be silently discarded.
- * Re-enable reading from the socket using the [[ResumeReading]] command.
+ * Re-enable reading from the socket using the `ResumeReading` command.
*/
case object SuspendReading extends Command
/**
* This message must be sent to the listener actor to re-enable reading from
- * the socket after a [[SuspendReading]] command.
+ * the socket after a `SuspendReading` command.
*/
case object ResumeReading extends Command
@@ -161,7 +161,7 @@ object Udp extends ExtensionId[UdpExt] with ExtensionIdProvider {
case object SimpleSenderReady extends SimpleSenderReady
/**
- * This message is sent by the listener actor in response to an [[Unbind]] command
+ * This message is sent by the listener actor in response to an `Unbind` command
* after the socket has been closed.
*/
sealed trait Unbound
@@ -312,13 +312,13 @@ object UdpMessage {
* Send this message to a listener actor (which sent a [[Udp.Bound]] message) to
* have it stop reading datagrams from the network. If the O/S kernel’s receive
* buffer runs full then subsequent datagrams will be silently discarded.
- * Re-enable reading from the socket using the [[Udp.ResumeReading]] command.
+ * Re-enable reading from the socket using the `Udp.ResumeReading` command.
*/
def suspendReading: Command = SuspendReading
/**
* This message must be sent to the listener actor to re-enable reading from
- * the socket after a [[Udp.SuspendReading]] command.
+ * the socket after a `Udp.SuspendReading` command.
*/
def resumeReading: Command = ResumeReading
}
diff --git a/akka-actor/src/main/scala/akka/io/UdpConnected.scala b/akka-actor/src/main/scala/akka/io/UdpConnected.scala
index d92da0e45d..f643b5c231 100644
--- a/akka-actor/src/main/scala/akka/io/UdpConnected.scala
+++ b/akka-actor/src/main/scala/akka/io/UdpConnected.scala
@@ -100,13 +100,13 @@ object UdpConnected extends ExtensionId[UdpConnectedExt] with ExtensionIdProvide
* Send this message to a listener actor (which sent a [[Udp.Bound]] message) to
* have it stop reading datagrams from the network. If the O/S kernel’s receive
* buffer runs full then subsequent datagrams will be silently discarded.
- * Re-enable reading from the socket using the [[ResumeReading]] command.
+ * Re-enable reading from the socket using the `ResumeReading` command.
*/
case object SuspendReading extends Command
/**
* This message must be sent to the listener actor to re-enable reading from
- * the socket after a [[SuspendReading]] command.
+ * the socket after a `SuspendReading` command.
*/
case object ResumeReading extends Command
@@ -137,7 +137,7 @@ object UdpConnected extends ExtensionId[UdpConnectedExt] with ExtensionIdProvide
/**
* This message is sent by the connection actor to the actor which sent the
- * [[Disconnect]] message when the UDP socket has been closed.
+ * `Disconnect` message when the UDP socket has been closed.
*/
sealed trait Disconnected extends Event
case object Disconnected extends Disconnected
@@ -231,13 +231,13 @@ object UdpConnectedMessage {
* Send this message to a listener actor (which sent a [[Udp.Bound]] message) to
* have it stop reading datagrams from the network. If the O/S kernel’s receive
* buffer runs full then subsequent datagrams will be silently discarded.
- * Re-enable reading from the socket using the [[UdpConnected.ResumeReading]] command.
+ * Re-enable reading from the socket using the `UdpConnected.ResumeReading` command.
*/
def suspendReading: Command = SuspendReading
/**
* This message must be sent to the listener actor to re-enable reading from
- * the socket after a [[UdpConnected.SuspendReading]] command.
+ * the socket after a `UdpConnected.SuspendReading` command.
*/
def resumeReading: Command = ResumeReading
diff --git a/akka-actor/src/main/scala/akka/pattern/CircuitBreaker.scala b/akka-actor/src/main/scala/akka/pattern/CircuitBreaker.scala
index 556e678882..d6c02f2fee 100644
--- a/akka-actor/src/main/scala/akka/pattern/CircuitBreaker.scala
+++ b/akka-actor/src/main/scala/akka/pattern/CircuitBreaker.scala
@@ -108,9 +108,8 @@ class CircuitBreaker(scheduler: Scheduler, maxFailures: Int, callTimeout: Finite
* Wraps invocations of asynchronous calls that need to be protected
*
* @param body Call needing protected
- * @tparam T return type from call
* @return [[scala.concurrent.Future]] containing the call result or a
- * [[scala.concurrent.TimeoutException]] if the call timed out
+ * `scala.concurrent.TimeoutException` if the call timed out
*
*/
def withCircuitBreaker[T](body: ⇒ Future[T]): Future[T] = currentState.invoke(body)
@@ -119,9 +118,8 @@ class CircuitBreaker(scheduler: Scheduler, maxFailures: Int, callTimeout: Finite
* Java API for [[#withCircuitBreaker]]
*
* @param body Call needing protected
- * @tparam T return type from call
* @return [[scala.concurrent.Future]] containing the call result or a
- * [[scala.concurrent.TimeoutException]] if the call timed out
+ * `scala.concurrent.TimeoutException` if the call timed out
*/
def callWithCircuitBreaker[T](body: Callable[Future[T]]): Future[T] = withCircuitBreaker(body.call)
@@ -129,12 +127,12 @@ class CircuitBreaker(scheduler: Scheduler, maxFailures: Int, callTimeout: Finite
* Wraps invocations of synchronous calls that need to be protected
*
* Calls are run in caller's thread. Because of the synchronous nature of
- * this call the [[scala.concurrent.TimeoutException]] will only be thrown
+ * this call the `scala.concurrent.TimeoutException` will only be thrown
* after the body has completed.
*
+ * Throws java.util.concurrent.TimeoutException if the call timed out.
+ *
* @param body Call needing protected
- * @tparam T return type from call
- * @throws scala.concurrent.TimeoutException if the call timed out
* @return The result of the call
*/
def withSyncCircuitBreaker[T](body: ⇒ T): T =
@@ -143,11 +141,9 @@ class CircuitBreaker(scheduler: Scheduler, maxFailures: Int, callTimeout: Finite
callTimeout)
/**
- * Java API for [[#withSyncCircuitBreaker]]
+ * Java API for [[#withSyncCircuitBreaker]]. Throws [[java.util.concurrent.TimeoutException]] if the call timed out.
*
* @param body Call needing protected
- * @tparam T return type from call
- * @throws scala.concurrent.TimeoutException if the call timed out
* @return The result of the call
*/
def callWithSyncCircuitBreaker[T](body: Callable[T]): T = withSyncCircuitBreaker(body.call)
@@ -223,11 +219,10 @@ class CircuitBreaker(scheduler: Scheduler, maxFailures: Int, callTimeout: Finite
private[akka] def currentFailureCount: Int = Closed.get
/**
- * Implements consistent transition between states
+ * Implements consistent transition between states. Throws IllegalStateException if an invalid transition is attempted.
*
* @param fromState State being transitioning from
* @param toState State being transitioning from
- * @throws IllegalStateException if an invalid transition is attempted
*/
private def transition(fromState: State, toState: State): Unit =
if (swapState(fromState, toState))
@@ -254,6 +249,8 @@ class CircuitBreaker(scheduler: Scheduler, maxFailures: Int, callTimeout: Finite
*/
private def attemptReset(): Unit = transition(Open, HalfOpen)
+ private val timeoutFuture = Future.failed(new TimeoutException("Circuit Breaker Timed out.") with NoStackTrace)
+
/**
* Internal state abstraction
*/
@@ -264,7 +261,6 @@ class CircuitBreaker(scheduler: Scheduler, maxFailures: Int, callTimeout: Finite
* Add a listener function which is invoked on state entry
*
* @param listener listener implementation
- * @tparam T return type of listener, not used - but supplied for type inference purposes
*/
def addListener(listener: Runnable): Unit = listeners add listener
@@ -295,7 +291,6 @@ class CircuitBreaker(scheduler: Scheduler, maxFailures: Int, callTimeout: Finite
* call timeout is counted as a failed call, otherwise a successful call
*
* @param body Implementation of the call
- * @tparam T Return type of the call's implementation
* @return Future containing the result of the call
*/
def callThrough[T](body: ⇒ Future[T]): Future[T] = {
@@ -308,14 +303,13 @@ class CircuitBreaker(scheduler: Scheduler, maxFailures: Int, callTimeout: Finite
val p = Promise[T]()
implicit val ec = sameThreadExecutionContext
- p.future.onComplete({
+ p.future.onComplete {
case s: Success[_] ⇒ callSucceeds()
case _ ⇒ callFails()
- })
+ }
val timeout = scheduler.scheduleOnce(callTimeout) {
- p.tryCompleteWith(
- Future.failed(new TimeoutException("Circuit Breaker Timed out.")))
+ p tryCompleteWith timeoutFuture
}
materialize(body).onComplete { result ⇒
@@ -330,7 +324,6 @@ class CircuitBreaker(scheduler: Scheduler, maxFailures: Int, callTimeout: Finite
* Abstract entry point for all states
*
* @param body Implementation of the call that needs protected
- * @tparam T Return type of protected call
* @return Future containing result of protected call
*/
def invoke[T](body: ⇒ Future[T]): Future[T]
@@ -373,7 +366,6 @@ class CircuitBreaker(scheduler: Scheduler, maxFailures: Int, callTimeout: Finite
* Implementation of invoke, which simply attempts the call
*
* @param body Implementation of the call that needs protected
- * @tparam T Return type of protected call
* @return Future containing result of protected call
*/
override def invoke[T](body: ⇒ Future[T]): Future[T] = callThrough(body)
@@ -418,7 +410,6 @@ class CircuitBreaker(scheduler: Scheduler, maxFailures: Int, callTimeout: Finite
* If the call succeeds the breaker closes.
*
* @param body Implementation of the call that needs protected
- * @tparam T Return type of protected call
* @return Future containing result of protected call
*/
override def invoke[T](body: ⇒ Future[T]): Future[T] =
@@ -462,7 +453,6 @@ class CircuitBreaker(scheduler: Scheduler, maxFailures: Int, callTimeout: Finite
* Fail-fast on any invocation
*
* @param body Implementation of the call that needs protected
- * @tparam T Return type of protected call
* @return Future containing result of protected call
*/
override def invoke[T](body: ⇒ Future[T]): Future[T] =
diff --git a/akka-actor/src/main/scala/akka/pattern/FutureTimeoutSupport.scala b/akka-actor/src/main/scala/akka/pattern/FutureTimeoutSupport.scala
index e80a15e34a..fbec79d4b2 100644
--- a/akka-actor/src/main/scala/akka/pattern/FutureTimeoutSupport.scala
+++ b/akka-actor/src/main/scala/akka/pattern/FutureTimeoutSupport.scala
@@ -1,8 +1,7 @@
-package akka.pattern
-
/**
* Copyright (C) 2009-2015 Typesafe Inc. T or when the header cannot be found,
* the exception is returned in a [[scala.util.Failure]].
*
- *
* The CamelContext is accessible in a [[akka.camel.javaapi.UntypedConsumerActor]] and [[akka.camel.javaapi.UntypedProducerActor]]
* using the `getCamelContext` method, and is available on the [[akka.camel.CamelExtension]].
*
@@ -60,7 +59,7 @@ case class CamelMessage(body: Any, headers: Map[String, Any]) {
/**
* Java API: Returns the header by given name parameter. The header is converted to type T as defined by the clazz parameter.
* An exception is thrown when the conversion to the type T fails or when the header cannot be found.
- *
+ *
* The CamelContext is accessible in a [[akka.camel.javaapi.UntypedConsumerActor]] and [[akka.camel.javaapi.UntypedProducerActor]]
* using the `getCamelContext` method, and is available on the [[akka.camel.CamelExtension]].
*/
@@ -133,7 +132,6 @@ case class CamelMessage(body: Any, headers: Map[String, Any]) {
/**
* Companion object of CamelMessage class.
- *
*/
object CamelMessage {
@@ -176,7 +174,6 @@ object CamelMessage {
/**
* Positive acknowledgement message (used for application-acknowledged message receipts).
* When `autoAck` is set to false in the [[akka.camel.Consumer]], you can send an `Ack` to the sender of the CamelMessage.
- *
*/
case object Ack {
/** Java API to get the Ack singleton */
diff --git a/akka-camel/src/main/scala/akka/camel/Consumer.scala b/akka-camel/src/main/scala/akka/camel/Consumer.scala
index 4c6807dfb5..63236f4224 100644
--- a/akka-camel/src/main/scala/akka/camel/Consumer.scala
+++ b/akka-camel/src/main/scala/akka/camel/Consumer.scala
@@ -14,8 +14,6 @@ import scala.language.existentials
/**
* Mixed in by Actor implementations that consume message from Camel endpoints.
- *
- *
*/
trait Consumer extends Actor with CamelSupport {
import Consumer._
diff --git a/akka-camel/src/main/scala/akka/camel/Producer.scala b/akka-camel/src/main/scala/akka/camel/Producer.scala
index c640b1a53c..68740ae349 100644
--- a/akka-camel/src/main/scala/akka/camel/Producer.scala
+++ b/akka-camel/src/main/scala/akka/camel/Producer.scala
@@ -54,7 +54,7 @@ trait ProducerSupport extends Actor with CamelSupport {
* actually sent it is pre-processed by calling transformOutgoingMessage. If oneway
* is true, an in-only message exchange is initiated, otherwise an in-out message exchange.
*
- * @see Producer#produce(Any, ExchangePattern)
+ * @see Producer#produce
*/
protected def produce: Receive = {
case CamelProducerObjects(endpoint, processor) ⇒
@@ -119,7 +119,7 @@ trait ProducerSupport extends Actor with CamelSupport {
* response is received asynchronously, the receiveAfterProduce is called
* asynchronously. The original sender is preserved.
*
- * @see CamelMessage#canonicalize(Any)
+ * @see CamelMessage#canonicalize
* @param endpoint the endpoint
* @param processor the processor
* @param msg message to produce
diff --git a/akka-camel/src/main/scala/akka/camel/internal/ActivationMessage.scala b/akka-camel/src/main/scala/akka/camel/internal/ActivationMessage.scala
index 5e6caae747..8113379df5 100644
--- a/akka-camel/src/main/scala/akka/camel/internal/ActivationMessage.scala
+++ b/akka-camel/src/main/scala/akka/camel/internal/ActivationMessage.scala
@@ -1,8 +1,7 @@
-package akka.camel.internal
-
/**
* Copyright (C) 2009-2015 Typesafe Inc. [actorPath]?[options]%s,
* where [actorPath] refers to the actor path to the actor.
- *
- *
*/
private[camel] class ActorEndpoint(uri: String,
comp: ActorComponent,
@@ -87,7 +83,6 @@ private[camel] class ActorEndpoint(uri: String,
/**
* INTERNAL API
* Configures the `ActorEndpoint`. This needs to be a `bean` for Camel purposes.
- *
*/
private[camel] trait ActorEndpointConfig {
def path: ActorEndpointPath
@@ -99,12 +94,10 @@ private[camel] trait ActorEndpointConfig {
}
/**
- * Sends the in-message of an exchange to an untyped actor, identified by an [[akka.camel.internal.component.ActorEndPoint]]
- *
- * @see akka.camel.component.ActorComponent
- * @see akka.camel.component.ActorEndpoint
- *
+ * Sends the in-message of an exchange to an untyped actor, identified by an [[akka.camel.internal.component.ActorEndpoint]]
*
+ * @see akka.camel.internal.component.ActorComponent
+ * @see akka.camel.internal.component.ActorEndpoint
*/
private[camel] class ActorProducer(val endpoint: ActorEndpoint, camel: Camel) extends DefaultProducer(endpoint) with AsyncProcessor {
/**
diff --git a/akka-camel/src/main/scala/akka/camel/javaapi/UntypedProducerActor.scala b/akka-camel/src/main/scala/akka/camel/javaapi/UntypedProducerActor.scala
index de95337f29..1ea80f1f3e 100644
--- a/akka-camel/src/main/scala/akka/camel/javaapi/UntypedProducerActor.scala
+++ b/akka-camel/src/main/scala/akka/camel/javaapi/UntypedProducerActor.scala
@@ -11,8 +11,6 @@ import org.apache.camel.impl.DefaultCamelContext
/**
* Subclass this abstract class to create an untyped producer actor. This class is meant to be used from Java.
- *
- *
*/
abstract class UntypedProducerActor extends UntypedActor with ProducerSupport {
/**
diff --git a/akka-cluster-metrics/src/main/scala/akka/cluster/metrics/ClusterMetricsCollector.scala b/akka-cluster-metrics/src/main/scala/akka/cluster/metrics/ClusterMetricsCollector.scala
index d655d9735e..7e041c3255 100644
--- a/akka-cluster-metrics/src/main/scala/akka/cluster/metrics/ClusterMetricsCollector.scala
+++ b/akka-cluster-metrics/src/main/scala/akka/cluster/metrics/ClusterMetricsCollector.scala
@@ -202,7 +202,7 @@ private[metrics] class ClusterMetricsCollector extends Actor with ActorLogging {
}
/**
- * Updates the initial node ring for those nodes that are [[akka.cluster.MemberStatus.Up]].
+ * Updates the initial node ring for those nodes that are [[akka.cluster.MemberStatus]] `Up`.
*/
def receiveState(state: CurrentClusterState): Unit =
nodes = (state.members -- state.unreachable) collect { case m if m.status == MemberStatus.Up ⇒ m.address }
diff --git a/akka-cluster-metrics/src/main/scala/akka/cluster/metrics/ClusterMetricsStrategy.scala b/akka-cluster-metrics/src/main/scala/akka/cluster/metrics/ClusterMetricsStrategy.scala
index 9ca51114f9..0c39dcc136 100644
--- a/akka-cluster-metrics/src/main/scala/akka/cluster/metrics/ClusterMetricsStrategy.scala
+++ b/akka-cluster-metrics/src/main/scala/akka/cluster/metrics/ClusterMetricsStrategy.scala
@@ -10,7 +10,7 @@ import akka.util.Helpers.ConfigOps
/**
* Default [[ClusterMetricsSupervisor]] strategy:
- * A configurable [[OneForOneStrategy]] with restart-on-throwable decider.
+ * A configurable [[akka.actor.OneForOneStrategy]] with restart-on-throwable decider.
*/
class ClusterMetricsStrategy(config: Config) extends OneForOneStrategy(
maxNrOfRetries = config.getInt("maxNrOfRetries"),
@@ -25,7 +25,7 @@ object ClusterMetricsStrategy {
import akka.actor.SupervisorStrategy._
/**
- * [[SupervisorStrategy.Decider]] which allows to survive intermittent Sigar native method calls failures.
+ * [[akka.actor.SupervisorStrategy]] `Decider` which allows to survive intermittent Sigar native method calls failures.
*/
val metricsDecider: SupervisorStrategy.Decider = {
case _: ActorInitializationException ⇒ Stop
diff --git a/akka-cluster-metrics/src/main/scala/akka/cluster/metrics/EWMA.scala b/akka-cluster-metrics/src/main/scala/akka/cluster/metrics/EWMA.scala
index 2ea389ad11..396b48cc36 100644
--- a/akka-cluster-metrics/src/main/scala/akka/cluster/metrics/EWMA.scala
+++ b/akka-cluster-metrics/src/main/scala/akka/cluster/metrics/EWMA.scala
@@ -33,7 +33,7 @@ final case class EWMA(value: Double, alpha: Double) {
* Calculates the exponentially weighted moving average for a given monitored data set.
*
* @param xn the new data point
- * @return a new [[akka.cluster.EWMA]] with the updated value
+ * @return a new EWMA with the updated value
*/
def :+(xn: Double): EWMA = {
val newValue = (alpha * xn) + (1 - alpha) * value
diff --git a/akka-cluster-metrics/src/main/scala/akka/cluster/metrics/Metric.scala b/akka-cluster-metrics/src/main/scala/akka/cluster/metrics/Metric.scala
index c01a4e0498..7c2a1e3457 100644
--- a/akka-cluster-metrics/src/main/scala/akka/cluster/metrics/Metric.scala
+++ b/akka-cluster-metrics/src/main/scala/akka/cluster/metrics/Metric.scala
@@ -109,7 +109,7 @@ object StandardMetrics {
final val SystemLoadAverage = "system-load-average"
final val Processors = "processors"
// In latest Linux kernels: CpuCombined + CpuStolen + CpuIdle = 1.0 or 100%.
- /** Sum of User + Sys + Nice + Wait. See [[org.hyperic.sigar.CpuPerc]] */
+ /** Sum of User + Sys + Nice + Wait. See `org.hyperic.sigar.CpuPerc` */
final val CpuCombined = "cpu-combined"
/** The amount of CPU 'stolen' from this virtual machine by the hypervisor for other tasks (such as running another virtual machine). */
final val CpuStolen = "cpu-stolen"
@@ -146,9 +146,9 @@ object StandardMetrics {
}
/**
- * The amount of used and committed memory will always be <= max if max is defined.
- * A memory allocation may fail if it attempts to increase the used memory such that used > committed
- * even if used <= max is true (e.g. when the system virtual memory is low).
+ * The amount of used and committed memory will always be <= max if max is defined.
+ * A memory allocation may fail if it attempts to increase the used memory such that used > committed
+ * even if used <= max is true (e.g. when the system virtual memory is low).
*
* @param address [[akka.actor.Address]] of the node the metrics are gathered at
* @param timestamp the time of sampling, in milliseconds since midnight, January 1, 1970 UTC
@@ -269,7 +269,7 @@ private[metrics] trait MetricNumericConverter {
*
* @param address [[akka.actor.Address]] of the node the metrics are gathered at
* @param timestamp the time of sampling, in milliseconds since midnight, January 1, 1970 UTC
- * @param metrics the set of sampled [[akka.actor.Metric]]
+ * @param metrics the set of sampled [[akka.cluster.metrics.Metric]]
*/
@SerialVersionUID(1L)
final case class NodeMetrics(address: Address, timestamp: Long, metrics: Set[Metric] = Set.empty[Metric]) {
@@ -344,7 +344,7 @@ private[metrics] object MetricsGossip {
private[metrics] final case class MetricsGossip(nodes: Set[NodeMetrics]) {
/**
- * Removes nodes if their correlating node ring members are not [[akka.cluster.MemberStatus.Up]]
+ * Removes nodes if their correlating node ring members are not [[akka.cluster.MemberStatus]] `Up`.
*/
def remove(node: Address): MetricsGossip = copy(nodes = nodes filterNot (_.address == node))
diff --git a/akka-cluster-metrics/src/main/scala/akka/cluster/metrics/MetricsCollector.scala b/akka-cluster-metrics/src/main/scala/akka/cluster/metrics/MetricsCollector.scala
index ffdc9b0485..e40d2b47fa 100644
--- a/akka-cluster-metrics/src/main/scala/akka/cluster/metrics/MetricsCollector.scala
+++ b/akka-cluster-metrics/src/main/scala/akka/cluster/metrics/MetricsCollector.scala
@@ -237,7 +237,7 @@ class SigarMetricsCollector(address: Address, decayFactor: Double, sigar: SigarP
* theoretically. Note that 99% CPU utilization can be optimal or indicative of failure.
*
* In the data stream, this will sometimes return with a valid metric value, and sometimes as a NaN or Infinite.
- * Documented bug https://bugzilla.redhat.com/show_bug.cgi?id=749121 and several others.
+ * Documented bug 749121 and several others.
*
* Creates a new instance each time.
*/
@@ -248,8 +248,8 @@ class SigarMetricsCollector(address: Address, decayFactor: Double, sigar: SigarP
/**
* (SIGAR) Returns the stolen CPU time. Relevant to virtual hosting environments.
- * For details please see: [[http://en.wikipedia.org/wiki/CPU_time#Subdivision Wikipedia - CPU time subdivision]] and
- * [[https://www.datadoghq.com/2013/08/understanding-aws-stolen-cpu-and-how-it-affects-your-apps/ Understanding AWS stolen CPU and how it affects your apps]]
+ * For details please see: Wikipedia - CPU time subdivision and
+ * Understanding AWS stolen CPU and how it affects your apps
*
* Creates a new instance each time.
*/
diff --git a/akka-cluster-metrics/src/main/scala/akka/cluster/metrics/Provision.scala b/akka-cluster-metrics/src/main/scala/akka/cluster/metrics/Provision.scala
index 3bf11d6ff6..19da06186d 100644
--- a/akka-cluster-metrics/src/main/scala/akka/cluster/metrics/Provision.scala
+++ b/akka-cluster-metrics/src/main/scala/akka/cluster/metrics/Provision.scala
@@ -15,14 +15,14 @@ import scala.util.Failure
import scala.util.Try
/**
- * Provide sigar instance as [[SigarProxy]].
+ * Provide sigar instance as `SigarProxy`.
*
* User can provision sigar classes and native library in one of the following ways:
*
- * 1) Use [[https://github.com/kamon-io/sigar-loader Kamon sigar-loader]] as a project dependency for the user project.
+ * 1) Use Kamon sigar-loader as a project dependency for the user project.
* Metrics extension will extract and load sigar library on demand with help of Kamon sigar provisioner.
*
- * 2) Use [[https://github.com/kamon-io/sigar-loader Kamon sigar-loader]] as java agent: `java -javaagent:/path/to/sigar-loader.jar`
+ * 2) Use Kamon sigar-loader as java agent: `java -javaagent:/path/to/sigar-loader.jar`
* Kamon sigar loader agent will extract and load sigar library during JVM start.
*
* 3) Place `sigar.jar` on the `classpath` and sigar native library for the o/s on the `java.library.path`
@@ -79,7 +79,7 @@ object SigarProvider {
/**
* Release underlying sigar proxy resources.
*
- * Note: [[SigarProxy]] is not [[Sigar]] during tests.
+ * Note: `SigarProxy` is not `Sigar` during tests.
*/
def close(sigar: SigarProxy) = {
if (sigar.isInstanceOf[Sigar]) sigar.asInstanceOf[Sigar].close()
@@ -87,7 +87,7 @@ object SigarProvider {
}
/**
- * Provide sigar instance as [[SigarProxy]] with configured location via [[ClusterMetricsSettings]].
+ * Provide sigar instance as `SigarProxy` with configured location via [[ClusterMetricsSettings]].
*/
case class DefaultSigarProvider(settings: ClusterMetricsSettings) extends SigarProvider {
def extractFolder = settings.NativeLibraryExtractFolder
diff --git a/akka-cluster-metrics/src/main/scala/akka/cluster/metrics/protobuf/MessageSerializer.scala b/akka-cluster-metrics/src/main/scala/akka/cluster/metrics/protobuf/MessageSerializer.scala
index 8075c7d53e..9f74c76dba 100644
--- a/akka-cluster-metrics/src/main/scala/akka/cluster/metrics/protobuf/MessageSerializer.scala
+++ b/akka-cluster-metrics/src/main/scala/akka/cluster/metrics/protobuf/MessageSerializer.scala
@@ -19,7 +19,7 @@ import scala.annotation.tailrec
import scala.collection.JavaConverters.{ asJavaIterableConverter, asScalaBufferConverter, setAsJavaSetConverter }
/**
- * Protobuf serializer for [[ClusterMetricsMessage]] types.
+ * Protobuf serializer for [[akka.cluster.metrics.ClusterMetricsMessage]] types.
*/
class MessageSerializer(val system: ExtendedActorSystem) extends BaseSerializer {
diff --git a/akka-cluster-sharding/src/main/scala/akka/cluster/sharding/ClusterSharding.scala b/akka-cluster-sharding/src/main/scala/akka/cluster/sharding/ClusterSharding.scala
index 1544993fbe..66773593f4 100644
--- a/akka-cluster-sharding/src/main/scala/akka/cluster/sharding/ClusterSharding.scala
+++ b/akka-cluster-sharding/src/main/scala/akka/cluster/sharding/ClusterSharding.scala
@@ -1433,7 +1433,7 @@ object ShardCoordinator {
* supposed to discard their known location of the shard, i.e. start buffering
* incoming messages for the shard. They reply with [[BeginHandOffAck]].
* When all have replied the `ShardCoordinator` continues by sending
- * [[HandOff]] to the `ShardRegion` responsible for the shard.
+ * `HandOff` to the `ShardRegion` responsible for the shard.
*/
@SerialVersionUID(1L) final case class BeginHandOff(shard: ShardId) extends CoordinatorMessage
/**
@@ -1441,14 +1441,14 @@ object ShardCoordinator {
*/
@SerialVersionUID(1L) final case class BeginHandOffAck(shard: ShardId) extends CoordinatorCommand
/**
- * When all `ShardRegion` actors have acknoledged the [[BeginHandOff]] the
- * ShardCoordinator` sends this message to the `ShardRegion` responsible for the
+ * When all `ShardRegion` actors have acknoledged the `BeginHandOff` the
+ * `ShardCoordinator` sends this message to the `ShardRegion` responsible for the
* shard. The `ShardRegion` is supposed to stop all entries in that shard and when
* all entries have terminated reply with `ShardStopped` to the `ShardCoordinator`.
*/
@SerialVersionUID(1L) final case class HandOff(shard: ShardId) extends CoordinatorMessage
/**
- * Reply to [[HandOff]] when all entries in the shard have been terminated.
+ * Reply to `HandOff` when all entries in the shard have been terminated.
*/
@SerialVersionUID(1L) final case class ShardStopped(shard: ShardId) extends CoordinatorCommand
@@ -1527,8 +1527,8 @@ object ShardCoordinator {
/**
* INTERNAL API. Rebalancing process is performed by this actor.
- * It sends [[BeginHandOff]] to all `ShardRegion` actors followed by
- * [[HandOff]] to the `ShardRegion` responsible for the shard.
+ * It sends `BeginHandOff` to all `ShardRegion` actors followed by
+ * `HandOff` to the `ShardRegion` responsible for the shard.
* When the handoff is completed it sends [[RebalanceDone]] to its
* parent `ShardCoordinator`. If the process takes longer than the
* `handOffTimeout` it also sends [[RebalanceDone]].
diff --git a/akka-cluster-tools/src/main/scala/akka/cluster/client/ClusterClient.scala b/akka-cluster-tools/src/main/scala/akka/cluster/client/ClusterClient.scala
index 38070f6e00..ba28f6b117 100644
--- a/akka-cluster-tools/src/main/scala/akka/cluster/client/ClusterClient.scala
+++ b/akka-cluster-tools/src/main/scala/akka/cluster/client/ClusterClient.scala
@@ -194,9 +194,9 @@ class ClusterClient(
}
/**
- * Extension that starts [[ClusterReceptionist]] and accompanying [[DistributedPubSubMediator]]
+ * Extension that starts [[ClusterReceptionist]] and accompanying [[akka.cluster.pubsub.DistributedPubSubMediator]]
* with settings defined in config section `akka.cluster.client.receptionist`.
- * The [[DistributedPubSubMediator]] is started by the [[DistributedPubSubExtension]].
+ * The [[akka.cluster.pubsub.DistributedPubSubMediator]] is started by the [[akka.cluster.pubsub.DistributedPubSubExtension]].
*/
object ClusterReceptionistExtension extends ExtensionId[ClusterReceptionistExtension] with ExtensionIdProvider {
override def get(system: ActorSystem): ClusterReceptionistExtension = super.get(system)
@@ -347,12 +347,12 @@ object ClusterReceptionist {
* The receptionist can be started with the [[ClusterReceptionistExtension]] or as an
* ordinary actor (use the factory method [[ClusterReceptionist#props]]).
*
- * The receptionist forwards messages from the client to the associated [[DistributedPubSubMediator]],
+ * The receptionist forwards messages from the client to the associated [[akka.cluster.pubsub.DistributedPubSubMediator]],
* i.e. the client can send messages to any actor in the cluster that is registered in the
* `DistributedPubSubMediator`. Messages from the client are wrapped in
- * [[DistributedPubSubMediator.Send]], [[DistributedPubSubMediator.SendToAll]]
- * or [[DistributedPubSubMediator.Publish]] with the semantics described in
- * [[DistributedPubSubMediator]].
+ * [[akka.cluster.pubsub.DistributedPubSubMediator.Send]], [[akka.cluster.pubsub.DistributedPubSubMediator.SendToAll]]
+ * or [[akka.cluster.pubsub.DistributedPubSubMediator.Publish]] with the semantics described in
+ * [[akka.cluster.pubsub.DistributedPubSubMediator]].
*
* Response messages from the destination actor are tunneled via the receptionist
* to avoid inbound connections from other cluster nodes to the client, i.e.
diff --git a/akka-cluster-tools/src/main/scala/akka/cluster/pubsub/DistributedPubSubMediator.scala b/akka-cluster-tools/src/main/scala/akka/cluster/pubsub/DistributedPubSubMediator.scala
index 1f4354ce67..bf43737429 100644
--- a/akka-cluster-tools/src/main/scala/akka/cluster/pubsub/DistributedPubSubMediator.scala
+++ b/akka-cluster-tools/src/main/scala/akka/cluster/pubsub/DistributedPubSubMediator.scala
@@ -258,12 +258,12 @@ object DistributedPubSubMediator {
}
/**
- * Mediator uses [[Router]] to send messages to multiple destinations, Router in general
- * unwraps messages from [[RouterEnvelope]] and sends the contents to [[Routee]]s.
+ * Mediator uses [[akka.routing.Router]] to send messages to multiple destinations, Router in general
+ * unwraps messages from [[akka.routing.RouterEnvelope]] and sends the contents to [[akka.routing.Routee]]s.
*
* Using mediator services should not have an undesired effect of unwrapping messages
- * out of [[RouterEnvelope]]. For this reason user messages are wrapped in
- * [[MediatorRouterEnvelope]] which will be unwrapped by the [[Router]] leaving original
+ * out of [[akka.routing.RouterEnvelope]]. For this reason user messages are wrapped in
+ * [[MediatorRouterEnvelope]] which will be unwrapped by the [[akka.routing.Router]] leaving original
* user message.
*/
def wrapIfNeeded: Any ⇒ Any = {
diff --git a/akka-cluster-tools/src/multi-jvm/scala/akka/cluster/singleton/ClusterSingletonManagerLeaveSpec.scala b/akka-cluster-tools/src/multi-jvm/scala/akka/cluster/singleton/ClusterSingletonManagerLeaveSpec.scala
index dfad366be4..6312ea0604 100644
--- a/akka-cluster-tools/src/multi-jvm/scala/akka/cluster/singleton/ClusterSingletonManagerLeaveSpec.scala
+++ b/akka-cluster-tools/src/multi-jvm/scala/akka/cluster/singleton/ClusterSingletonManagerLeaveSpec.scala
@@ -106,6 +106,7 @@ class ClusterSingletonManagerLeaveSpec extends MultiNodeSpec(ClusterSingletonMan
within(10.seconds) {
awaitAssert(cluster.state.members.count(m ⇒ m.status == MemberStatus.Up) should be(3))
}
+ enterBarrier("all-up")
runOn(second) {
cluster.leave(node(first).address)
diff --git a/akka-cluster/src/main/scala/akka/cluster/Cluster.scala b/akka-cluster/src/main/scala/akka/cluster/Cluster.scala
index dc56109ae3..a8bf06bbe1 100644
--- a/akka-cluster/src/main/scala/akka/cluster/Cluster.scala
+++ b/akka-cluster/src/main/scala/akka/cluster/Cluster.scala
@@ -50,7 +50,7 @@ object Cluster extends ExtensionId[Cluster] with ExtensionIdProvider {
*
* Each cluster [[Member]] is identified by its [[akka.actor.Address]], and
* the cluster address of this actor system is [[#selfAddress]]. A member also has a status;
- * initially [[MemberStatus.Joining]] followed by [[MemberStatus.Up]].
+ * initially [[MemberStatus]] `Joining` followed by [[MemberStatus]] `Up`.
*/
class Cluster(val system: ExtendedActorSystem) extends Extension {
@@ -212,11 +212,11 @@ class Cluster(val system: ExtendedActorSystem) extends Extension {
* The `to` classes can be [[akka.cluster.ClusterEvent.ClusterDomainEvent]]
* or subclasses.
*
- * If `initialStateMode` is [[ClusterEvent.InitialStateAsEvents]] the events corresponding
+ * If `initialStateMode` is `ClusterEvent.InitialStateAsEvents` the events corresponding
* to the current state will be sent to the subscriber to mimic what you would
* have seen if you were listening to the events when they occurred in the past.
*
- * If `initialStateMode` is [[ClusterEvent.InitialStateAsSnapshot]] a snapshot of
+ * If `initialStateMode` is `ClusterEvent.InitialStateAsSnapshot` a snapshot of
* [[akka.cluster.ClusterEvent.CurrentClusterState]] will be sent to the subscriber as the
* first message.
*
@@ -274,8 +274,8 @@ class Cluster(val system: ExtendedActorSystem) extends Extension {
/**
* Send command to issue state transition to LEAVING for the node specified by 'address'.
- * The member will go through the status changes [[MemberStatus.Leaving]] (not published to
- * subscribers) followed by [[MemberStatus.Exiting]] and finally [[MemberStatus.Removed]].
+ * The member will go through the status changes [[MemberStatus]] `Leaving` (not published to
+ * subscribers) followed by [[MemberStatus]] `Exiting` and finally [[MemberStatus]] `Removed`.
*
* Note that this command can be issued to any member in the cluster, not necessarily the
* one that is leaving. The cluster extension, but not the actor system or JVM, of the
@@ -300,7 +300,7 @@ class Cluster(val system: ExtendedActorSystem) extends Extension {
/**
* The supplied thunk will be run, once, when current cluster member is `Up`.
- * Typically used together with configuration option `akka.cluster.min-nr-of-members'
+ * Typically used together with configuration option `akka.cluster.min-nr-of-members`
* to defer some action, such as starting actors, until the cluster has reached
* a certain size.
*/
@@ -309,7 +309,7 @@ class Cluster(val system: ExtendedActorSystem) extends Extension {
/**
* Java API: The supplied callback will be run, once, when current cluster member is `Up`.
- * Typically used together with configuration option `akka.cluster.min-nr-of-members'
+ * Typically used together with configuration option `akka.cluster.min-nr-of-members`
* to defer some action, such as starting actors, until the cluster has reached
* a certain size.
*/
@@ -344,7 +344,7 @@ class Cluster(val system: ExtendedActorSystem) extends Extension {
* Shuts down all connections to other members, the cluster daemon and the periodic gossip and cleanup tasks.
*
* Should not called by the user. The user can issue a LEAVE command which will tell the node
- * to go through graceful handoff process `LEAVE -> EXITING -> REMOVED -> SHUTDOWN`.
+ * to go through graceful handoff process `LEAVE -> EXITING -> REMOVED -> SHUTDOWN`.
*/
private[cluster] def shutdown(): Unit = {
if (_isTerminated.compareAndSet(false, true)) {
diff --git a/akka-cluster/src/main/scala/akka/cluster/ClusterDaemon.scala b/akka-cluster/src/main/scala/akka/cluster/ClusterDaemon.scala
index f91cf9f972..6d5fa627d8 100644
--- a/akka-cluster/src/main/scala/akka/cluster/ClusterDaemon.scala
+++ b/akka-cluster/src/main/scala/akka/cluster/ClusterDaemon.scala
@@ -87,19 +87,19 @@ private[cluster] object InternalClusterAction {
case object JoinSeedNode
/**
- * @see JoinSeedNode
+ * see JoinSeedNode
*/
@SerialVersionUID(1L)
case object InitJoin extends ClusterMessage
/**
- * @see JoinSeedNode
+ * see JoinSeedNode
*/
@SerialVersionUID(1L)
final case class InitJoinAck(address: Address) extends ClusterMessage
/**
- * @see JoinSeedNode
+ * see JoinSeedNode
*/
@SerialVersionUID(1L)
final case class InitJoinNack(address: Address) extends ClusterMessage
diff --git a/akka-cluster/src/main/scala/akka/cluster/ClusterEvent.scala b/akka-cluster/src/main/scala/akka/cluster/ClusterEvent.scala
index 6e154c3f5a..8332ae8b86 100644
--- a/akka-cluster/src/main/scala/akka/cluster/ClusterEvent.scala
+++ b/akka-cluster/src/main/scala/akka/cluster/ClusterEvent.scala
@@ -129,7 +129,7 @@ object ClusterEvent {
}
/**
- * Member status changed to [[MemberStatus.Exiting]] and will be removed
+ * Member status changed to `MemberStatus.Exiting` and will be removed
* when all members have seen the `Exiting` status.
*/
final case class MemberExited(member: Member) extends MemberEvent {
@@ -178,7 +178,7 @@ object ClusterEvent {
final case object ClusterShuttingDown extends ClusterDomainEvent
/**
- * Java API: get the singleton instance of [[ClusterShuttingDown]] event
+ * Java API: get the singleton instance of `ClusterShuttingDown` event
*/
def getClusterShuttingDownInstance = ClusterShuttingDown
diff --git a/akka-cluster/src/main/scala/akka/cluster/ClusterMetricsCollector.scala b/akka-cluster/src/main/scala/akka/cluster/ClusterMetricsCollector.scala
index 80126a269d..bcf35858b1 100644
--- a/akka-cluster/src/main/scala/akka/cluster/ClusterMetricsCollector.scala
+++ b/akka-cluster/src/main/scala/akka/cluster/ClusterMetricsCollector.scala
@@ -119,7 +119,7 @@ private[cluster] class ClusterMetricsCollector(publisher: ActorRef) extends Acto
}
/**
- * Updates the initial node ring for those nodes that are [[akka.cluster.MemberStatus.Up]].
+ * Updates the initial node ring for those nodes that are [[akka.cluster.MemberStatus]] `Up`.
*/
def receiveState(state: CurrentClusterState): Unit =
nodes = state.members collect { case m if m.status == Up ⇒ m.address }
@@ -128,7 +128,7 @@ private[cluster] class ClusterMetricsCollector(publisher: ActorRef) extends Acto
* Samples the latest metrics for the node, updates metrics statistics in
* [[akka.cluster.MetricsGossip]], and publishes the change to the event bus.
*
- * @see [[akka.cluster.ClusterMetricsCollector.collect( )]]
+ * @see [[akka.cluster.ClusterMetricsCollector#collect]]
*/
def collect(): Unit = {
latestGossip :+= collector.sample()
@@ -189,7 +189,7 @@ private[cluster] object MetricsGossip {
private[cluster] final case class MetricsGossip(nodes: Set[NodeMetrics]) {
/**
- * Removes nodes if their correlating node ring members are not [[akka.cluster.MemberStatus.Up]]
+ * Removes nodes if their correlating node ring members are not [[akka.cluster.MemberStatus]] `Up`.
*/
def remove(node: Address): MetricsGossip = copy(nodes = nodes filterNot (_.address == node))
@@ -386,7 +386,7 @@ object Metric extends MetricNumericConverter {
*
* @param address [[akka.actor.Address]] of the node the metrics are gathered at
* @param timestamp the time of sampling, in milliseconds since midnight, January 1, 1970 UTC
- * @param metrics the set of sampled [[akka.actor.Metric]]
+ * @param metrics the set of sampled [[akka.cluster.Metric]]
*/
@SerialVersionUID(1L)
@deprecated("Superseded by akka.cluster.metrics (in akka-cluster-metrics jar)", "2.4")
diff --git a/akka-contrib/src/main/scala/akka/contrib/pattern/Aggregator.scala b/akka-contrib/src/main/scala/akka/contrib/pattern/Aggregator.scala
index 10696d4fc1..64ed357bdc 100644
--- a/akka-contrib/src/main/scala/akka/contrib/pattern/Aggregator.scala
+++ b/akka-contrib/src/main/scala/akka/contrib/pattern/Aggregator.scala
@@ -88,7 +88,6 @@ object WorkList {
* Singly linked list entry implementation for WorkList.
* @param ref The item reference, None for head entry
* @param permanent If the entry is to be kept after processing
- * @tparam T The type of the item
*/
class Entry[T](val ref: Option[T], val permanent: Boolean) {
var next: Entry[T] = null
@@ -101,7 +100,6 @@ object WorkList {
* The list is not thread safe! However it is expected to be reentrant. This means a processing function can add/remove
* entries from the list while processing. Most important, a processing function can remove its own entry from the list.
* The first remove must return true and any subsequent removes must return false.
- * @tparam T The type of the item
*/
class WorkList[T] {
diff --git a/akka-contrib/src/main/scala/akka/contrib/pattern/ReliableProxy.scala b/akka-contrib/src/main/scala/akka/contrib/pattern/ReliableProxy.scala
index d1ed3b2149..cba19c70b6 100644
--- a/akka-contrib/src/main/scala/akka/contrib/pattern/ReliableProxy.scala
+++ b/akka-contrib/src/main/scala/akka/contrib/pattern/ReliableProxy.scala
@@ -173,22 +173,22 @@ import ReliableProxy._
* transition callbacks to those actors which subscribe using the
* ``SubscribeTransitionCallBack`` and ``UnsubscribeTransitionCallBack``
* messages; see [[akka.actor.FSM]] for more documentation. The proxy will
- * transition into [[ReliableProxy.Active]] state when ACKs
- * are outstanding and return to the [[ReliableProxy.Idle]]
+ * transition into `ReliableProxy.Active` state when ACKs
+ * are outstanding and return to the `ReliableProxy.Idle`
* state when every message send so far has been confirmed by the peer end-point.
*
- * The initial state of the proxy is [[ReliableProxy.Connecting]]. In this state the
+ * The initial state of the proxy is `ReliableProxy.Connecting`. In this state the
* proxy will repeatedly send [[akka.actor.Identify]] messages to `ActorSelection(targetPath)`
* in order to obtain a new `ActorRef` for the target. When an [[akka.actor.ActorIdentity]]
* for the target is received a new tunnel will be created, a [[ReliableProxy.TargetChanged]]
* message containing the target `ActorRef` will be sent to the proxy's transition subscribers
- * and the proxy will transition into either the [[ReliableProxy.Idle]] or [[ReliableProxy.Active]]
+ * and the proxy will transition into either the `ReliableProxy.Idle` or `ReliableProxy.Active`
* state, depending if there are any outstanding messages that need to be delivered. If
* `maxConnectAttempts` is defined this actor will stop itself after `Identify` is sent
* `maxConnectAttempts` times.
*
* While in the `Idle` or `Active` states, if a communication failure causes the tunnel to
- * terminate via Remote Deathwatch the proxy will transition into the [[ReliableProxy.Connecting]]
+ * terminate via Remote Deathwatch the proxy will transition into the `ReliableProxy.Connecting`
* state as described above. After reconnecting `TargetChanged` will be sent only if the target
* `ActorRef` has changed.
*
diff --git a/akka-docs/rst/general/configuration.rst b/akka-docs/rst/general/configuration.rst
index 3009ecceb7..e86f45e40c 100644
--- a/akka-docs/rst/general/configuration.rst
+++ b/akka-docs/rst/general/configuration.rst
@@ -204,7 +204,7 @@ before or after using them to construct an actor system:
.. parsed-literal::
- Welcome to Scala version @scalaVersion@ (Java HotSpot(TM) 64-Bit Server VM, Java 1.6.0_27).
+ Welcome to Scala version @scalaVersion@ (Java HotSpot(TM) 64-Bit Server VM, Java 1.8.0).
Type in expressions to have them evaluated.
Type :help for more information.
diff --git a/akka-docs/rst/intro/getting-started.rst b/akka-docs/rst/intro/getting-started.rst
index 933ff1d316..6f38a94d34 100644
--- a/akka-docs/rst/intro/getting-started.rst
+++ b/akka-docs/rst/intro/getting-started.rst
@@ -4,9 +4,11 @@ Getting Started
Prerequisites
-------------
-Akka requires that you have `Java 1.6