diff --git a/akka-actor/src/main/java/akka/dispatch/AbstractBoundedNodeQueue.java b/akka-actor/src/main/java/akka/dispatch/AbstractBoundedNodeQueue.java index ed506282d9..4ae78fee31 100644 --- a/akka-actor/src/main/java/akka/dispatch/AbstractBoundedNodeQueue.java +++ b/akka-actor/src/main/java/akka/dispatch/AbstractBoundedNodeQueue.java @@ -137,7 +137,7 @@ public abstract class AbstractBoundedNodeQueue { */ public final int size() { //Order of operations is extremely important here - // If no item was dequeued between when we looked at the count of the enqueueing end, + // If no item was dequeued between when we looked at the count of the enqueuing end, // there should be no out-of-bounds for(;;) { final int deqCountBefore = getDeq().count; diff --git a/akka-actor/src/main/scala/akka/actor/ActorRef.scala b/akka-actor/src/main/scala/akka/actor/ActorRef.scala index e22fe6d34e..341071cd9e 100644 --- a/akka-actor/src/main/scala/akka/actor/ActorRef.scala +++ b/akka-actor/src/main/scala/akka/actor/ActorRef.scala @@ -465,7 +465,7 @@ private[akka] trait MinimalActorRef extends InternalActorRef with LocalRef { protected def writeReplace(): AnyRef = SerializedActorRef(this) } -/** Subscribe to this class to be notified about all DeadLetters (also the supressed ones). */ +/** Subscribe to this class to be notified about all DeadLetters (also the suppressed ones). */ sealed trait AllDeadLetters { def message: Any def sender: ActorRef diff --git a/akka-actor/src/main/scala/akka/actor/ActorSystem.scala b/akka-actor/src/main/scala/akka/actor/ActorSystem.scala index 85b37681e5..5df607a02d 100644 --- a/akka-actor/src/main/scala/akka/actor/ActorSystem.scala +++ b/akka-actor/src/main/scala/akka/actor/ActorSystem.scala @@ -133,7 +133,7 @@ object ActorSystem { * falls back to the ClassLoader associated with the ActorSystem class. * If an ExecutionContext is given, it will be used as the default executor inside this ActorSystem. * If no ExecutionContext is given, the system will fallback to the executor configured under "akka.actor.default-dispatcher.default-executor.fallback". - * The system will use the passed in config, or falls back to the deafult reference configuration using the ClassLoader. + * The system will use the passed in config, or falls back to the default reference configuration using the ClassLoader. * * @see The Typesafe Config Library API Documentation */ @@ -828,7 +828,7 @@ private[akka] class ActorSystemImpl( private[this] final val done = Promise[T]() private[this] final val ref = new AtomicReference(done) - // onComplete never fires twice so safe to avoid nullcheck + // onComplete never fires twice so safe to avoid null check upStreamTerminated onComplete { t ⇒ ref.getAndSet(null).complete(t) } /** diff --git a/akka-actor/src/main/scala/akka/actor/FSM.scala b/akka-actor/src/main/scala/akka/actor/FSM.scala index 09e3597ff1..8986aad909 100644 --- a/akka-actor/src/main/scala/akka/actor/FSM.scala +++ b/akka-actor/src/main/scala/akka/actor/FSM.scala @@ -212,7 +212,7 @@ object FSM { final case class Event[D](event: Any, stateData: D) extends NoSerializationVerificationNeeded /** - * Case class representing the state of the [[akka.actor.FSM]] whithin the + * Case class representing the state of the [[akka.actor.FSM]] within the * `onTermination` block. */ final case class StopEvent[S, D](reason: Reason, currentState: S, stateData: D) extends NoSerializationVerificationNeeded diff --git a/akka-actor/src/main/scala/akka/actor/Scheduler.scala b/akka-actor/src/main/scala/akka/actor/Scheduler.scala index 0ea99a562a..61f6fbf22f 100644 --- a/akka-actor/src/main/scala/akka/actor/Scheduler.scala +++ b/akka-actor/src/main/scala/akka/actor/Scheduler.scala @@ -213,7 +213,7 @@ class LightArrayRevolverScheduler(config: Config, try { ((d + TickDuration - oneNs) / TickDuration).toLong * TickDuration } catch { - case _: IllegalArgumentException ⇒ d // rouding up Long.MaxValue.nanos overflows + case _: IllegalArgumentException ⇒ d // rounding up Long.MaxValue.nanos overflows } /** diff --git a/akka-actor/src/main/scala/akka/actor/TypedActor.scala b/akka-actor/src/main/scala/akka/actor/TypedActor.scala index 50198a61aa..b19cdefcdb 100644 --- a/akka-actor/src/main/scala/akka/actor/TypedActor.scala +++ b/akka-actor/src/main/scala/akka/actor/TypedActor.scala @@ -59,7 +59,7 @@ trait TypedActorFactory { } /** - * Returns wether the supplied AnyRef is a TypedActor proxy or not + * Returns whether the supplied AnyRef is a TypedActor proxy or not */ def isTypedActor(proxyOrNot: AnyRef): Boolean @@ -654,7 +654,7 @@ class TypedActorExtension(val system: ExtendedActorSystem) extends TypedActorFac } /** - * Returns wether the supplied AnyRef is a TypedActor proxy or not + * Returns whether the supplied AnyRef is a TypedActor proxy or not */ def isTypedActor(proxyOrNot: AnyRef): Boolean = invocationHandlerFor(proxyOrNot) ne null @@ -675,7 +675,7 @@ class TypedActorExtension(val system: ExtendedActorSystem) extends TypedActorFac proxy } else { proxyVar set proxy // Chicken and egg situation we needed to solve, set the proxy so that we can set the self-reference inside each receive - actorVar set actorRef //Make sure the InvocationHandler gets ahold of the actor reference, this is not a problem since the proxy hasn't escaped this method yet + actorVar set actorRef //Make sure the InvocationHandler gets a hold of the actor reference, this is not a problem since the proxy hasn't escaped this method yet proxyVar.get } } diff --git a/akka-actor/src/main/scala/akka/dispatch/BatchingExecutor.scala b/akka-actor/src/main/scala/akka/dispatch/BatchingExecutor.scala index bd39c73c25..a53f5e7771 100644 --- a/akka-actor/src/main/scala/akka/dispatch/BatchingExecutor.scala +++ b/akka-actor/src/main/scala/akka/dispatch/BatchingExecutor.scala @@ -59,7 +59,7 @@ private[akka] trait BatchingExecutor extends Executor { protected final def resubmitUnbatched(): Boolean = { val current = _tasksLocal.get() _tasksLocal.remove() - if ((current eq this) && !current.isEmpty) { // Resubmit outselves if something bad happened and we still have work to do + if ((current eq this) && !current.isEmpty) { // Resubmit ourselves if something bad happened and we still have work to do unbatchedExecute(current) //TODO what if this submission fails? true } else false diff --git a/akka-actor/src/main/scala/akka/dispatch/Future.scala b/akka-actor/src/main/scala/akka/dispatch/Future.scala index 6af5853aa1..f753967487 100644 --- a/akka-actor/src/main/scala/akka/dispatch/Future.scala +++ b/akka-actor/src/main/scala/akka/dispatch/Future.scala @@ -88,7 +88,7 @@ object Futures { * * The result becomes available once the asynchronous computation is completed. * - * @param body the asychronous computation + * @param body the asynchronous computation * @param executor the execution context on which the future is run * @return the `Future` holding the result of the computation */ diff --git a/akka-actor/src/main/scala/akka/event/EventStream.scala b/akka-actor/src/main/scala/akka/event/EventStream.scala index 2efc240b27..3a7b1069e9 100644 --- a/akka-actor/src/main/scala/akka/event/EventStream.scala +++ b/akka-actor/src/main/scala/akka/event/EventStream.scala @@ -126,7 +126,7 @@ class EventStream(sys: ActorSystem, private val debug: Boolean) extends LoggingB * * The actual check if the subscriber still has subscriptions is performed by the `EventStreamUnsubscriber`, * because it's an expensive operation, and we don want to block client-code for that long, the Actor will eventually - * catch up and perform the apropriate operation. + * catch up and perform the appropriate operation. */ @tailrec private def unregisterIfNoMoreSubscribedChannels(subscriber: ActorRef): Unit = { diff --git a/akka-actor/src/main/scala/akka/event/Logging.scala b/akka-actor/src/main/scala/akka/event/Logging.scala index d77976a0b0..f8c659fc38 100644 --- a/akka-actor/src/main/scala/akka/event/Logging.scala +++ b/akka-actor/src/main/scala/akka/event/Logging.scala @@ -224,7 +224,7 @@ trait LoggingBus extends ActorEventBus { * } * * class MyClass extends MyType { - * val sys = ActorSyste("sys") + * val sys = ActorSystem("sys") * val log = Logging(sys, this) // will use "hallo,akka://sys" as logSource * def name = "hallo" * } diff --git a/akka-actor/src/main/scala/akka/io/SelectionHandler.scala b/akka-actor/src/main/scala/akka/io/SelectionHandler.scala index 0046bebe72..42886eeda6 100644 --- a/akka-actor/src/main/scala/akka/io/SelectionHandler.scala +++ b/akka-actor/src/main/scala/akka/io/SelectionHandler.scala @@ -52,7 +52,7 @@ private[io] trait ChannelRegistry { /** * Implementations of this interface are sent as actor messages back to a channel actor as * a result of it having called `register` on the `ChannelRegistry`. - * Enables a channel actor to directly schedule interest setting tasks to the selector mgmt. dispatcher. + * Enables a channel actor to directly schedule interest setting tasks to the selector management dispatcher. */ private[io] trait ChannelRegistration extends NoSerializationVerificationNeeded { def enableInterest(op: Int) diff --git a/akka-actor/src/main/scala/akka/io/Tcp.scala b/akka-actor/src/main/scala/akka/io/Tcp.scala index cb4f964a40..0d166dbc73 100644 --- a/akka-actor/src/main/scala/akka/io/Tcp.scala +++ b/akka-actor/src/main/scala/akka/io/Tcp.scala @@ -26,7 +26,7 @@ import java.lang.{ Iterable ⇒ JIterable } * In order to open an outbound connection send a [[Tcp.Connect]] message * to the [[TcpExt#manager]]. * - * In order to start listening for inbound connetions send a [[Tcp.Bind]] + * In order to start listening for inbound connections send a [[Tcp.Bind]] * message to the [[TcpExt#manager]]. * * The Java API for generating TCP commands is available at [[TcpMessage]]. diff --git a/akka-actor/src/main/scala/akka/io/TcpManager.scala b/akka-actor/src/main/scala/akka/io/TcpManager.scala index eca97acdb6..d03ff3dae7 100644 --- a/akka-actor/src/main/scala/akka/io/TcpManager.scala +++ b/akka-actor/src/main/scala/akka/io/TcpManager.scala @@ -49,11 +49,11 @@ private[io] class TcpManager(tcp: TcpExt) def receive = workerForCommandHandler { case c: Connect ⇒ - val commander = sender() // cache because we create a function that will run asyncly + val commander = sender() // cache because we create a function that will run asynchly (registry ⇒ Props(classOf[TcpOutgoingConnection], tcp, registry, commander, c)) case b: Bind ⇒ - val commander = sender() // cache because we create a function that will run asyncly + val commander = sender() // cache because we create a function that will run asynchly (registry ⇒ Props(classOf[TcpListener], selectorPool, tcp, registry, commander, b)) } diff --git a/akka-actor/src/main/scala/akka/io/UdpConnectedManager.scala b/akka-actor/src/main/scala/akka/io/UdpConnectedManager.scala index bcef7e4f73..51688af8a9 100644 --- a/akka-actor/src/main/scala/akka/io/UdpConnectedManager.scala +++ b/akka-actor/src/main/scala/akka/io/UdpConnectedManager.scala @@ -14,7 +14,7 @@ private[io] class UdpConnectedManager(udpConn: UdpConnectedExt) def receive = workerForCommandHandler { case c: Connect ⇒ - val commander = sender() // cache because we create a function that will run asyncly + val commander = sender() // cache because we create a function that will run asynchly registry ⇒ Props(classOf[UdpConnection], udpConn, registry, commander, c) } diff --git a/akka-actor/src/main/scala/akka/io/UdpManager.scala b/akka-actor/src/main/scala/akka/io/UdpManager.scala index 1d943a3f05..4bf011f76d 100644 --- a/akka-actor/src/main/scala/akka/io/UdpManager.scala +++ b/akka-actor/src/main/scala/akka/io/UdpManager.scala @@ -47,11 +47,11 @@ private[io] class UdpManager(udp: UdpExt) extends SelectionHandler.SelectorBased def receive = workerForCommandHandler { case b: Bind ⇒ - val commander = sender() // cache because we create a function that will run asyncly + val commander = sender() // cache because we create a function that will run asynchly (registry ⇒ Props(classOf[UdpListener], udp, registry, commander, b)) case SimpleSender(options) ⇒ - val commander = sender() // cache because we create a function that will run asyncly + val commander = sender() // cache because we create a function that will run asynchly (registry ⇒ Props(classOf[UdpSender], udp, registry, commander, options)) } diff --git a/akka-actor/src/main/scala/akka/routing/Balancing.scala b/akka-actor/src/main/scala/akka/routing/Balancing.scala index 3c5f226320..7cbd19c6de 100644 --- a/akka-actor/src/main/scala/akka/routing/Balancing.scala +++ b/akka-actor/src/main/scala/akka/routing/Balancing.scala @@ -126,7 +126,7 @@ final case class BalancingPool( } /** - * Uses the supervisor strategy of the given Routerconfig + * Uses the supervisor strategy of the given RouterConfig * if this RouterConfig doesn't have one. */ override def withFallback(other: RouterConfig): RouterConfig = diff --git a/akka-actor/src/main/scala/akka/routing/Broadcast.scala b/akka-actor/src/main/scala/akka/routing/Broadcast.scala index 7320778dd7..7110ea50cb 100644 --- a/akka-actor/src/main/scala/akka/routing/Broadcast.scala +++ b/akka-actor/src/main/scala/akka/routing/Broadcast.scala @@ -97,7 +97,7 @@ final case class BroadcastPool( def withDispatcher(dispatcherId: String): BroadcastPool = copy(routerDispatcher = dispatcherId) /** - * Uses the resizer and/or the supervisor strategy of the given Routerconfig + * Uses the resizer and/or the supervisor strategy of the given RouterConfig * if this RouterConfig doesn't have one, i.e. the resizer defined in code is used if * resizer was not defined in config. */ diff --git a/akka-actor/src/main/scala/akka/routing/ConsistentHashing.scala b/akka-actor/src/main/scala/akka/routing/ConsistentHashing.scala index 3286384559..89371f67e2 100644 --- a/akka-actor/src/main/scala/akka/routing/ConsistentHashing.scala +++ b/akka-actor/src/main/scala/akka/routing/ConsistentHashing.scala @@ -314,7 +314,7 @@ final case class ConsistentHashingPool( copy(hashMapping = ConsistentHashingRouter.hashMappingAdapter(mapper)) /** - * Uses the resizer and/or the supervisor strategy of the given Routerconfig + * Uses the resizer and/or the supervisor strategy of the given RouterConfig * if this RouterConfig doesn't have one, i.e. the resizer defined in code is used if * resizer was not defined in config. * Uses the the `hashMapping` defined in code, since that can't be defined in configuration. diff --git a/akka-actor/src/main/scala/akka/routing/Random.scala b/akka-actor/src/main/scala/akka/routing/Random.scala index a4c353ce5e..ccd5d08b97 100644 --- a/akka-actor/src/main/scala/akka/routing/Random.scala +++ b/akka-actor/src/main/scala/akka/routing/Random.scala @@ -98,7 +98,7 @@ final case class RandomPool( def withDispatcher(dispatcherId: String): RandomPool = copy(routerDispatcher = dispatcherId) /** - * Uses the resizer and/or the supervisor strategy of the given Routerconfig + * Uses the resizer and/or the supervisor strategy of the given RouterConfig * if this RouterConfig doesn't have one, i.e. the resizer defined in code is used if * resizer was not defined in config. */ diff --git a/akka-actor/src/main/scala/akka/routing/Resizer.scala b/akka-actor/src/main/scala/akka/routing/Resizer.scala index ff7c8f4669..37b2f93127 100644 --- a/akka-actor/src/main/scala/akka/routing/Resizer.scala +++ b/akka-actor/src/main/scala/akka/routing/Resizer.scala @@ -49,7 +49,7 @@ trait Resizer { * returns true and no other resize is in progress. * * Return the number of routees to add or remove. Negative value will remove that number of routees. - * Positive value will add that number of routess. 0 will not change the routees. + * Positive value will add that number of routees. 0 will not change the routees. * * This method is invoked only in the context of the Router actor. */ diff --git a/akka-actor/src/main/scala/akka/routing/RoundRobin.scala b/akka-actor/src/main/scala/akka/routing/RoundRobin.scala index bd1c5effd1..71955ce552 100644 --- a/akka-actor/src/main/scala/akka/routing/RoundRobin.scala +++ b/akka-actor/src/main/scala/akka/routing/RoundRobin.scala @@ -102,7 +102,7 @@ final case class RoundRobinPool( def withDispatcher(dispatcherId: String): RoundRobinPool = copy(routerDispatcher = dispatcherId) /** - * Uses the resizer and/or the supervisor strategy of the given Routerconfig + * Uses the resizer and/or the supervisor strategy of the given RouterConfig * if this RouterConfig doesn't have one, i.e. the resizer defined in code is used if * resizer was not defined in config. */ diff --git a/akka-actor/src/main/scala/akka/routing/RoutedActorCell.scala b/akka-actor/src/main/scala/akka/routing/RoutedActorCell.scala index 88066f3e31..9a424f6c15 100644 --- a/akka-actor/src/main/scala/akka/routing/RoutedActorCell.scala +++ b/akka-actor/src/main/scala/akka/routing/RoutedActorCell.scala @@ -119,7 +119,7 @@ private[akka] class RoutedActorCell( } /** - * Called when `router` is initalized but before `super.start()` to + * Called when `router` is initialized but before `super.start()` to * be able to do extra initialization in subclass. */ protected def preSuperStart(): Unit = () diff --git a/akka-actor/src/main/scala/akka/routing/RouterConfig.scala b/akka-actor/src/main/scala/akka/routing/RouterConfig.scala index f8c6d56921..6a8a4570ee 100644 --- a/akka-actor/src/main/scala/akka/routing/RouterConfig.scala +++ b/akka-actor/src/main/scala/akka/routing/RouterConfig.scala @@ -54,7 +54,7 @@ trait RouterConfig extends Serializable { * Possibility to define an actor for controlling the routing * logic from external stimuli (e.g. monitoring metrics). * This actor will be a child of the router "head" actor. - * Managment messages not handled by the "head" actor are + * Management messages not handled by the "head" actor are * delegated to this controller actor. */ def routingLogicController(routingLogic: RoutingLogic): Option[Props] = None diff --git a/akka-actor/src/main/scala/akka/routing/ScatterGatherFirstCompleted.scala b/akka-actor/src/main/scala/akka/routing/ScatterGatherFirstCompleted.scala index ffe59fb9c1..8ed0d77a58 100644 --- a/akka-actor/src/main/scala/akka/routing/ScatterGatherFirstCompleted.scala +++ b/akka-actor/src/main/scala/akka/routing/ScatterGatherFirstCompleted.scala @@ -136,7 +136,7 @@ final case class ScatterGatherFirstCompletedPool( def withDispatcher(dispatcherId: String): ScatterGatherFirstCompletedPool = copy(routerDispatcher = dispatcherId) /** - * Uses the resizer and/or the supervisor strategy of the given Routerconfig + * Uses the resizer and/or the supervisor strategy of the given RouterConfig * if this RouterConfig doesn't have one, i.e. the resizer defined in code is used if * resizer was not defined in config. */ diff --git a/akka-actor/src/main/scala/akka/routing/SmallestMailbox.scala b/akka-actor/src/main/scala/akka/routing/SmallestMailbox.scala index b8c6b0e337..e212844fb0 100644 --- a/akka-actor/src/main/scala/akka/routing/SmallestMailbox.scala +++ b/akka-actor/src/main/scala/akka/routing/SmallestMailbox.scala @@ -213,7 +213,7 @@ final case class SmallestMailboxPool( def withDispatcher(dispatcherId: String): SmallestMailboxPool = copy(routerDispatcher = dispatcherId) /** - * Uses the resizer and/or the supervisor strategy of the given Routerconfig + * Uses the resizer and/or the supervisor strategy of the given RouterConfig * if this RouterConfig doesn't have one, i.e. the resizer defined in code is used if * resizer was not defined in config. */ diff --git a/akka-actor/src/main/scala/akka/routing/TailChopping.scala b/akka-actor/src/main/scala/akka/routing/TailChopping.scala index 72241c67a1..7cfcb66b9e 100644 --- a/akka-actor/src/main/scala/akka/routing/TailChopping.scala +++ b/akka-actor/src/main/scala/akka/routing/TailChopping.scala @@ -83,7 +83,7 @@ private[akka] final case class TailChoppingRoutees( } /** - * A router poll thats sends the message to a first, random picked, routee, + * A router poll that sends the message to a first, random picked, routee, * then wait a specified `interval` and then send to a second, random picked, and so on till one full cycle.. * * The configuration parameter trumps the constructor arguments. This means that @@ -169,7 +169,7 @@ final case class TailChoppingPool( def withDispatcher(dispatcherId: String): TailChoppingPool = copy(routerDispatcher = dispatcherId) /** - * Uses the resizer and/or the supervisor strategy of the given Routerconfig + * Uses the resizer and/or the supervisor strategy of the given RouterConfig * if this RouterConfig doesn't have one, i.e. the resizer defined in code is used if * resizer was not defined in config. */ diff --git a/akka-actor/src/main/scala/akka/util/LockUtil.scala b/akka-actor/src/main/scala/akka/util/LockUtil.scala index 05a624622c..c09ffe4392 100644 --- a/akka-actor/src/main/scala/akka/util/LockUtil.scala +++ b/akka-actor/src/main/scala/akka/util/LockUtil.scala @@ -138,7 +138,7 @@ class Switch(startAsOn: Boolean = false) { def isOn: Boolean = switch.get /** - * Returns whether the switch is IMMEDDIATELY off (no locking) + * Returns whether the switch is IMMEDIATELY off (no locking) */ def isOff: Boolean = !isOn } diff --git a/akka-camel/src/main/scala/akka/camel/Camel.scala b/akka-camel/src/main/scala/akka/camel/Camel.scala index dd482cc081..6804105dc9 100644 --- a/akka-camel/src/main/scala/akka/camel/Camel.scala +++ b/akka-camel/src/main/scala/akka/camel/Camel.scala @@ -103,7 +103,7 @@ class CamelSettings private[camel] (config: Config, dynamicAccess: DynamicAccess (s: String, r: RouteDefinition) ⇒ conversions.get(s).fold(r)(r.convertBodyTo) } /** - * Configured setting, determine the class used to load/retrive the instance of the Camel Context + * Configured setting, determine the class used to load/retrieve the instance of the Camel Context */ final val ContextProvider: ContextProvider = { val fqcn = config.getString("akka.camel.context-provider") diff --git a/akka-cluster-metrics/src/main/scala/akka/cluster/metrics/ClusterMetricsRouting.scala b/akka-cluster-metrics/src/main/scala/akka/cluster/metrics/ClusterMetricsRouting.scala index 152550409b..33be051515 100644 --- a/akka-cluster-metrics/src/main/scala/akka/cluster/metrics/ClusterMetricsRouting.scala +++ b/akka-cluster-metrics/src/main/scala/akka/cluster/metrics/ClusterMetricsRouting.scala @@ -163,7 +163,7 @@ final case class AdaptiveLoadBalancingPool( def withDispatcher(dispatcherId: String): AdaptiveLoadBalancingPool = copy(routerDispatcher = dispatcherId) /** - * Uses the supervisor strategy of the given Routerconfig + * Uses the supervisor strategy of the given RouterConfig * if this RouterConfig doesn't have one */ override def withFallback(other: RouterConfig): RouterConfig = diff --git a/akka-cluster-metrics/src/main/scala/akka/cluster/metrics/MetricsCollector.scala b/akka-cluster-metrics/src/main/scala/akka/cluster/metrics/MetricsCollector.scala index e40d2b47fa..ee92a26b2e 100644 --- a/akka-cluster-metrics/src/main/scala/akka/cluster/metrics/MetricsCollector.scala +++ b/akka-cluster-metrics/src/main/scala/akka/cluster/metrics/MetricsCollector.scala @@ -81,7 +81,7 @@ private[metrics] object MetricsCollector { * Loads JVM and system metrics through JMX monitoring beans. * * @param address The [[akka.actor.Address]] of the node being sampled - * @param decay how quickly the exponential weighting of past data is decayed + * @param decayFactor how quickly the exponential weighting of past data is decayed */ class JmxMetricsCollector(address: Address, decayFactor: Double) extends MetricsCollector { import StandardMetrics._ @@ -184,7 +184,7 @@ class JmxMetricsCollector(address: Address, decayFactor: Double) extends Metrics * to missing classes or native libraries. * * @param address The [[akka.actor.Address]] of the node being sampled - * @param decay how quickly the exponential weighting of past data is decayed + * @param decayFactor how quickly the exponential weighting of past data is decayed * @param sigar the org.hyperic.Sigar instance */ class SigarMetricsCollector(address: Address, decayFactor: Double, sigar: SigarProxy) diff --git a/akka-cluster-sharding/src/main/scala/akka/cluster/sharding/ClusterSharding.scala b/akka-cluster-sharding/src/main/scala/akka/cluster/sharding/ClusterSharding.scala index ce01d33609..06768b62ed 100644 --- a/akka-cluster-sharding/src/main/scala/akka/cluster/sharding/ClusterSharding.scala +++ b/akka-cluster-sharding/src/main/scala/akka/cluster/sharding/ClusterSharding.scala @@ -399,7 +399,7 @@ private[akka] object ClusterShardingGuardian { } /** - * INTERNAL API. [[ShardRegion]] and [[ShardCoordinator]] actors are createad as children + * INTERNAL API. [[ShardRegion]] and [[ShardCoordinator]] actors are created as children * of this actor. */ private[akka] class ClusterShardingGuardian extends Actor { @@ -1467,7 +1467,7 @@ object ShardCoordinator { */ @SerialVersionUID(1L) final case class RegisterProxy(shardRegionProxy: ActorRef) extends CoordinatorCommand /** - * Acknowledgement from `ShardCoordinator` that [[Register]] or [[RegisterProxy]] was sucessful. + * Acknowledgement from `ShardCoordinator` that [[Register]] or [[RegisterProxy]] was successful. */ @SerialVersionUID(1L) final case class RegisterAck(coordinator: ActorRef) extends CoordinatorMessage /** @@ -1484,7 +1484,7 @@ object ShardCoordinator { */ @SerialVersionUID(1L) final case class ShardHome(shard: ShardId, ref: ActorRef) extends CoordinatorMessage /** - * `ShardCoodinator` informs a `ShardRegion` that it is hosting this shard + * `ShardCoordinator` informs a `ShardRegion` that it is hosting this shard */ @SerialVersionUID(1L) final case class HostShard(shard: ShardId) extends CoordinatorMessage /** @@ -1505,7 +1505,7 @@ object ShardCoordinator { */ @SerialVersionUID(1L) final case class BeginHandOffAck(shard: ShardId) extends CoordinatorCommand /** - * When all `ShardRegion` actors have acknoledged the `BeginHandOff` the + * When all `ShardRegion` actors have acknowledged the `BeginHandOff` the * `ShardCoordinator` sends this message to the `ShardRegion` responsible for the * shard. The `ShardRegion` is supposed to stop all entries in that shard and when * all entries have terminated reply with `ShardStopped` to the `ShardCoordinator`. @@ -1872,7 +1872,7 @@ class ShardCoordinator(handOffTimeout: FiniteDuration, shardStartTimeout: Finite case ShardHome(_, _) ⇒ //On rebalance, we send ourselves a GetShardHome message to reallocate a - // shard. This recieve handles the "response" from that message. i.e. Ingores it. + // shard. This receive handles the "response" from that message. i.e. ignores it. case ClusterShuttingDown ⇒ log.debug("Shutting down ShardCoordinator") diff --git a/akka-cluster-sharding/src/multi-jvm/scala/akka/cluster/sharding/ClusterShardingSpec.scala b/akka-cluster-sharding/src/multi-jvm/scala/akka/cluster/sharding/ClusterShardingSpec.scala index ba9748aad2..b97eec6898 100644 --- a/akka-cluster-sharding/src/multi-jvm/scala/akka/cluster/sharding/ClusterShardingSpec.scala +++ b/akka-cluster-sharding/src/multi-jvm/scala/akka/cluster/sharding/ClusterShardingSpec.scala @@ -88,7 +88,7 @@ object ClusterShardingSpec extends MultiNodeConfig { override def postStop(): Unit = { super.postStop() - // Simulate that the passivation takes some time, to verify passivation bufffering + // Simulate that the passivation takes some time, to verify passivation buffering Thread.sleep(500) } //#counter-actor diff --git a/akka-cluster-tools/src/main/scala/akka/cluster/singleton/ClusterSingletonManager.scala b/akka-cluster-tools/src/main/scala/akka/cluster/singleton/ClusterSingletonManager.scala index e1a4bddc9c..f581e78663 100644 --- a/akka-cluster-tools/src/main/scala/akka/cluster/singleton/ClusterSingletonManager.scala +++ b/akka-cluster-tools/src/main/scala/akka/cluster/singleton/ClusterSingletonManager.scala @@ -154,7 +154,7 @@ object ClusterSingletonManager { */ private object Internal { /** - * Sent from new oldest to previous oldest to initate the + * Sent from new oldest to previous oldest to initiate the * hand-over process. `HandOverInProgress` and `HandOverDone` * are expected replies. */ diff --git a/akka-cluster-tools/src/multi-jvm/scala/akka/cluster/singleton/ClusterSingletonManagerSpec.scala b/akka-cluster-tools/src/multi-jvm/scala/akka/cluster/singleton/ClusterSingletonManagerSpec.scala index 4178c5ef03..08eac16745 100644 --- a/akka-cluster-tools/src/multi-jvm/scala/akka/cluster/singleton/ClusterSingletonManagerSpec.scala +++ b/akka-cluster-tools/src/multi-jvm/scala/akka/cluster/singleton/ClusterSingletonManagerSpec.scala @@ -70,7 +70,7 @@ object ClusterSingletonManagerSpec extends MultiNodeConfig { } /** - * This channel is extremly strict with regards to + * This channel is extremely strict with regards to * registration and unregistration of consumer to * be able to detect misbehaviour (e.g. two active * singleton instances). diff --git a/akka-cluster/src/main/scala/akka/cluster/ClusterDaemon.scala b/akka-cluster/src/main/scala/akka/cluster/ClusterDaemon.scala index 0682a1cd3d..99257f76a7 100644 --- a/akka-cluster/src/main/scala/akka/cluster/ClusterDaemon.scala +++ b/akka-cluster/src/main/scala/akka/cluster/ClusterDaemon.scala @@ -128,7 +128,7 @@ private[cluster] object InternalClusterAction { final case class PublisherCreated(publisher: ActorRef) /** - * Comand to [[akka.cluster.ClusterDaemon]] to create a + * Command to [[akka.cluster.ClusterDaemon]] to create a * [[akka.cluster.OnMemberStatusChangedListener]]. */ final case class AddOnMemberUpListener(callback: Runnable) extends NoSerializationVerificationNeeded diff --git a/akka-cluster/src/main/scala/akka/cluster/routing/AdaptiveLoadBalancing.scala b/akka-cluster/src/main/scala/akka/cluster/routing/AdaptiveLoadBalancing.scala index db9d8591e6..2ec432b9ec 100644 --- a/akka-cluster/src/main/scala/akka/cluster/routing/AdaptiveLoadBalancing.scala +++ b/akka-cluster/src/main/scala/akka/cluster/routing/AdaptiveLoadBalancing.scala @@ -174,7 +174,7 @@ final case class AdaptiveLoadBalancingPool( def withDispatcher(dispatcherId: String): AdaptiveLoadBalancingPool = copy(routerDispatcher = dispatcherId) /** - * Uses the supervisor strategy of the given Routerconfig + * Uses the supervisor strategy of the given RouterConfig * if this RouterConfig doesn't have one */ override def withFallback(other: RouterConfig): RouterConfig = diff --git a/akka-cluster/src/multi-jvm/scala/akka/cluster/ClusterAccrualFailureDetectorSpec.scala b/akka-cluster/src/multi-jvm/scala/akka/cluster/ClusterAccrualFailureDetectorSpec.scala index baf70ec741..4991b22da9 100644 --- a/akka-cluster/src/multi-jvm/scala/akka/cluster/ClusterAccrualFailureDetectorSpec.scala +++ b/akka-cluster/src/multi-jvm/scala/akka/cluster/ClusterAccrualFailureDetectorSpec.scala @@ -96,7 +96,7 @@ abstract class ClusterAccrualFailureDetectorSpec enterBarrier("third-shutdown") runOn(first, second) { - // remaning nodes should detect failure... + // remaining nodes should detect failure... awaitCond(!cluster.failureDetector.isAvailable(third), 15.seconds) // other connections still ok cluster.failureDetector.isAvailable(first) should ===(true) diff --git a/akka-cluster/src/multi-jvm/scala/akka/cluster/MultiNodeClusterSpec.scala b/akka-cluster/src/multi-jvm/scala/akka/cluster/MultiNodeClusterSpec.scala index bb7b8cce92..523dcca93f 100644 --- a/akka-cluster/src/multi-jvm/scala/akka/cluster/MultiNodeClusterSpec.scala +++ b/akka-cluster/src/multi-jvm/scala/akka/cluster/MultiNodeClusterSpec.scala @@ -343,7 +343,7 @@ trait MultiNodeClusterSpec extends Suite with STMultiNodeSpec with WatchedByCoro */ def markNodeAsUnavailable(address: Address): Unit = { if (isFailureDetectorPuppet) { - // before marking it as unavailble there should be at least one heartbeat + // before marking it as unavailable there should be at least one heartbeat // to create the FailureDetectorPuppet in the FailureDetectorRegistry cluster.failureDetector.heartbeat(address) failureDetectorPuppet(address) foreach (_.markNodeAsUnavailable()) diff --git a/akka-docs/_sphinx/themes/akka/static/effects.core.js b/akka-docs/_sphinx/themes/akka/static/effects.core.js index ed4fd37741..53cb941a35 100644 --- a/akka-docs/_sphinx/themes/akka/static/effects.core.js +++ b/akka-docs/_sphinx/themes/akka/static/effects.core.js @@ -94,7 +94,7 @@ $.extend($.effects, { // The main function to form the object for animation for(var n in newStyle) { if( typeof newStyle[n] != "function" && newStyle[n] /* No functions and null properties */ - && n.indexOf("Moz") == -1 && n.indexOf("length") == -1 /* No mozilla spezific render properties. */ + && n.indexOf("Moz") == -1 && n.indexOf("length") == -1 /* No mozilla specific render properties. */ && newStyle[n] != oldStyle[n] /* Only values that have changed are used for the animation */ && (n.match(/color/i) || (!n.match(/color/i) && !isNaN(parseInt(newStyle[n],10)))) /* Only things that can be parsed to integers or colors */ && (oldStyle.position != "static" || (oldStyle.position == "static" && !n.match(/left|top|bottom|right/))) /* No need for positions when dealing with static positions */ diff --git a/akka-osgi/src/main/scala/akka/osgi/ActorSystemActivator.scala b/akka-osgi/src/main/scala/akka/osgi/ActorSystemActivator.scala index dda89a2913..0a5ebbad16 100644 --- a/akka-osgi/src/main/scala/akka/osgi/ActorSystemActivator.scala +++ b/akka-osgi/src/main/scala/akka/osgi/ActorSystemActivator.scala @@ -63,7 +63,7 @@ abstract class ActorSystemActivator extends BundleActivator { val filter = s"(objectclass=${classOf[LogService].getName})" context.addServiceListener(logServiceListner, filter) - //Small trick to create an event if the service is registred before this start listing for + //Small trick to create an event if the service is registered before this start listing for Option(context.getServiceReference(classOf[LogService].getName)).foreach(x ⇒ { logServiceListner.serviceChanged(new ServiceEvent(ServiceEvent.REGISTERED, x)) }) diff --git a/akka-osgi/src/main/scala/akka/osgi/DefaultOSGiLogger.scala b/akka-osgi/src/main/scala/akka/osgi/DefaultOSGiLogger.scala index ffbe5fc2ed..68a380eeb9 100644 --- a/akka-osgi/src/main/scala/akka/osgi/DefaultOSGiLogger.scala +++ b/akka-osgi/src/main/scala/akka/osgi/DefaultOSGiLogger.scala @@ -47,8 +47,8 @@ class DefaultOSGiLogger extends DefaultLogger { } /** - * Behaviour of the Eventhanlder that is setup (has received a LogService) - * @param logService registrered OSGi LogService + * Behaviour of the Event handler that is setup (has received a LogService) + * @param logService registered OSGi LogService * @return Receive : Logs LogEvent or go back to the uninitialised state */ def initialisedReceive(logService: LogService): Receive = { @@ -65,7 +65,7 @@ class DefaultOSGiLogger extends DefaultLogger { * Logs a message in an OSGi LogService * * @param logService OSGi LogService registered and used for logging - * @param event akka LogEvent that is log unsing the LogService + * @param event akka LogEvent that is logged using the LogService */ def logMessage(logService: LogService, event: LogEvent) { event match { @@ -79,7 +79,7 @@ class DefaultOSGiLogger extends DefaultLogger { } /** - * Message sent when LogService is unregistred. + * Message sent when LogService is unregistered. * Sent from the ActorSystemActivator to a logger (as DefaultOsgiLogger). */ case object UnregisteringLogService diff --git a/akka-persistence-tck/src/main/scala/akka/persistence/PluginSpec.scala b/akka-persistence-tck/src/main/scala/akka/persistence/PluginSpec.scala index b0a5314ec8..7dc0a34d80 100644 --- a/akka-persistence-tck/src/main/scala/akka/persistence/PluginSpec.scala +++ b/akka-persistence-tck/src/main/scala/akka/persistence/PluginSpec.scala @@ -18,7 +18,7 @@ abstract class PluginSpec(val config: Config) extends TestKitBase with WordSpecL private var _pid: String = _ // used to avoid messages be delivered to a restarted actor, - // this is akka-persistence internals and journals themselfes don't really care + // this is akka-persistence internals and journals themselves don't really care protected val actorInstanceId = 1 override protected def beforeEach(): Unit = diff --git a/akka-persistence-tck/src/main/scala/akka/persistence/japi/journal/JavaJournalPerfSpec.scala b/akka-persistence-tck/src/main/scala/akka/persistence/japi/journal/JavaJournalPerfSpec.scala index 39ec50ee1d..744bab944a 100644 --- a/akka-persistence-tck/src/main/scala/akka/persistence/japi/journal/JavaJournalPerfSpec.scala +++ b/akka-persistence-tck/src/main/scala/akka/persistence/japi/journal/JavaJournalPerfSpec.scala @@ -20,7 +20,7 @@ import org.scalatest.junit.JUnitRunner * It is *NOT* meant to be a comprehensive benchmark, but rather aims to help plugin developers to easily determine * if their plugin's performance is roughly as expected. It also validates the plugin still works under "more messages" scenarios. * - * The measurements are by default printed to `System.out`, if you want to customise this please override the [[#info]] method. + * The measurements are by default printed to `System.out`, if you want to customize this please override the [[#info]] method. * * The benchmark iteration and message counts are easily customisable by overriding these methods: * @@ -36,7 +36,7 @@ import org.scalatest.junit.JUnitRunner * }}} * * In case your journal plugin needs some kind of setup or teardown, override the `beforeAll` or `afterAll` - * methods (don't forget to call `super` in your overriden methods). + * methods (don't forget to call `super` in your overridden methods). * * @see [[akka.persistence.journal.JournalSpec]] * @see [[akka.persistence.journal.JournalPerfSpec]] diff --git a/akka-persistence-tck/src/main/scala/akka/persistence/japi/journal/JavaJournalSpec.scala b/akka-persistence-tck/src/main/scala/akka/persistence/japi/journal/JavaJournalSpec.scala index 621689e164..9ff59fdca2 100644 --- a/akka-persistence-tck/src/main/scala/akka/persistence/japi/journal/JavaJournalSpec.scala +++ b/akka-persistence-tck/src/main/scala/akka/persistence/japi/journal/JavaJournalSpec.scala @@ -14,7 +14,7 @@ import org.scalatest.junit.JUnitRunner * Java / JUnit API for [[akka.persistence.journal.JournalSpec]]. * * In case your journal plugin needs some kind of setup or teardown, override the `beforeAll` or `afterAll` - * methods (don't forget to call `super` in your overriden methods). + * methods (don't forget to call `super` in your overridden methods). * * @see [[akka.persistence.journal.JournalSpec]] * @see [[akka.persistence.journal.JournalPerfSpec]] diff --git a/akka-persistence-tck/src/main/scala/akka/persistence/japi/snapshot/JavaSnapshotStoreSpec.scala b/akka-persistence-tck/src/main/scala/akka/persistence/japi/snapshot/JavaSnapshotStoreSpec.scala index 65148c7919..360629a535 100644 --- a/akka-persistence-tck/src/main/scala/akka/persistence/japi/snapshot/JavaSnapshotStoreSpec.scala +++ b/akka-persistence-tck/src/main/scala/akka/persistence/japi/snapshot/JavaSnapshotStoreSpec.scala @@ -15,7 +15,7 @@ import org.scalatest.junit.JUnitRunner * Plugin authors are highly encouraged to include it in their plugin's test suites. * * In case your snapshot-store plugin needs some kind of setup or teardown, override the `beforeAll` or `afterAll` - * methods (don't forget to call `super` in your overriden methods). + * methods (don't forget to call `super` in your overridden methods). * * @see [[akka.persistence.snapshot.SnapshotStoreSpec]] */ diff --git a/akka-persistence-tck/src/main/scala/akka/persistence/journal/JournalPerfSpec.scala b/akka-persistence-tck/src/main/scala/akka/persistence/journal/JournalPerfSpec.scala index ce94465252..a830bf48d1 100644 --- a/akka-persistence-tck/src/main/scala/akka/persistence/journal/JournalPerfSpec.scala +++ b/akka-persistence-tck/src/main/scala/akka/persistence/journal/JournalPerfSpec.scala @@ -69,7 +69,7 @@ object JournalPerfSpec { * if their plugin's performance is roughly as expected. It also validates the plugin still works under "more messages" scenarios. * * In case your journal plugin needs some kind of setup or teardown, override the `beforeAll` or `afterAll` - * methods (don't forget to call `super` in your overriden methods). + * methods (don't forget to call `super` in your overridden methods). * * For a Java and JUnit consumable version of the TCK please refer to [[akka.persistence.japi.journal.JavaJournalPerfSpec]]. * @@ -87,7 +87,7 @@ abstract class JournalPerfSpec(config: Config) extends JournalSpec(config) { testProbe.expectMsg(awaitDuration, cmnds.last) } - /** Executes a block of code multiple times (no warmup) */ + /** Executes a block of code multiple times (no warm-up) */ def measure(msg: Duration ⇒ String)(block: ⇒ Unit): Unit = { val measurements = Array.ofDim[Duration](measurementIterations) var i = 0 @@ -112,7 +112,7 @@ abstract class JournalPerfSpec(config: Config) extends JournalSpec(config) { /** Override in order to customize timeouts used for expectMsg, in order to tune the awaits to your journal's perf */ private def awaitDuration: FiniteDuration = awaitDurationMillis.millis - /** Numbe of messages sent to the PersistentActor under test for each test iteration */ + /** Number of messages sent to the PersistentActor under test for each test iteration */ def eventsCount: Int = 10 * 1000 /** Number of measurement iterations each test will be run. */ diff --git a/akka-persistence-tck/src/main/scala/akka/persistence/journal/JournalSpec.scala b/akka-persistence-tck/src/main/scala/akka/persistence/journal/JournalSpec.scala index 48f33d658a..5d9d20a4c5 100644 --- a/akka-persistence-tck/src/main/scala/akka/persistence/journal/JournalSpec.scala +++ b/akka-persistence-tck/src/main/scala/akka/persistence/journal/JournalSpec.scala @@ -21,7 +21,7 @@ object JournalSpec { * Plugin authors are highly encouraged to include it in their plugin's test suites. * * In case your journal plugin needs some kind of setup or teardown, override the `beforeAll` or `afterAll` - * methods (don't forget to call `super` in your overriden methods). + * methods (don't forget to call `super` in your overridden methods). * * For a Java and JUnit consumable version of the TCK please refer to [[akka.persistence.japi.journal.JavaJournalSpec]]. * diff --git a/akka-persistence-tck/src/main/scala/akka/persistence/snapshot/SnapshotStoreSpec.scala b/akka-persistence-tck/src/main/scala/akka/persistence/snapshot/SnapshotStoreSpec.scala index be937dea0e..373c2bc3ba 100644 --- a/akka-persistence-tck/src/main/scala/akka/persistence/snapshot/SnapshotStoreSpec.scala +++ b/akka-persistence-tck/src/main/scala/akka/persistence/snapshot/SnapshotStoreSpec.scala @@ -17,7 +17,7 @@ object SnapshotStoreSpec { * Plugin authors are highly encouraged to include it in their plugin's test suites. * * In case your journal plugin needs some kind of setup or teardown, override the `beforeAll` or `afterAll` - * methods (don't forget to call `super` in your overriden methods). + * methods (don't forget to call `super` in your overridden methods). * * For a Java and JUnit consumable version of the TCK please refer to [[akka.persistence.japi.snapshot.JavaSnapshotStoreSpec]]. * diff --git a/akka-persistence/src/main/scala/akka/persistence/Eventsourced.scala b/akka-persistence/src/main/scala/akka/persistence/Eventsourced.scala index f5e79be932..d8b2f87792 100644 --- a/akka-persistence/src/main/scala/akka/persistence/Eventsourced.scala +++ b/akka-persistence/src/main/scala/akka/persistence/Eventsourced.scala @@ -25,7 +25,7 @@ private[persistence] object Eventsourced { def evt: Any def handler: Any ⇒ Unit } - /** forces actor to stash incoming commands untill all these invocations are handled */ + /** forces actor to stash incoming commands until all these invocations are handled */ private final case class StashingHandlerInvocation(evt: Any, handler: Any ⇒ Unit) extends PendingHandlerInvocation /** does not force the actor to stash commands; Originates from either `persistAsync` or `defer` calls */ private final case class AsyncHandlerInvocation(evt: Any, handler: Any ⇒ Unit) extends PendingHandlerInvocation diff --git a/akka-remote-tests/src/multi-jvm/scala/akka/remote/RemoteNodeShutdownAndComesBackSpec.scala b/akka-remote-tests/src/multi-jvm/scala/akka/remote/RemoteNodeShutdownAndComesBackSpec.scala index 1236110c10..293183712d 100644 --- a/akka-remote-tests/src/multi-jvm/scala/akka/remote/RemoteNodeShutdownAndComesBackSpec.scala +++ b/akka-remote-tests/src/multi-jvm/scala/akka/remote/RemoteNodeShutdownAndComesBackSpec.scala @@ -111,7 +111,7 @@ abstract class RemoteNodeShutdownAndComesBackSpec expectTerminated(subject) // Establish watch with the new system. This triggers additional system message traffic. If buffers are out - // of synch the remote system will be quarantined and the rest of the test will fail (or even in earlier + // of sync the remote system will be quarantined and the rest of the test will fail (or even in earlier // stages depending on circumstances). system.actorSelection(RootActorPath(secondAddress) / "user" / "subject") ! Identify("subject") val subjectNew = expectMsgType[ActorIdentity].ref.get diff --git a/akka-remote-tests/src/multi-jvm/scala/akka/remote/Ticket15109Spec.scala b/akka-remote-tests/src/multi-jvm/scala/akka/remote/Ticket15109Spec.scala index 86d0cff73a..3235270788 100644 --- a/akka-remote-tests/src/multi-jvm/scala/akka/remote/Ticket15109Spec.scala +++ b/akka-remote-tests/src/multi-jvm/scala/akka/remote/Ticket15109Spec.scala @@ -89,8 +89,8 @@ abstract class Ticket15109Spec extends MultiNodeSpec(Ticket15109Spec) enterBarrier("actor-identified") runOn(second) { - // Force a dissassociation. Using the message Shutdown, which is suboptimal here, but this is the only - // DisassoicateInfo that triggers the code-path we want to test + // Force a disassociation. Using the message Shutdown, which is suboptimal here, but this is the only + // DisassociateInfo that triggers the code-path we want to test Await.result(RARP(system).provider.transport.managementCommand( ForceDisassociateExplicitly(node(first).address, AssociationHandle.Shutdown)), 3.seconds) } diff --git a/akka-remote-tests/src/multi-jvm/scala/akka/remote/routing/RemoteRoundRobinSpec.scala b/akka-remote-tests/src/multi-jvm/scala/akka/remote/routing/RemoteRoundRobinSpec.scala index afe4f1d45a..25d3ab6621 100644 --- a/akka-remote-tests/src/multi-jvm/scala/akka/remote/routing/RemoteRoundRobinSpec.scala +++ b/akka-remote-tests/src/multi-jvm/scala/akka/remote/routing/RemoteRoundRobinSpec.scala @@ -134,7 +134,7 @@ class RemoteRoundRobinSpec extends MultiNodeSpec(RemoteRoundRobinMultiJvmSpec) actor.isInstanceOf[RoutedActorRef] should ===(true) actor ! GetRoutees - // initial nrOfInstances 1 + inital resize => 2 + // initial nrOfInstances 1 + initial resize => 2 expectMsgType[Routees].routees.size should ===(2) val repliesFrom: Set[ActorRef] = diff --git a/akka-remote/src/main/scala/akka/remote/Endpoint.scala b/akka-remote/src/main/scala/akka/remote/Endpoint.scala index ae6ecda855..fb0b53ac73 100644 --- a/akka-remote/src/main/scala/akka/remote/Endpoint.scala +++ b/akka-remote/src/main/scala/akka/remote/Endpoint.scala @@ -703,7 +703,7 @@ private[remote] class EndpointWriter( val backoffNanos = backoffDeadlinelineNanoTime - System.nanoTime if (backoffNanos > 0) { LockSupport.parkNanos(backoffNanos) - // parkNanos allows for spurious wakeup, check again + // parkNanos allows for spurious wake-up, check again backoff() } } @@ -791,7 +791,7 @@ private[remote] class EndpointWriter( case Some(r) ⇒ r.tell(s, replyTo) case None ⇒ - // initalizing, buffer and take care of it later when buffer is sent + // initializing, buffer and take care of it later when buffer is sent enqueueInBuffer(s) } case TakeOver(newHandle, replyTo) ⇒ diff --git a/akka-remote/src/main/scala/akka/remote/FailureDetectorRegistry.scala b/akka-remote/src/main/scala/akka/remote/FailureDetectorRegistry.scala index 658d32b36f..7cedd56775 100644 --- a/akka-remote/src/main/scala/akka/remote/FailureDetectorRegistry.scala +++ b/akka-remote/src/main/scala/akka/remote/FailureDetectorRegistry.scala @@ -32,7 +32,7 @@ trait FailureDetectorRegistry[A] { /** * Records a heartbeat for a resource. If the resource is not yet registered (i.e. this is the first heartbeat) then - * it is automatially registered. + * it is automatically registered. */ def heartbeat(resource: A): Unit diff --git a/akka-remote/src/main/scala/akka/remote/PhiAccrualFailureDetector.scala b/akka-remote/src/main/scala/akka/remote/PhiAccrualFailureDetector.scala index 0f71f8b4a6..819e187b4c 100644 --- a/akka-remote/src/main/scala/akka/remote/PhiAccrualFailureDetector.scala +++ b/akka-remote/src/main/scala/akka/remote/PhiAccrualFailureDetector.scala @@ -186,7 +186,7 @@ private[akka] object HeartbeatHistory { * Create an empty HeartbeatHistory, without any history. * Can only be used as starting point for appending intervals. * The stats (mean, variance, stdDeviation) are not defined for - * for empty HeartbeatHistory, i.e. throws AritmeticException. + * for empty HeartbeatHistory, i.e. throws ArithmeticException. */ def apply(maxSampleSize: Int): HeartbeatHistory = HeartbeatHistory( maxSampleSize = maxSampleSize, @@ -201,7 +201,7 @@ private[akka] object HeartbeatHistory { * It is capped by the number of samples specified in `maxSampleSize`. * * The stats (mean, variance, stdDeviation) are not defined for - * for empty HeartbeatHistory, i.e. throws AritmeticException. + * for empty HeartbeatHistory, i.e. throws ArithmeticException. */ private[akka] final case class HeartbeatHistory private ( maxSampleSize: Int, diff --git a/akka-remote/src/main/scala/akka/remote/RemoteTransport.scala b/akka-remote/src/main/scala/akka/remote/RemoteTransport.scala index cb6556a5c4..ae95ffded1 100644 --- a/akka-remote/src/main/scala/akka/remote/RemoteTransport.scala +++ b/akka-remote/src/main/scala/akka/remote/RemoteTransport.scala @@ -47,7 +47,7 @@ private[akka] abstract class RemoteTransport(val system: ExtendedActorSystem, va def addresses: immutable.Set[Address] /** - * The default transport address of the actorsystem + * The default transport address of the ActorSystem * @return The listen address of the default transport */ def defaultAddress: Address diff --git a/akka-remote/src/main/scala/akka/remote/RemoteWatcher.scala b/akka-remote/src/main/scala/akka/remote/RemoteWatcher.scala index 26b655738f..ca9f8a3ee8 100644 --- a/akka-remote/src/main/scala/akka/remote/RemoteWatcher.scala +++ b/akka-remote/src/main/scala/akka/remote/RemoteWatcher.scala @@ -68,7 +68,7 @@ private[akka] object RemoteWatcher { * For a new node to be watched this actor periodically sends `RemoteWatcher.Heartbeat` * to the peer actor on the other node, which replies with [[RemoteWatcher.HeartbeatRsp]] * message back. The failure detector on the watching side monitors these heartbeat messages. - * If arrival of hearbeat messages stops it will be detected and this actor will publish + * If arrival of heartbeat messages stops it will be detected and this actor will publish * [[akka.actor.AddressTerminated]] to the [[akka.event.AddressTerminatedTopic]]. * * When all actors on a node have been unwatched it will stop sending heartbeat messages. diff --git a/akka-remote/src/main/scala/akka/remote/serialization/DaemonMsgCreateSerializer.scala b/akka-remote/src/main/scala/akka/remote/serialization/DaemonMsgCreateSerializer.scala index adf02f2cf5..2478d722f8 100644 --- a/akka-remote/src/main/scala/akka/remote/serialization/DaemonMsgCreateSerializer.scala +++ b/akka-remote/src/main/scala/akka/remote/serialization/DaemonMsgCreateSerializer.scala @@ -16,7 +16,7 @@ import scala.reflect.ClassTag import util.{ Failure, Success } /** - * Serializes akka's internal DaemonMsgCreate using protobuf + * Serializes Akka's internal DaemonMsgCreate using protobuf * for the core structure of DaemonMsgCreate, Props and Deploy. * Serialization of contained RouterConfig, Config, and Scope * is done with configured serializer for those classes, by diff --git a/akka-remote/src/main/scala/akka/remote/transport/TestTransport.scala b/akka-remote/src/main/scala/akka/remote/transport/TestTransport.scala index be365af7c5..4bcb31b97b 100644 --- a/akka-remote/src/main/scala/akka/remote/transport/TestTransport.scala +++ b/akka-remote/src/main/scala/akka/remote/transport/TestTransport.scala @@ -20,7 +20,7 @@ import scala.concurrent.ExecutionContext.Implicits.global * * The TestTransport is basically a shared memory between actor systems. The TestTransport could be programmed to * emulate different failure modes of a Transport implementation. TestTransport keeps a log of the activities it was - * requested to do. This class is not optimized for performace and MUST not be used as an in-memory transport in + * requested to do. This class is not optimized for performance and MUST not be used as an in-memory transport in * production systems. */ class TestTransport( @@ -177,7 +177,7 @@ object TestTransport { /** * Test utility to make behavior of functions that return some Future[B] controllable from tests. This tool is able * to overwrite default behavior with any generic behavior, including failure, and exposes control to the timing of - * the completition of the returned future. + * the completion of the returned future. * * The utility is implemented as a stack of behaviors, where the behavior on the top of the stack represents the * currently active behavior. The bottom of the stack always contains the defaultBehavior which can not be popped diff --git a/akka-remote/src/main/scala/akka/remote/transport/ThrottlerTransportAdapter.scala b/akka-remote/src/main/scala/akka/remote/transport/ThrottlerTransportAdapter.scala index 42d6b574e1..31c567056a 100644 --- a/akka-remote/src/main/scala/akka/remote/transport/ThrottlerTransportAdapter.scala +++ b/akka-remote/src/main/scala/akka/remote/transport/ThrottlerTransportAdapter.scala @@ -145,13 +145,13 @@ object ThrottlerTransportAdapter { } /** - * Management Command to force dissocation of an address. + * Management Command to force disassociation of an address. */ @SerialVersionUID(1L) final case class ForceDisassociate(address: Address) /** - * Management Command to force dissocation of an address with an explicit error. + * Management Command to force disassociation of an address with an explicit error. */ @SerialVersionUID(1L) final case class ForceDisassociateExplicitly(address: Address, reason: DisassociateInfo) diff --git a/akka-remote/src/main/scala/akka/remote/transport/netty/NettyTransport.scala b/akka-remote/src/main/scala/akka/remote/transport/netty/NettyTransport.scala index 2bceae4d55..d1d3202c84 100644 --- a/akka-remote/src/main/scala/akka/remote/transport/netty/NettyTransport.scala +++ b/akka-remote/src/main/scala/akka/remote/transport/netty/NettyTransport.scala @@ -432,7 +432,7 @@ class NettyTransport(val settings: NettyTransportSettings, val system: ExtendedA } catch { case NonFatal(e) ⇒ { log.error("failed to bind to {}, shutting down Netty transport", address) - try { shutdown() } catch { case NonFatal(e) ⇒ } // ingore possible exception during shutdown + try { shutdown() } catch { case NonFatal(e) ⇒ } // ignore possible exception during shutdown throw e } } diff --git a/akka-samples/akka-sample-cluster-java/src/multi-jvm/scala/sample/cluster/transformation/TransformationSampleSpec.scala b/akka-samples/akka-sample-cluster-java/src/multi-jvm/scala/sample/cluster/transformation/TransformationSampleSpec.scala index 48456980fb..ea075f2ed0 100644 --- a/akka-samples/akka-sample-cluster-java/src/multi-jvm/scala/sample/cluster/transformation/TransformationSampleSpec.scala +++ b/akka-samples/akka-sample-cluster-java/src/multi-jvm/scala/sample/cluster/transformation/TransformationSampleSpec.scala @@ -81,7 +81,7 @@ abstract class TransformationSampleSpec extends MultiNodeSpec(TransformationSamp val transformationFrontend = system.actorOf(Props[TransformationFrontend], name = "frontend") transformationFrontend ! new TransformationJob("hello") expectMsgPF() { - // no backends yet, service unavailble + // no backends yet, service unavailable case f: JobFailed => } } diff --git a/akka-samples/akka-sample-cluster-scala/src/multi-jvm/scala/sample/cluster/transformation/TransformationSampleSpec.scala b/akka-samples/akka-sample-cluster-scala/src/multi-jvm/scala/sample/cluster/transformation/TransformationSampleSpec.scala index 0a8d7eda04..d21388b393 100644 --- a/akka-samples/akka-sample-cluster-scala/src/multi-jvm/scala/sample/cluster/transformation/TransformationSampleSpec.scala +++ b/akka-samples/akka-sample-cluster-scala/src/multi-jvm/scala/sample/cluster/transformation/TransformationSampleSpec.scala @@ -79,7 +79,7 @@ abstract class TransformationSampleSpec extends MultiNodeSpec(TransformationSamp val transformationFrontend = system.actorOf(Props[TransformationFrontend], name = "frontend") transformationFrontend ! TransformationJob("hello") expectMsgPF() { - // no backends yet, service unavailble + // no backends yet, service unavailable case JobFailed(_, TransformationJob("hello")) => } } diff --git a/akka-samples/akka-sample-fsm-java-lambda/src/main/java/sample/fsm/DiningHakkersOnFsm.java b/akka-samples/akka-sample-fsm-java-lambda/src/main/java/sample/fsm/DiningHakkersOnFsm.java index 2a03fdb4e5..c61cbbe62d 100644 --- a/akka-samples/akka-sample-fsm-java-lambda/src/main/java/sample/fsm/DiningHakkersOnFsm.java +++ b/akka-samples/akka-sample-fsm-java-lambda/src/main/java/sample/fsm/DiningHakkersOnFsm.java @@ -54,7 +54,7 @@ public class DiningHakkersOnFsm { event((event, data) -> (event == Put) && (data.hakker == sender()), (event, data) -> goTo(CS.Available).using(new TakenBy(context().system().deadLetters())))); - // Initialze the chopstick + // Initialize the chopstick initialize(); } } diff --git a/akka-samples/akka-sample-fsm-scala/src/main/scala/sample/become/DiningHakkersOnBecome.scala b/akka-samples/akka-sample-fsm-scala/src/main/scala/sample/become/DiningHakkersOnBecome.scala index 558c1df637..9e5bc17c14 100644 --- a/akka-samples/akka-sample-fsm-scala/src/main/scala/sample/become/DiningHakkersOnBecome.scala +++ b/akka-samples/akka-sample-fsm-scala/src/main/scala/sample/become/DiningHakkersOnBecome.scala @@ -46,7 +46,7 @@ class Chopstick extends Actor { } /* -* A hakker is an awesome dude or dudett who either thinks about hacking or has to eat ;-) +* A hakker is an awesome dude or dudette who either thinks about hacking or has to eat ;-) */ class Hakker(name: String, left: ActorRef, right: ActorRef) extends Actor { diff --git a/akka-samples/akka-sample-fsm-scala/src/main/scala/sample/fsm/DiningHakkersOnFsm.scala b/akka-samples/akka-sample-fsm-scala/src/main/scala/sample/fsm/DiningHakkersOnFsm.scala index 3d56bee290..5df85df01f 100644 --- a/akka-samples/akka-sample-fsm-scala/src/main/scala/sample/fsm/DiningHakkersOnFsm.scala +++ b/akka-samples/akka-sample-fsm-scala/src/main/scala/sample/fsm/DiningHakkersOnFsm.scala @@ -53,7 +53,7 @@ class Chopstick extends Actor with FSM[ChopstickState, TakenBy] { goto(Available) using TakenBy(system.deadLetters) } - // Initialze the chopstick + // Initialize the chopstick initialize() } diff --git a/akka-samples/akka-sample-osgi-dining-hakkers/core/src/main/scala/akka/sample/osgi/internal/Hakker.scala b/akka-samples/akka-sample-osgi-dining-hakkers/core/src/main/scala/akka/sample/osgi/internal/Hakker.scala index 7f2a6f00b1..96a754b8bb 100644 --- a/akka-samples/akka-sample-osgi-dining-hakkers/core/src/main/scala/akka/sample/osgi/internal/Hakker.scala +++ b/akka-samples/akka-sample-osgi-dining-hakkers/core/src/main/scala/akka/sample/osgi/internal/Hakker.scala @@ -46,7 +46,7 @@ class Chopstick extends Actor { } /* -* A hakker is an awesome dude or dudett who either thinks about hacking or has to eat ;-) +* A hakker is an awesome dude or dudette who either thinks about hacking or has to eat ;-) */ class Hakker(name: String, chair: Int) extends Actor { diff --git a/akka-testkit/src/main/scala/akka/testkit/TestActorRef.scala b/akka-testkit/src/main/scala/akka/testkit/TestActorRef.scala index 065904bcbc..be0702fe6f 100644 --- a/akka-testkit/src/main/scala/akka/testkit/TestActorRef.scala +++ b/akka-testkit/src/main/scala/akka/testkit/TestActorRef.scala @@ -105,7 +105,7 @@ class TestActorRef[T <: Actor]( def watch(subject: ActorRef): ActorRef = underlying.watch(subject) /** - * Deregisters this actor from being a death monitor of the provided ActorRef + * Unregisters this actor from being a death monitor of the provided ActorRef * This means that this actor will not get a Terminated()-message when the provided actor * is permanently terminated. * diff --git a/akka-testkit/src/main/scala/akka/testkit/TestEventListener.scala b/akka-testkit/src/main/scala/akka/testkit/TestEventListener.scala index 90690a6eaa..37fdf2eb86 100644 --- a/akka-testkit/src/main/scala/akka/testkit/TestEventListener.scala +++ b/akka-testkit/src/main/scala/akka/testkit/TestEventListener.scala @@ -20,7 +20,7 @@ import akka.util.BoxedType /** * Implementation helpers of the EventFilter facilities: send `Mute` * to the TestEventListener to install a filter, and `UnMute` to - * deinstall it. + * uninstall it. * * You should always prefer the filter methods in the package object * (see [[akka.testkit]] `filterEvents` and `filterException`) or on the @@ -31,7 +31,7 @@ sealed trait TestEvent /** * Implementation helpers of the EventFilter facilities: send Mute * to the TestEventFilter to install a filter, and UnMute to - * deinstall it. + * uninstall it. * * You should always prefer the filter methods in the package object * (see [[akka.testkit]] `filterEvents` and `filterException`) or on the diff --git a/akka-testkit/src/main/scala/akka/testkit/TestFSMRef.scala b/akka-testkit/src/main/scala/akka/testkit/TestFSMRef.scala index 2fbd1d4c0b..cf3ecf509d 100644 --- a/akka-testkit/src/main/scala/akka/testkit/TestFSMRef.scala +++ b/akka-testkit/src/main/scala/akka/testkit/TestFSMRef.scala @@ -13,7 +13,7 @@ import akka.dispatch.MailboxType import scala.reflect.ClassTag /** - * This is a specialised form of the TestActorRef with support for querying and + * This is a specialized form of the TestActorRef with support for querying and * setting the state of a FSM. Use a LoggingFSM with this class if you also * need to inspect event traces. * diff --git a/akka-testkit/src/main/scala/akka/testkit/TestKit.scala b/akka-testkit/src/main/scala/akka/testkit/TestKit.scala index 295381f156..f253e8fbcf 100644 --- a/akka-testkit/src/main/scala/akka/testkit/TestKit.scala +++ b/akka-testkit/src/main/scala/akka/testkit/TestKit.scala @@ -389,7 +389,7 @@ trait TestKitBase { * partial function matches and returns false. Use it to ignore certain * messages while waiting for a specific message. * - * @return the last received messsage, i.e. the first one for which the + * @return the last received message, i.e. the first one for which the * partial function returned true */ def fishForMessage(max: Duration = Duration.Undefined, hint: String = "")(f: PartialFunction[Any, Boolean]): Any = { diff --git a/akka-testkit/src/test/scala/akka/testkit/metrics/KnownOpsInTimespanTimer.scala b/akka-testkit/src/test/scala/akka/testkit/metrics/KnownOpsInTimespanTimer.scala index aa5ea48ba5..816a0baad7 100644 --- a/akka-testkit/src/test/scala/akka/testkit/metrics/KnownOpsInTimespanTimer.scala +++ b/akka-testkit/src/test/scala/akka/testkit/metrics/KnownOpsInTimespanTimer.scala @@ -8,7 +8,7 @@ import java.util.concurrent.atomic.AtomicLong import java.util.concurrent.TimeUnit._ /** - * Specialised "one-shot" Timer. + * Specialized "one-shot" Timer. * Given a known number of operations performed within a time span (to be measured) it displays the average time one operation took. * * Please note that this is a *very coarse* estimation; The gain though is that we do not have to perform the counting inside of the measured thing (we can adding in tight loops). diff --git a/akka-testkit/src/test/scala/akka/testkit/metrics/MetricsKitOps.scala b/akka-testkit/src/test/scala/akka/testkit/metrics/MetricsKitOps.scala index 68e94f452c..bf084521e9 100644 --- a/akka-testkit/src/test/scala/akka/testkit/metrics/MetricsKitOps.scala +++ b/akka-testkit/src/test/scala/akka/testkit/metrics/MetricsKitOps.scala @@ -27,7 +27,7 @@ private[akka] trait MetricsKitOps extends MetricKeyDSL { /** * Used to measure timing of known number of operations over time. - * While not being the most percise, it allows to measure a coarse op/s without injecting counters to the measured operation (potentially hot-loop). + * While not being the most precise, it allows to measure a coarse op/s without injecting counters to the measured operation (potentially hot-loop). * * Do not use for short running pieces of code. */ diff --git a/akka-typed/src/main/scala/akka/typed/Behavior.scala b/akka-typed/src/main/scala/akka/typed/Behavior.scala index 9783fc0248..27a2fc5ba7 100644 --- a/akka-typed/src/main/scala/akka/typed/Behavior.scala +++ b/akka-typed/src/main/scala/akka/typed/Behavior.scala @@ -104,7 +104,7 @@ final case class PostRestart(failure: Throwable) extends Signal * * IMPORTANT NOTE: if the actor terminated by switching to the * `Stopped` behavior then this signal will be ignored (i.e. the - * Stopped behvavior will do nothing in reaction to it). + * Stopped behavior will do nothing in reaction to it). */ @SerialVersionUID(1L) final case object PostStop extends Signal diff --git a/akka-typed/src/main/scala/akka/typed/patterns/Receptionist.scala b/akka-typed/src/main/scala/akka/typed/patterns/Receptionist.scala index 19c32d966d..b64987e9d4 100644 --- a/akka-typed/src/main/scala/akka/typed/patterns/Receptionist.scala +++ b/akka-typed/src/main/scala/akka/typed/patterns/Receptionist.scala @@ -53,7 +53,7 @@ object Receptionist { final case class Find[T](key: ServiceKey[T])(val replyTo: ActorRef[Listing[T]]) extends Command /** - * Confirmtion that the given [[akka.typed.ActorRef]] has been associated with the [[ServiceKey]]. + * Confirmation that the given [[akka.typed.ActorRef]] has been associated with the [[ServiceKey]]. */ final case class Registered[T](key: ServiceKey[T], address: ActorRef[T]) /** diff --git a/project/OSGi.scala b/project/OSGi.scala index 5ed7fca5cd..94afa1e0e2 100644 --- a/project/OSGi.scala +++ b/project/OSGi.scala @@ -48,7 +48,7 @@ object OSGi { val osgiOptionalImports = Seq( // needed because testkit is normally not used in the application bundle, // but it should still be included as transitive dependency and used by BundleDelegatingClassLoader - // to be able to find refererence.conf + // to be able to find reference.conf "akka.testkit", "com.google.protobuf") diff --git a/project/Sample.scala b/project/Sample.scala index 5d537ca2bb..9e0bb4c893 100644 --- a/project/Sample.scala +++ b/project/Sample.scala @@ -41,7 +41,7 @@ object Sample { val dependencies = buildDependencies.value val classpathWithProjectDependencies = dependencies.classpath.map { case (proj, deps) if proj.project == project.id => - // add project dependency for every akka library dependnecy + // add project dependency for every akka library dependency (proj, deps ++ projectDependencies.map(ResolvedClasspathDependency(_, None))) case (project, deps) => (project, deps) }