=doc #17329 Fixed and normalized spellings in ScalaDoc and comments
This commit is contained in:
parent
10f039f70d
commit
bc791eb86c
77 changed files with 100 additions and 100 deletions
|
|
@ -137,7 +137,7 @@ public abstract class AbstractBoundedNodeQueue<T> {
|
|||
*/
|
||||
public final int size() {
|
||||
//Order of operations is extremely important here
|
||||
// If no item was dequeued between when we looked at the count of the enqueueing end,
|
||||
// If no item was dequeued between when we looked at the count of the enqueuing end,
|
||||
// there should be no out-of-bounds
|
||||
for(;;) {
|
||||
final int deqCountBefore = getDeq().count;
|
||||
|
|
|
|||
|
|
@ -465,7 +465,7 @@ private[akka] trait MinimalActorRef extends InternalActorRef with LocalRef {
|
|||
protected def writeReplace(): AnyRef = SerializedActorRef(this)
|
||||
}
|
||||
|
||||
/** Subscribe to this class to be notified about all DeadLetters (also the supressed ones). */
|
||||
/** Subscribe to this class to be notified about all DeadLetters (also the suppressed ones). */
|
||||
sealed trait AllDeadLetters {
|
||||
def message: Any
|
||||
def sender: ActorRef
|
||||
|
|
|
|||
|
|
@ -133,7 +133,7 @@ object ActorSystem {
|
|||
* falls back to the ClassLoader associated with the ActorSystem class.
|
||||
* If an ExecutionContext is given, it will be used as the default executor inside this ActorSystem.
|
||||
* If no ExecutionContext is given, the system will fallback to the executor configured under "akka.actor.default-dispatcher.default-executor.fallback".
|
||||
* The system will use the passed in config, or falls back to the deafult reference configuration using the ClassLoader.
|
||||
* The system will use the passed in config, or falls back to the default reference configuration using the ClassLoader.
|
||||
*
|
||||
* @see <a href="http://typesafehub.github.io/config/v1.3.0/" target="_blank">The Typesafe Config Library API Documentation</a>
|
||||
*/
|
||||
|
|
@ -828,7 +828,7 @@ private[akka] class ActorSystemImpl(
|
|||
private[this] final val done = Promise[T]()
|
||||
private[this] final val ref = new AtomicReference(done)
|
||||
|
||||
// onComplete never fires twice so safe to avoid nullcheck
|
||||
// onComplete never fires twice so safe to avoid null check
|
||||
upStreamTerminated onComplete { t ⇒ ref.getAndSet(null).complete(t) }
|
||||
|
||||
/**
|
||||
|
|
|
|||
|
|
@ -212,7 +212,7 @@ object FSM {
|
|||
final case class Event[D](event: Any, stateData: D) extends NoSerializationVerificationNeeded
|
||||
|
||||
/**
|
||||
* Case class representing the state of the [[akka.actor.FSM]] whithin the
|
||||
* Case class representing the state of the [[akka.actor.FSM]] within the
|
||||
* `onTermination` block.
|
||||
*/
|
||||
final case class StopEvent[S, D](reason: Reason, currentState: S, stateData: D) extends NoSerializationVerificationNeeded
|
||||
|
|
|
|||
|
|
@ -213,7 +213,7 @@ class LightArrayRevolverScheduler(config: Config,
|
|||
try {
|
||||
((d + TickDuration - oneNs) / TickDuration).toLong * TickDuration
|
||||
} catch {
|
||||
case _: IllegalArgumentException ⇒ d // rouding up Long.MaxValue.nanos overflows
|
||||
case _: IllegalArgumentException ⇒ d // rounding up Long.MaxValue.nanos overflows
|
||||
}
|
||||
|
||||
/**
|
||||
|
|
|
|||
|
|
@ -59,7 +59,7 @@ trait TypedActorFactory {
|
|||
}
|
||||
|
||||
/**
|
||||
* Returns wether the supplied AnyRef is a TypedActor proxy or not
|
||||
* Returns whether the supplied AnyRef is a TypedActor proxy or not
|
||||
*/
|
||||
def isTypedActor(proxyOrNot: AnyRef): Boolean
|
||||
|
||||
|
|
@ -654,7 +654,7 @@ class TypedActorExtension(val system: ExtendedActorSystem) extends TypedActorFac
|
|||
}
|
||||
|
||||
/**
|
||||
* Returns wether the supplied AnyRef is a TypedActor proxy or not
|
||||
* Returns whether the supplied AnyRef is a TypedActor proxy or not
|
||||
*/
|
||||
def isTypedActor(proxyOrNot: AnyRef): Boolean = invocationHandlerFor(proxyOrNot) ne null
|
||||
|
||||
|
|
@ -675,7 +675,7 @@ class TypedActorExtension(val system: ExtendedActorSystem) extends TypedActorFac
|
|||
proxy
|
||||
} else {
|
||||
proxyVar set proxy // Chicken and egg situation we needed to solve, set the proxy so that we can set the self-reference inside each receive
|
||||
actorVar set actorRef //Make sure the InvocationHandler gets ahold of the actor reference, this is not a problem since the proxy hasn't escaped this method yet
|
||||
actorVar set actorRef //Make sure the InvocationHandler gets a hold of the actor reference, this is not a problem since the proxy hasn't escaped this method yet
|
||||
proxyVar.get
|
||||
}
|
||||
}
|
||||
|
|
|
|||
|
|
@ -59,7 +59,7 @@ private[akka] trait BatchingExecutor extends Executor {
|
|||
protected final def resubmitUnbatched(): Boolean = {
|
||||
val current = _tasksLocal.get()
|
||||
_tasksLocal.remove()
|
||||
if ((current eq this) && !current.isEmpty) { // Resubmit outselves if something bad happened and we still have work to do
|
||||
if ((current eq this) && !current.isEmpty) { // Resubmit ourselves if something bad happened and we still have work to do
|
||||
unbatchedExecute(current) //TODO what if this submission fails?
|
||||
true
|
||||
} else false
|
||||
|
|
|
|||
|
|
@ -88,7 +88,7 @@ object Futures {
|
|||
*
|
||||
* The result becomes available once the asynchronous computation is completed.
|
||||
*
|
||||
* @param body the asychronous computation
|
||||
* @param body the asynchronous computation
|
||||
* @param executor the execution context on which the future is run
|
||||
* @return the `Future` holding the result of the computation
|
||||
*/
|
||||
|
|
|
|||
|
|
@ -126,7 +126,7 @@ class EventStream(sys: ActorSystem, private val debug: Boolean) extends LoggingB
|
|||
*
|
||||
* The actual check if the subscriber still has subscriptions is performed by the `EventStreamUnsubscriber`,
|
||||
* because it's an expensive operation, and we don want to block client-code for that long, the Actor will eventually
|
||||
* catch up and perform the apropriate operation.
|
||||
* catch up and perform the appropriate operation.
|
||||
*/
|
||||
@tailrec
|
||||
private def unregisterIfNoMoreSubscribedChannels(subscriber: ActorRef): Unit = {
|
||||
|
|
|
|||
|
|
@ -224,7 +224,7 @@ trait LoggingBus extends ActorEventBus {
|
|||
* }
|
||||
*
|
||||
* class MyClass extends MyType {
|
||||
* val sys = ActorSyste("sys")
|
||||
* val sys = ActorSystem("sys")
|
||||
* val log = Logging(sys, this) // will use "hallo,akka://sys" as logSource
|
||||
* def name = "hallo"
|
||||
* }
|
||||
|
|
|
|||
|
|
@ -52,7 +52,7 @@ private[io] trait ChannelRegistry {
|
|||
/**
|
||||
* Implementations of this interface are sent as actor messages back to a channel actor as
|
||||
* a result of it having called `register` on the `ChannelRegistry`.
|
||||
* Enables a channel actor to directly schedule interest setting tasks to the selector mgmt. dispatcher.
|
||||
* Enables a channel actor to directly schedule interest setting tasks to the selector management dispatcher.
|
||||
*/
|
||||
private[io] trait ChannelRegistration extends NoSerializationVerificationNeeded {
|
||||
def enableInterest(op: Int)
|
||||
|
|
|
|||
|
|
@ -26,7 +26,7 @@ import java.lang.{ Iterable ⇒ JIterable }
|
|||
* In order to open an outbound connection send a [[Tcp.Connect]] message
|
||||
* to the [[TcpExt#manager]].
|
||||
*
|
||||
* In order to start listening for inbound connetions send a [[Tcp.Bind]]
|
||||
* In order to start listening for inbound connections send a [[Tcp.Bind]]
|
||||
* message to the [[TcpExt#manager]].
|
||||
*
|
||||
* The Java API for generating TCP commands is available at [[TcpMessage]].
|
||||
|
|
|
|||
|
|
@ -49,11 +49,11 @@ private[io] class TcpManager(tcp: TcpExt)
|
|||
|
||||
def receive = workerForCommandHandler {
|
||||
case c: Connect ⇒
|
||||
val commander = sender() // cache because we create a function that will run asyncly
|
||||
val commander = sender() // cache because we create a function that will run asynchly
|
||||
(registry ⇒ Props(classOf[TcpOutgoingConnection], tcp, registry, commander, c))
|
||||
|
||||
case b: Bind ⇒
|
||||
val commander = sender() // cache because we create a function that will run asyncly
|
||||
val commander = sender() // cache because we create a function that will run asynchly
|
||||
(registry ⇒ Props(classOf[TcpListener], selectorPool, tcp, registry, commander, b))
|
||||
}
|
||||
|
||||
|
|
|
|||
|
|
@ -14,7 +14,7 @@ private[io] class UdpConnectedManager(udpConn: UdpConnectedExt)
|
|||
|
||||
def receive = workerForCommandHandler {
|
||||
case c: Connect ⇒
|
||||
val commander = sender() // cache because we create a function that will run asyncly
|
||||
val commander = sender() // cache because we create a function that will run asynchly
|
||||
registry ⇒ Props(classOf[UdpConnection], udpConn, registry, commander, c)
|
||||
}
|
||||
|
||||
|
|
|
|||
|
|
@ -47,11 +47,11 @@ private[io] class UdpManager(udp: UdpExt) extends SelectionHandler.SelectorBased
|
|||
|
||||
def receive = workerForCommandHandler {
|
||||
case b: Bind ⇒
|
||||
val commander = sender() // cache because we create a function that will run asyncly
|
||||
val commander = sender() // cache because we create a function that will run asynchly
|
||||
(registry ⇒ Props(classOf[UdpListener], udp, registry, commander, b))
|
||||
|
||||
case SimpleSender(options) ⇒
|
||||
val commander = sender() // cache because we create a function that will run asyncly
|
||||
val commander = sender() // cache because we create a function that will run asynchly
|
||||
(registry ⇒ Props(classOf[UdpSender], udp, registry, commander, options))
|
||||
}
|
||||
|
||||
|
|
|
|||
|
|
@ -126,7 +126,7 @@ final case class BalancingPool(
|
|||
}
|
||||
|
||||
/**
|
||||
* Uses the supervisor strategy of the given Routerconfig
|
||||
* Uses the supervisor strategy of the given RouterConfig
|
||||
* if this RouterConfig doesn't have one.
|
||||
*/
|
||||
override def withFallback(other: RouterConfig): RouterConfig =
|
||||
|
|
|
|||
|
|
@ -97,7 +97,7 @@ final case class BroadcastPool(
|
|||
def withDispatcher(dispatcherId: String): BroadcastPool = copy(routerDispatcher = dispatcherId)
|
||||
|
||||
/**
|
||||
* Uses the resizer and/or the supervisor strategy of the given Routerconfig
|
||||
* Uses the resizer and/or the supervisor strategy of the given RouterConfig
|
||||
* if this RouterConfig doesn't have one, i.e. the resizer defined in code is used if
|
||||
* resizer was not defined in config.
|
||||
*/
|
||||
|
|
|
|||
|
|
@ -314,7 +314,7 @@ final case class ConsistentHashingPool(
|
|||
copy(hashMapping = ConsistentHashingRouter.hashMappingAdapter(mapper))
|
||||
|
||||
/**
|
||||
* Uses the resizer and/or the supervisor strategy of the given Routerconfig
|
||||
* Uses the resizer and/or the supervisor strategy of the given RouterConfig
|
||||
* if this RouterConfig doesn't have one, i.e. the resizer defined in code is used if
|
||||
* resizer was not defined in config.
|
||||
* Uses the the `hashMapping` defined in code, since that can't be defined in configuration.
|
||||
|
|
|
|||
|
|
@ -98,7 +98,7 @@ final case class RandomPool(
|
|||
def withDispatcher(dispatcherId: String): RandomPool = copy(routerDispatcher = dispatcherId)
|
||||
|
||||
/**
|
||||
* Uses the resizer and/or the supervisor strategy of the given Routerconfig
|
||||
* Uses the resizer and/or the supervisor strategy of the given RouterConfig
|
||||
* if this RouterConfig doesn't have one, i.e. the resizer defined in code is used if
|
||||
* resizer was not defined in config.
|
||||
*/
|
||||
|
|
|
|||
|
|
@ -49,7 +49,7 @@ trait Resizer {
|
|||
* returns true and no other resize is in progress.
|
||||
*
|
||||
* Return the number of routees to add or remove. Negative value will remove that number of routees.
|
||||
* Positive value will add that number of routess. 0 will not change the routees.
|
||||
* Positive value will add that number of routees. 0 will not change the routees.
|
||||
*
|
||||
* This method is invoked only in the context of the Router actor.
|
||||
*/
|
||||
|
|
|
|||
|
|
@ -102,7 +102,7 @@ final case class RoundRobinPool(
|
|||
def withDispatcher(dispatcherId: String): RoundRobinPool = copy(routerDispatcher = dispatcherId)
|
||||
|
||||
/**
|
||||
* Uses the resizer and/or the supervisor strategy of the given Routerconfig
|
||||
* Uses the resizer and/or the supervisor strategy of the given RouterConfig
|
||||
* if this RouterConfig doesn't have one, i.e. the resizer defined in code is used if
|
||||
* resizer was not defined in config.
|
||||
*/
|
||||
|
|
|
|||
|
|
@ -119,7 +119,7 @@ private[akka] class RoutedActorCell(
|
|||
}
|
||||
|
||||
/**
|
||||
* Called when `router` is initalized but before `super.start()` to
|
||||
* Called when `router` is initialized but before `super.start()` to
|
||||
* be able to do extra initialization in subclass.
|
||||
*/
|
||||
protected def preSuperStart(): Unit = ()
|
||||
|
|
|
|||
|
|
@ -54,7 +54,7 @@ trait RouterConfig extends Serializable {
|
|||
* Possibility to define an actor for controlling the routing
|
||||
* logic from external stimuli (e.g. monitoring metrics).
|
||||
* This actor will be a child of the router "head" actor.
|
||||
* Managment messages not handled by the "head" actor are
|
||||
* Management messages not handled by the "head" actor are
|
||||
* delegated to this controller actor.
|
||||
*/
|
||||
def routingLogicController(routingLogic: RoutingLogic): Option[Props] = None
|
||||
|
|
|
|||
|
|
@ -136,7 +136,7 @@ final case class ScatterGatherFirstCompletedPool(
|
|||
def withDispatcher(dispatcherId: String): ScatterGatherFirstCompletedPool = copy(routerDispatcher = dispatcherId)
|
||||
|
||||
/**
|
||||
* Uses the resizer and/or the supervisor strategy of the given Routerconfig
|
||||
* Uses the resizer and/or the supervisor strategy of the given RouterConfig
|
||||
* if this RouterConfig doesn't have one, i.e. the resizer defined in code is used if
|
||||
* resizer was not defined in config.
|
||||
*/
|
||||
|
|
|
|||
|
|
@ -213,7 +213,7 @@ final case class SmallestMailboxPool(
|
|||
def withDispatcher(dispatcherId: String): SmallestMailboxPool = copy(routerDispatcher = dispatcherId)
|
||||
|
||||
/**
|
||||
* Uses the resizer and/or the supervisor strategy of the given Routerconfig
|
||||
* Uses the resizer and/or the supervisor strategy of the given RouterConfig
|
||||
* if this RouterConfig doesn't have one, i.e. the resizer defined in code is used if
|
||||
* resizer was not defined in config.
|
||||
*/
|
||||
|
|
|
|||
|
|
@ -83,7 +83,7 @@ private[akka] final case class TailChoppingRoutees(
|
|||
}
|
||||
|
||||
/**
|
||||
* A router poll thats sends the message to a first, random picked, routee,
|
||||
* A router poll that sends the message to a first, random picked, routee,
|
||||
* then wait a specified `interval` and then send to a second, random picked, and so on till one full cycle..
|
||||
*
|
||||
* The configuration parameter trumps the constructor arguments. This means that
|
||||
|
|
@ -169,7 +169,7 @@ final case class TailChoppingPool(
|
|||
def withDispatcher(dispatcherId: String): TailChoppingPool = copy(routerDispatcher = dispatcherId)
|
||||
|
||||
/**
|
||||
* Uses the resizer and/or the supervisor strategy of the given Routerconfig
|
||||
* Uses the resizer and/or the supervisor strategy of the given RouterConfig
|
||||
* if this RouterConfig doesn't have one, i.e. the resizer defined in code is used if
|
||||
* resizer was not defined in config.
|
||||
*/
|
||||
|
|
|
|||
|
|
@ -138,7 +138,7 @@ class Switch(startAsOn: Boolean = false) {
|
|||
def isOn: Boolean = switch.get
|
||||
|
||||
/**
|
||||
* Returns whether the switch is IMMEDDIATELY off (no locking)
|
||||
* Returns whether the switch is IMMEDIATELY off (no locking)
|
||||
*/
|
||||
def isOff: Boolean = !isOn
|
||||
}
|
||||
|
|
|
|||
|
|
@ -103,7 +103,7 @@ class CamelSettings private[camel] (config: Config, dynamicAccess: DynamicAccess
|
|||
(s: String, r: RouteDefinition) ⇒ conversions.get(s).fold(r)(r.convertBodyTo)
|
||||
}
|
||||
/**
|
||||
* Configured setting, determine the class used to load/retrive the instance of the Camel Context
|
||||
* Configured setting, determine the class used to load/retrieve the instance of the Camel Context
|
||||
*/
|
||||
final val ContextProvider: ContextProvider = {
|
||||
val fqcn = config.getString("akka.camel.context-provider")
|
||||
|
|
|
|||
|
|
@ -163,7 +163,7 @@ final case class AdaptiveLoadBalancingPool(
|
|||
def withDispatcher(dispatcherId: String): AdaptiveLoadBalancingPool = copy(routerDispatcher = dispatcherId)
|
||||
|
||||
/**
|
||||
* Uses the supervisor strategy of the given Routerconfig
|
||||
* Uses the supervisor strategy of the given RouterConfig
|
||||
* if this RouterConfig doesn't have one
|
||||
*/
|
||||
override def withFallback(other: RouterConfig): RouterConfig =
|
||||
|
|
|
|||
|
|
@ -81,7 +81,7 @@ private[metrics] object MetricsCollector {
|
|||
* Loads JVM and system metrics through JMX monitoring beans.
|
||||
*
|
||||
* @param address The [[akka.actor.Address]] of the node being sampled
|
||||
* @param decay how quickly the exponential weighting of past data is decayed
|
||||
* @param decayFactor how quickly the exponential weighting of past data is decayed
|
||||
*/
|
||||
class JmxMetricsCollector(address: Address, decayFactor: Double) extends MetricsCollector {
|
||||
import StandardMetrics._
|
||||
|
|
@ -184,7 +184,7 @@ class JmxMetricsCollector(address: Address, decayFactor: Double) extends Metrics
|
|||
* to missing classes or native libraries.
|
||||
*
|
||||
* @param address The [[akka.actor.Address]] of the node being sampled
|
||||
* @param decay how quickly the exponential weighting of past data is decayed
|
||||
* @param decayFactor how quickly the exponential weighting of past data is decayed
|
||||
* @param sigar the org.hyperic.Sigar instance
|
||||
*/
|
||||
class SigarMetricsCollector(address: Address, decayFactor: Double, sigar: SigarProxy)
|
||||
|
|
|
|||
|
|
@ -399,7 +399,7 @@ private[akka] object ClusterShardingGuardian {
|
|||
}
|
||||
|
||||
/**
|
||||
* INTERNAL API. [[ShardRegion]] and [[ShardCoordinator]] actors are createad as children
|
||||
* INTERNAL API. [[ShardRegion]] and [[ShardCoordinator]] actors are created as children
|
||||
* of this actor.
|
||||
*/
|
||||
private[akka] class ClusterShardingGuardian extends Actor {
|
||||
|
|
@ -1467,7 +1467,7 @@ object ShardCoordinator {
|
|||
*/
|
||||
@SerialVersionUID(1L) final case class RegisterProxy(shardRegionProxy: ActorRef) extends CoordinatorCommand
|
||||
/**
|
||||
* Acknowledgement from `ShardCoordinator` that [[Register]] or [[RegisterProxy]] was sucessful.
|
||||
* Acknowledgement from `ShardCoordinator` that [[Register]] or [[RegisterProxy]] was successful.
|
||||
*/
|
||||
@SerialVersionUID(1L) final case class RegisterAck(coordinator: ActorRef) extends CoordinatorMessage
|
||||
/**
|
||||
|
|
@ -1484,7 +1484,7 @@ object ShardCoordinator {
|
|||
*/
|
||||
@SerialVersionUID(1L) final case class ShardHome(shard: ShardId, ref: ActorRef) extends CoordinatorMessage
|
||||
/**
|
||||
* `ShardCoodinator` informs a `ShardRegion` that it is hosting this shard
|
||||
* `ShardCoordinator` informs a `ShardRegion` that it is hosting this shard
|
||||
*/
|
||||
@SerialVersionUID(1L) final case class HostShard(shard: ShardId) extends CoordinatorMessage
|
||||
/**
|
||||
|
|
@ -1505,7 +1505,7 @@ object ShardCoordinator {
|
|||
*/
|
||||
@SerialVersionUID(1L) final case class BeginHandOffAck(shard: ShardId) extends CoordinatorCommand
|
||||
/**
|
||||
* When all `ShardRegion` actors have acknoledged the `BeginHandOff` the
|
||||
* When all `ShardRegion` actors have acknowledged the `BeginHandOff` the
|
||||
* `ShardCoordinator` sends this message to the `ShardRegion` responsible for the
|
||||
* shard. The `ShardRegion` is supposed to stop all entries in that shard and when
|
||||
* all entries have terminated reply with `ShardStopped` to the `ShardCoordinator`.
|
||||
|
|
@ -1872,7 +1872,7 @@ class ShardCoordinator(handOffTimeout: FiniteDuration, shardStartTimeout: Finite
|
|||
|
||||
case ShardHome(_, _) ⇒
|
||||
//On rebalance, we send ourselves a GetShardHome message to reallocate a
|
||||
// shard. This recieve handles the "response" from that message. i.e. Ingores it.
|
||||
// shard. This receive handles the "response" from that message. i.e. ignores it.
|
||||
|
||||
case ClusterShuttingDown ⇒
|
||||
log.debug("Shutting down ShardCoordinator")
|
||||
|
|
|
|||
|
|
@ -88,7 +88,7 @@ object ClusterShardingSpec extends MultiNodeConfig {
|
|||
|
||||
override def postStop(): Unit = {
|
||||
super.postStop()
|
||||
// Simulate that the passivation takes some time, to verify passivation bufffering
|
||||
// Simulate that the passivation takes some time, to verify passivation buffering
|
||||
Thread.sleep(500)
|
||||
}
|
||||
//#counter-actor
|
||||
|
|
|
|||
|
|
@ -154,7 +154,7 @@ object ClusterSingletonManager {
|
|||
*/
|
||||
private object Internal {
|
||||
/**
|
||||
* Sent from new oldest to previous oldest to initate the
|
||||
* Sent from new oldest to previous oldest to initiate the
|
||||
* hand-over process. `HandOverInProgress` and `HandOverDone`
|
||||
* are expected replies.
|
||||
*/
|
||||
|
|
|
|||
|
|
@ -70,7 +70,7 @@ object ClusterSingletonManagerSpec extends MultiNodeConfig {
|
|||
}
|
||||
|
||||
/**
|
||||
* This channel is extremly strict with regards to
|
||||
* This channel is extremely strict with regards to
|
||||
* registration and unregistration of consumer to
|
||||
* be able to detect misbehaviour (e.g. two active
|
||||
* singleton instances).
|
||||
|
|
|
|||
|
|
@ -128,7 +128,7 @@ private[cluster] object InternalClusterAction {
|
|||
final case class PublisherCreated(publisher: ActorRef)
|
||||
|
||||
/**
|
||||
* Comand to [[akka.cluster.ClusterDaemon]] to create a
|
||||
* Command to [[akka.cluster.ClusterDaemon]] to create a
|
||||
* [[akka.cluster.OnMemberStatusChangedListener]].
|
||||
*/
|
||||
final case class AddOnMemberUpListener(callback: Runnable) extends NoSerializationVerificationNeeded
|
||||
|
|
|
|||
|
|
@ -174,7 +174,7 @@ final case class AdaptiveLoadBalancingPool(
|
|||
def withDispatcher(dispatcherId: String): AdaptiveLoadBalancingPool = copy(routerDispatcher = dispatcherId)
|
||||
|
||||
/**
|
||||
* Uses the supervisor strategy of the given Routerconfig
|
||||
* Uses the supervisor strategy of the given RouterConfig
|
||||
* if this RouterConfig doesn't have one
|
||||
*/
|
||||
override def withFallback(other: RouterConfig): RouterConfig =
|
||||
|
|
|
|||
|
|
@ -96,7 +96,7 @@ abstract class ClusterAccrualFailureDetectorSpec
|
|||
enterBarrier("third-shutdown")
|
||||
|
||||
runOn(first, second) {
|
||||
// remaning nodes should detect failure...
|
||||
// remaining nodes should detect failure...
|
||||
awaitCond(!cluster.failureDetector.isAvailable(third), 15.seconds)
|
||||
// other connections still ok
|
||||
cluster.failureDetector.isAvailable(first) should ===(true)
|
||||
|
|
|
|||
|
|
@ -343,7 +343,7 @@ trait MultiNodeClusterSpec extends Suite with STMultiNodeSpec with WatchedByCoro
|
|||
*/
|
||||
def markNodeAsUnavailable(address: Address): Unit = {
|
||||
if (isFailureDetectorPuppet) {
|
||||
// before marking it as unavailble there should be at least one heartbeat
|
||||
// before marking it as unavailable there should be at least one heartbeat
|
||||
// to create the FailureDetectorPuppet in the FailureDetectorRegistry
|
||||
cluster.failureDetector.heartbeat(address)
|
||||
failureDetectorPuppet(address) foreach (_.markNodeAsUnavailable())
|
||||
|
|
|
|||
|
|
@ -94,7 +94,7 @@ $.extend($.effects, {
|
|||
// The main function to form the object for animation
|
||||
for(var n in newStyle) {
|
||||
if( typeof newStyle[n] != "function" && newStyle[n] /* No functions and null properties */
|
||||
&& n.indexOf("Moz") == -1 && n.indexOf("length") == -1 /* No mozilla spezific render properties. */
|
||||
&& n.indexOf("Moz") == -1 && n.indexOf("length") == -1 /* No mozilla specific render properties. */
|
||||
&& newStyle[n] != oldStyle[n] /* Only values that have changed are used for the animation */
|
||||
&& (n.match(/color/i) || (!n.match(/color/i) && !isNaN(parseInt(newStyle[n],10)))) /* Only things that can be parsed to integers or colors */
|
||||
&& (oldStyle.position != "static" || (oldStyle.position == "static" && !n.match(/left|top|bottom|right/))) /* No need for positions when dealing with static positions */
|
||||
|
|
|
|||
|
|
@ -63,7 +63,7 @@ abstract class ActorSystemActivator extends BundleActivator {
|
|||
val filter = s"(objectclass=${classOf[LogService].getName})"
|
||||
context.addServiceListener(logServiceListner, filter)
|
||||
|
||||
//Small trick to create an event if the service is registred before this start listing for
|
||||
//Small trick to create an event if the service is registered before this start listing for
|
||||
Option(context.getServiceReference(classOf[LogService].getName)).foreach(x ⇒ {
|
||||
logServiceListner.serviceChanged(new ServiceEvent(ServiceEvent.REGISTERED, x))
|
||||
})
|
||||
|
|
|
|||
|
|
@ -47,8 +47,8 @@ class DefaultOSGiLogger extends DefaultLogger {
|
|||
}
|
||||
|
||||
/**
|
||||
* Behaviour of the Eventhanlder that is setup (has received a LogService)
|
||||
* @param logService registrered OSGi LogService
|
||||
* Behaviour of the Event handler that is setup (has received a LogService)
|
||||
* @param logService registered OSGi LogService
|
||||
* @return Receive : Logs LogEvent or go back to the uninitialised state
|
||||
*/
|
||||
def initialisedReceive(logService: LogService): Receive = {
|
||||
|
|
@ -65,7 +65,7 @@ class DefaultOSGiLogger extends DefaultLogger {
|
|||
* Logs a message in an OSGi LogService
|
||||
*
|
||||
* @param logService OSGi LogService registered and used for logging
|
||||
* @param event akka LogEvent that is log unsing the LogService
|
||||
* @param event akka LogEvent that is logged using the LogService
|
||||
*/
|
||||
def logMessage(logService: LogService, event: LogEvent) {
|
||||
event match {
|
||||
|
|
@ -79,7 +79,7 @@ class DefaultOSGiLogger extends DefaultLogger {
|
|||
}
|
||||
|
||||
/**
|
||||
* Message sent when LogService is unregistred.
|
||||
* Message sent when LogService is unregistered.
|
||||
* Sent from the ActorSystemActivator to a logger (as DefaultOsgiLogger).
|
||||
*/
|
||||
case object UnregisteringLogService
|
||||
|
|
|
|||
|
|
@ -18,7 +18,7 @@ abstract class PluginSpec(val config: Config) extends TestKitBase with WordSpecL
|
|||
private var _pid: String = _
|
||||
|
||||
// used to avoid messages be delivered to a restarted actor,
|
||||
// this is akka-persistence internals and journals themselfes don't really care
|
||||
// this is akka-persistence internals and journals themselves don't really care
|
||||
protected val actorInstanceId = 1
|
||||
|
||||
override protected def beforeEach(): Unit =
|
||||
|
|
|
|||
|
|
@ -20,7 +20,7 @@ import org.scalatest.junit.JUnitRunner
|
|||
* It is *NOT* meant to be a comprehensive benchmark, but rather aims to help plugin developers to easily determine
|
||||
* if their plugin's performance is roughly as expected. It also validates the plugin still works under "more messages" scenarios.
|
||||
*
|
||||
* The measurements are by default printed to `System.out`, if you want to customise this please override the [[#info]] method.
|
||||
* The measurements are by default printed to `System.out`, if you want to customize this please override the [[#info]] method.
|
||||
*
|
||||
* The benchmark iteration and message counts are easily customisable by overriding these methods:
|
||||
*
|
||||
|
|
@ -36,7 +36,7 @@ import org.scalatest.junit.JUnitRunner
|
|||
* }}}
|
||||
*
|
||||
* In case your journal plugin needs some kind of setup or teardown, override the `beforeAll` or `afterAll`
|
||||
* methods (don't forget to call `super` in your overriden methods).
|
||||
* methods (don't forget to call `super` in your overridden methods).
|
||||
*
|
||||
* @see [[akka.persistence.journal.JournalSpec]]
|
||||
* @see [[akka.persistence.journal.JournalPerfSpec]]
|
||||
|
|
|
|||
|
|
@ -14,7 +14,7 @@ import org.scalatest.junit.JUnitRunner
|
|||
* Java / JUnit API for [[akka.persistence.journal.JournalSpec]].
|
||||
*
|
||||
* In case your journal plugin needs some kind of setup or teardown, override the `beforeAll` or `afterAll`
|
||||
* methods (don't forget to call `super` in your overriden methods).
|
||||
* methods (don't forget to call `super` in your overridden methods).
|
||||
*
|
||||
* @see [[akka.persistence.journal.JournalSpec]]
|
||||
* @see [[akka.persistence.journal.JournalPerfSpec]]
|
||||
|
|
|
|||
|
|
@ -15,7 +15,7 @@ import org.scalatest.junit.JUnitRunner
|
|||
* Plugin authors are highly encouraged to include it in their plugin's test suites.
|
||||
*
|
||||
* In case your snapshot-store plugin needs some kind of setup or teardown, override the `beforeAll` or `afterAll`
|
||||
* methods (don't forget to call `super` in your overriden methods).
|
||||
* methods (don't forget to call `super` in your overridden methods).
|
||||
*
|
||||
* @see [[akka.persistence.snapshot.SnapshotStoreSpec]]
|
||||
*/
|
||||
|
|
|
|||
|
|
@ -69,7 +69,7 @@ object JournalPerfSpec {
|
|||
* if their plugin's performance is roughly as expected. It also validates the plugin still works under "more messages" scenarios.
|
||||
*
|
||||
* In case your journal plugin needs some kind of setup or teardown, override the `beforeAll` or `afterAll`
|
||||
* methods (don't forget to call `super` in your overriden methods).
|
||||
* methods (don't forget to call `super` in your overridden methods).
|
||||
*
|
||||
* For a Java and JUnit consumable version of the TCK please refer to [[akka.persistence.japi.journal.JavaJournalPerfSpec]].
|
||||
*
|
||||
|
|
@ -87,7 +87,7 @@ abstract class JournalPerfSpec(config: Config) extends JournalSpec(config) {
|
|||
testProbe.expectMsg(awaitDuration, cmnds.last)
|
||||
}
|
||||
|
||||
/** Executes a block of code multiple times (no warmup) */
|
||||
/** Executes a block of code multiple times (no warm-up) */
|
||||
def measure(msg: Duration ⇒ String)(block: ⇒ Unit): Unit = {
|
||||
val measurements = Array.ofDim[Duration](measurementIterations)
|
||||
var i = 0
|
||||
|
|
@ -112,7 +112,7 @@ abstract class JournalPerfSpec(config: Config) extends JournalSpec(config) {
|
|||
/** Override in order to customize timeouts used for expectMsg, in order to tune the awaits to your journal's perf */
|
||||
private def awaitDuration: FiniteDuration = awaitDurationMillis.millis
|
||||
|
||||
/** Numbe of messages sent to the PersistentActor under test for each test iteration */
|
||||
/** Number of messages sent to the PersistentActor under test for each test iteration */
|
||||
def eventsCount: Int = 10 * 1000
|
||||
|
||||
/** Number of measurement iterations each test will be run. */
|
||||
|
|
|
|||
|
|
@ -21,7 +21,7 @@ object JournalSpec {
|
|||
* Plugin authors are highly encouraged to include it in their plugin's test suites.
|
||||
*
|
||||
* In case your journal plugin needs some kind of setup or teardown, override the `beforeAll` or `afterAll`
|
||||
* methods (don't forget to call `super` in your overriden methods).
|
||||
* methods (don't forget to call `super` in your overridden methods).
|
||||
*
|
||||
* For a Java and JUnit consumable version of the TCK please refer to [[akka.persistence.japi.journal.JavaJournalSpec]].
|
||||
*
|
||||
|
|
|
|||
|
|
@ -17,7 +17,7 @@ object SnapshotStoreSpec {
|
|||
* Plugin authors are highly encouraged to include it in their plugin's test suites.
|
||||
*
|
||||
* In case your journal plugin needs some kind of setup or teardown, override the `beforeAll` or `afterAll`
|
||||
* methods (don't forget to call `super` in your overriden methods).
|
||||
* methods (don't forget to call `super` in your overridden methods).
|
||||
*
|
||||
* For a Java and JUnit consumable version of the TCK please refer to [[akka.persistence.japi.snapshot.JavaSnapshotStoreSpec]].
|
||||
*
|
||||
|
|
|
|||
|
|
@ -25,7 +25,7 @@ private[persistence] object Eventsourced {
|
|||
def evt: Any
|
||||
def handler: Any ⇒ Unit
|
||||
}
|
||||
/** forces actor to stash incoming commands untill all these invocations are handled */
|
||||
/** forces actor to stash incoming commands until all these invocations are handled */
|
||||
private final case class StashingHandlerInvocation(evt: Any, handler: Any ⇒ Unit) extends PendingHandlerInvocation
|
||||
/** does not force the actor to stash commands; Originates from either `persistAsync` or `defer` calls */
|
||||
private final case class AsyncHandlerInvocation(evt: Any, handler: Any ⇒ Unit) extends PendingHandlerInvocation
|
||||
|
|
|
|||
|
|
@ -111,7 +111,7 @@ abstract class RemoteNodeShutdownAndComesBackSpec
|
|||
expectTerminated(subject)
|
||||
|
||||
// Establish watch with the new system. This triggers additional system message traffic. If buffers are out
|
||||
// of synch the remote system will be quarantined and the rest of the test will fail (or even in earlier
|
||||
// of sync the remote system will be quarantined and the rest of the test will fail (or even in earlier
|
||||
// stages depending on circumstances).
|
||||
system.actorSelection(RootActorPath(secondAddress) / "user" / "subject") ! Identify("subject")
|
||||
val subjectNew = expectMsgType[ActorIdentity].ref.get
|
||||
|
|
|
|||
|
|
@ -89,8 +89,8 @@ abstract class Ticket15109Spec extends MultiNodeSpec(Ticket15109Spec)
|
|||
enterBarrier("actor-identified")
|
||||
|
||||
runOn(second) {
|
||||
// Force a dissassociation. Using the message Shutdown, which is suboptimal here, but this is the only
|
||||
// DisassoicateInfo that triggers the code-path we want to test
|
||||
// Force a disassociation. Using the message Shutdown, which is suboptimal here, but this is the only
|
||||
// DisassociateInfo that triggers the code-path we want to test
|
||||
Await.result(RARP(system).provider.transport.managementCommand(
|
||||
ForceDisassociateExplicitly(node(first).address, AssociationHandle.Shutdown)), 3.seconds)
|
||||
}
|
||||
|
|
|
|||
|
|
@ -134,7 +134,7 @@ class RemoteRoundRobinSpec extends MultiNodeSpec(RemoteRoundRobinMultiJvmSpec)
|
|||
actor.isInstanceOf[RoutedActorRef] should ===(true)
|
||||
|
||||
actor ! GetRoutees
|
||||
// initial nrOfInstances 1 + inital resize => 2
|
||||
// initial nrOfInstances 1 + initial resize => 2
|
||||
expectMsgType[Routees].routees.size should ===(2)
|
||||
|
||||
val repliesFrom: Set[ActorRef] =
|
||||
|
|
|
|||
|
|
@ -703,7 +703,7 @@ private[remote] class EndpointWriter(
|
|||
val backoffNanos = backoffDeadlinelineNanoTime - System.nanoTime
|
||||
if (backoffNanos > 0) {
|
||||
LockSupport.parkNanos(backoffNanos)
|
||||
// parkNanos allows for spurious wakeup, check again
|
||||
// parkNanos allows for spurious wake-up, check again
|
||||
backoff()
|
||||
}
|
||||
}
|
||||
|
|
@ -791,7 +791,7 @@ private[remote] class EndpointWriter(
|
|||
case Some(r) ⇒
|
||||
r.tell(s, replyTo)
|
||||
case None ⇒
|
||||
// initalizing, buffer and take care of it later when buffer is sent
|
||||
// initializing, buffer and take care of it later when buffer is sent
|
||||
enqueueInBuffer(s)
|
||||
}
|
||||
case TakeOver(newHandle, replyTo) ⇒
|
||||
|
|
|
|||
|
|
@ -32,7 +32,7 @@ trait FailureDetectorRegistry[A] {
|
|||
|
||||
/**
|
||||
* Records a heartbeat for a resource. If the resource is not yet registered (i.e. this is the first heartbeat) then
|
||||
* it is automatially registered.
|
||||
* it is automatically registered.
|
||||
*/
|
||||
def heartbeat(resource: A): Unit
|
||||
|
||||
|
|
|
|||
|
|
@ -186,7 +186,7 @@ private[akka] object HeartbeatHistory {
|
|||
* Create an empty HeartbeatHistory, without any history.
|
||||
* Can only be used as starting point for appending intervals.
|
||||
* The stats (mean, variance, stdDeviation) are not defined for
|
||||
* for empty HeartbeatHistory, i.e. throws AritmeticException.
|
||||
* for empty HeartbeatHistory, i.e. throws ArithmeticException.
|
||||
*/
|
||||
def apply(maxSampleSize: Int): HeartbeatHistory = HeartbeatHistory(
|
||||
maxSampleSize = maxSampleSize,
|
||||
|
|
@ -201,7 +201,7 @@ private[akka] object HeartbeatHistory {
|
|||
* It is capped by the number of samples specified in `maxSampleSize`.
|
||||
*
|
||||
* The stats (mean, variance, stdDeviation) are not defined for
|
||||
* for empty HeartbeatHistory, i.e. throws AritmeticException.
|
||||
* for empty HeartbeatHistory, i.e. throws ArithmeticException.
|
||||
*/
|
||||
private[akka] final case class HeartbeatHistory private (
|
||||
maxSampleSize: Int,
|
||||
|
|
|
|||
|
|
@ -47,7 +47,7 @@ private[akka] abstract class RemoteTransport(val system: ExtendedActorSystem, va
|
|||
def addresses: immutable.Set[Address]
|
||||
|
||||
/**
|
||||
* The default transport address of the actorsystem
|
||||
* The default transport address of the ActorSystem
|
||||
* @return The listen address of the default transport
|
||||
*/
|
||||
def defaultAddress: Address
|
||||
|
|
|
|||
|
|
@ -68,7 +68,7 @@ private[akka] object RemoteWatcher {
|
|||
* For a new node to be watched this actor periodically sends `RemoteWatcher.Heartbeat`
|
||||
* to the peer actor on the other node, which replies with [[RemoteWatcher.HeartbeatRsp]]
|
||||
* message back. The failure detector on the watching side monitors these heartbeat messages.
|
||||
* If arrival of hearbeat messages stops it will be detected and this actor will publish
|
||||
* If arrival of heartbeat messages stops it will be detected and this actor will publish
|
||||
* [[akka.actor.AddressTerminated]] to the [[akka.event.AddressTerminatedTopic]].
|
||||
*
|
||||
* When all actors on a node have been unwatched it will stop sending heartbeat messages.
|
||||
|
|
|
|||
|
|
@ -16,7 +16,7 @@ import scala.reflect.ClassTag
|
|||
import util.{ Failure, Success }
|
||||
|
||||
/**
|
||||
* Serializes akka's internal DaemonMsgCreate using protobuf
|
||||
* Serializes Akka's internal DaemonMsgCreate using protobuf
|
||||
* for the core structure of DaemonMsgCreate, Props and Deploy.
|
||||
* Serialization of contained RouterConfig, Config, and Scope
|
||||
* is done with configured serializer for those classes, by
|
||||
|
|
|
|||
|
|
@ -20,7 +20,7 @@ import scala.concurrent.ExecutionContext.Implicits.global
|
|||
*
|
||||
* The TestTransport is basically a shared memory between actor systems. The TestTransport could be programmed to
|
||||
* emulate different failure modes of a Transport implementation. TestTransport keeps a log of the activities it was
|
||||
* requested to do. This class is not optimized for performace and MUST not be used as an in-memory transport in
|
||||
* requested to do. This class is not optimized for performance and MUST not be used as an in-memory transport in
|
||||
* production systems.
|
||||
*/
|
||||
class TestTransport(
|
||||
|
|
@ -177,7 +177,7 @@ object TestTransport {
|
|||
/**
|
||||
* Test utility to make behavior of functions that return some Future[B] controllable from tests. This tool is able
|
||||
* to overwrite default behavior with any generic behavior, including failure, and exposes control to the timing of
|
||||
* the completition of the returned future.
|
||||
* the completion of the returned future.
|
||||
*
|
||||
* The utility is implemented as a stack of behaviors, where the behavior on the top of the stack represents the
|
||||
* currently active behavior. The bottom of the stack always contains the defaultBehavior which can not be popped
|
||||
|
|
|
|||
|
|
@ -145,13 +145,13 @@ object ThrottlerTransportAdapter {
|
|||
}
|
||||
|
||||
/**
|
||||
* Management Command to force dissocation of an address.
|
||||
* Management Command to force disassociation of an address.
|
||||
*/
|
||||
@SerialVersionUID(1L)
|
||||
final case class ForceDisassociate(address: Address)
|
||||
|
||||
/**
|
||||
* Management Command to force dissocation of an address with an explicit error.
|
||||
* Management Command to force disassociation of an address with an explicit error.
|
||||
*/
|
||||
@SerialVersionUID(1L)
|
||||
final case class ForceDisassociateExplicitly(address: Address, reason: DisassociateInfo)
|
||||
|
|
|
|||
|
|
@ -432,7 +432,7 @@ class NettyTransport(val settings: NettyTransportSettings, val system: ExtendedA
|
|||
} catch {
|
||||
case NonFatal(e) ⇒ {
|
||||
log.error("failed to bind to {}, shutting down Netty transport", address)
|
||||
try { shutdown() } catch { case NonFatal(e) ⇒ } // ingore possible exception during shutdown
|
||||
try { shutdown() } catch { case NonFatal(e) ⇒ } // ignore possible exception during shutdown
|
||||
throw e
|
||||
}
|
||||
}
|
||||
|
|
|
|||
|
|
@ -81,7 +81,7 @@ abstract class TransformationSampleSpec extends MultiNodeSpec(TransformationSamp
|
|||
val transformationFrontend = system.actorOf(Props[TransformationFrontend], name = "frontend")
|
||||
transformationFrontend ! new TransformationJob("hello")
|
||||
expectMsgPF() {
|
||||
// no backends yet, service unavailble
|
||||
// no backends yet, service unavailable
|
||||
case f: JobFailed =>
|
||||
}
|
||||
}
|
||||
|
|
|
|||
|
|
@ -79,7 +79,7 @@ abstract class TransformationSampleSpec extends MultiNodeSpec(TransformationSamp
|
|||
val transformationFrontend = system.actorOf(Props[TransformationFrontend], name = "frontend")
|
||||
transformationFrontend ! TransformationJob("hello")
|
||||
expectMsgPF() {
|
||||
// no backends yet, service unavailble
|
||||
// no backends yet, service unavailable
|
||||
case JobFailed(_, TransformationJob("hello")) =>
|
||||
}
|
||||
}
|
||||
|
|
|
|||
|
|
@ -54,7 +54,7 @@ public class DiningHakkersOnFsm {
|
|||
event((event, data) -> (event == Put) && (data.hakker == sender()), (event, data) ->
|
||||
goTo(CS.Available).using(new TakenBy(context().system().deadLetters()))));
|
||||
|
||||
// Initialze the chopstick
|
||||
// Initialize the chopstick
|
||||
initialize();
|
||||
}
|
||||
}
|
||||
|
|
|
|||
|
|
@ -46,7 +46,7 @@ class Chopstick extends Actor {
|
|||
}
|
||||
|
||||
/*
|
||||
* A hakker is an awesome dude or dudett who either thinks about hacking or has to eat ;-)
|
||||
* A hakker is an awesome dude or dudette who either thinks about hacking or has to eat ;-)
|
||||
*/
|
||||
class Hakker(name: String, left: ActorRef, right: ActorRef) extends Actor {
|
||||
|
||||
|
|
|
|||
|
|
@ -53,7 +53,7 @@ class Chopstick extends Actor with FSM[ChopstickState, TakenBy] {
|
|||
goto(Available) using TakenBy(system.deadLetters)
|
||||
}
|
||||
|
||||
// Initialze the chopstick
|
||||
// Initialize the chopstick
|
||||
initialize()
|
||||
}
|
||||
|
||||
|
|
|
|||
|
|
@ -46,7 +46,7 @@ class Chopstick extends Actor {
|
|||
}
|
||||
|
||||
/*
|
||||
* A hakker is an awesome dude or dudett who either thinks about hacking or has to eat ;-)
|
||||
* A hakker is an awesome dude or dudette who either thinks about hacking or has to eat ;-)
|
||||
*/
|
||||
class Hakker(name: String, chair: Int) extends Actor {
|
||||
|
||||
|
|
|
|||
|
|
@ -105,7 +105,7 @@ class TestActorRef[T <: Actor](
|
|||
def watch(subject: ActorRef): ActorRef = underlying.watch(subject)
|
||||
|
||||
/**
|
||||
* Deregisters this actor from being a death monitor of the provided ActorRef
|
||||
* Unregisters this actor from being a death monitor of the provided ActorRef
|
||||
* This means that this actor will not get a Terminated()-message when the provided actor
|
||||
* is permanently terminated.
|
||||
*
|
||||
|
|
|
|||
|
|
@ -20,7 +20,7 @@ import akka.util.BoxedType
|
|||
/**
|
||||
* Implementation helpers of the EventFilter facilities: send `Mute`
|
||||
* to the TestEventListener to install a filter, and `UnMute` to
|
||||
* deinstall it.
|
||||
* uninstall it.
|
||||
*
|
||||
* You should always prefer the filter methods in the package object
|
||||
* (see [[akka.testkit]] `filterEvents` and `filterException`) or on the
|
||||
|
|
@ -31,7 +31,7 @@ sealed trait TestEvent
|
|||
/**
|
||||
* Implementation helpers of the EventFilter facilities: send <code>Mute</code>
|
||||
* to the TestEventFilter to install a filter, and <code>UnMute</code> to
|
||||
* deinstall it.
|
||||
* uninstall it.
|
||||
*
|
||||
* You should always prefer the filter methods in the package object
|
||||
* (see [[akka.testkit]] `filterEvents` and `filterException`) or on the
|
||||
|
|
|
|||
|
|
@ -13,7 +13,7 @@ import akka.dispatch.MailboxType
|
|||
import scala.reflect.ClassTag
|
||||
|
||||
/**
|
||||
* This is a specialised form of the TestActorRef with support for querying and
|
||||
* This is a specialized form of the TestActorRef with support for querying and
|
||||
* setting the state of a FSM. Use a LoggingFSM with this class if you also
|
||||
* need to inspect event traces.
|
||||
*
|
||||
|
|
|
|||
|
|
@ -389,7 +389,7 @@ trait TestKitBase {
|
|||
* partial function matches and returns false. Use it to ignore certain
|
||||
* messages while waiting for a specific message.
|
||||
*
|
||||
* @return the last received messsage, i.e. the first one for which the
|
||||
* @return the last received message, i.e. the first one for which the
|
||||
* partial function returned true
|
||||
*/
|
||||
def fishForMessage(max: Duration = Duration.Undefined, hint: String = "")(f: PartialFunction[Any, Boolean]): Any = {
|
||||
|
|
|
|||
|
|
@ -8,7 +8,7 @@ import java.util.concurrent.atomic.AtomicLong
|
|||
import java.util.concurrent.TimeUnit._
|
||||
|
||||
/**
|
||||
* Specialised "one-shot" Timer.
|
||||
* Specialized "one-shot" Timer.
|
||||
* Given a known number of operations performed within a time span (to be measured) it displays the average time one operation took.
|
||||
*
|
||||
* Please note that this is a *very coarse* estimation; The gain though is that we do not have to perform the counting inside of the measured thing (we can adding in tight loops).
|
||||
|
|
|
|||
|
|
@ -27,7 +27,7 @@ private[akka] trait MetricsKitOps extends MetricKeyDSL {
|
|||
|
||||
/**
|
||||
* Used to measure timing of known number of operations over time.
|
||||
* While not being the most percise, it allows to measure a coarse op/s without injecting counters to the measured operation (potentially hot-loop).
|
||||
* While not being the most precise, it allows to measure a coarse op/s without injecting counters to the measured operation (potentially hot-loop).
|
||||
*
|
||||
* Do not use for short running pieces of code.
|
||||
*/
|
||||
|
|
|
|||
|
|
@ -104,7 +104,7 @@ final case class PostRestart(failure: Throwable) extends Signal
|
|||
*
|
||||
* <b>IMPORTANT NOTE:</b> if the actor terminated by switching to the
|
||||
* `Stopped` behavior then this signal will be ignored (i.e. the
|
||||
* Stopped behvavior will do nothing in reaction to it).
|
||||
* Stopped behavior will do nothing in reaction to it).
|
||||
*/
|
||||
@SerialVersionUID(1L)
|
||||
final case object PostStop extends Signal
|
||||
|
|
|
|||
|
|
@ -53,7 +53,7 @@ object Receptionist {
|
|||
final case class Find[T](key: ServiceKey[T])(val replyTo: ActorRef[Listing[T]]) extends Command
|
||||
|
||||
/**
|
||||
* Confirmtion that the given [[akka.typed.ActorRef]] has been associated with the [[ServiceKey]].
|
||||
* Confirmation that the given [[akka.typed.ActorRef]] has been associated with the [[ServiceKey]].
|
||||
*/
|
||||
final case class Registered[T](key: ServiceKey[T], address: ActorRef[T])
|
||||
/**
|
||||
|
|
|
|||
|
|
@ -48,7 +48,7 @@ object OSGi {
|
|||
val osgiOptionalImports = Seq(
|
||||
// needed because testkit is normally not used in the application bundle,
|
||||
// but it should still be included as transitive dependency and used by BundleDelegatingClassLoader
|
||||
// to be able to find refererence.conf
|
||||
// to be able to find reference.conf
|
||||
"akka.testkit",
|
||||
"com.google.protobuf")
|
||||
|
||||
|
|
|
|||
|
|
@ -41,7 +41,7 @@ object Sample {
|
|||
val dependencies = buildDependencies.value
|
||||
val classpathWithProjectDependencies = dependencies.classpath.map {
|
||||
case (proj, deps) if proj.project == project.id =>
|
||||
// add project dependency for every akka library dependnecy
|
||||
// add project dependency for every akka library dependency
|
||||
(proj, deps ++ projectDependencies.map(ResolvedClasspathDependency(_, None)))
|
||||
case (project, deps) => (project, deps)
|
||||
}
|
||||
|
|
|
|||
Loading…
Add table
Add a link
Reference in a new issue