update some random akka refs (#196)

This commit is contained in:
PJ Fanning 2023-02-16 10:39:18 +01:00 committed by GitHub
parent 5e8a278b1b
commit 3ee028f915
No known key found for this signature in database
GPG key ID: 4AEE18F83AFDEB23
49 changed files with 175 additions and 172 deletions

View file

@ -79,7 +79,7 @@ class PrimitivesSerializationSpec extends PekkoSpec(PrimitivesSerializationSpec.
}
"have right serializer id" in {
// checking because moved to akka-actor
// checking because moved to pekko-actor
serialization.serializerFor(1L.asInstanceOf[AnyRef].getClass).identifier === 18
}
@ -102,7 +102,7 @@ class PrimitivesSerializationSpec extends PekkoSpec(PrimitivesSerializationSpec.
}
"have right serializer id" in {
// checking because moved to akka-actor
// checking because moved to pekko-actor
serialization.serializerFor(1L.asInstanceOf[AnyRef].getClass).identifier === 19
}
}
@ -124,7 +124,7 @@ class PrimitivesSerializationSpec extends PekkoSpec(PrimitivesSerializationSpec.
}
"have right serializer id ($i)" in {
// checking because moved to akka-actor
// checking because moved to pekko-actor
serialization.serializerFor(true.asInstanceOf[AnyRef].getClass).identifier === 35
}
}
@ -148,7 +148,7 @@ class PrimitivesSerializationSpec extends PekkoSpec(PrimitivesSerializationSpec.
}
"have right serializer id" in {
// checking because moved to akka-actor
// checking because moved to pekko-actor
serialization.serializerFor(1L.asInstanceOf[AnyRef].getClass).identifier === 20
}
@ -178,7 +178,7 @@ class PrimitivesSerializationSpec extends PekkoSpec(PrimitivesSerializationSpec.
}
"have right serializer id" in {
// checking because moved to akka-actor
// checking because moved to pekko-actor
serialization.serializerFor(1L.asInstanceOf[AnyRef].getClass).identifier === 21
}

View file

@ -276,7 +276,7 @@ class SerializeSpec extends PekkoSpec(SerializationTests.serializeConf) {
}
"log warning if non-Pekko serializer is configured for Pekko message" in {
EventFilter.warning(pattern = ".*not implemented by Pekko.*", occurrences = 1).intercept {
EventFilter.warning(pattern = ".*not implemented by Apache Pekko.*", occurrences = 1).intercept {
ser.serialize(new Other).get
}
}

View file

@ -576,7 +576,7 @@ class BoundedBlockingQueueSpec
val TestContext(queue, _, _, _, _, _) = newBoundedBlockingQueue(2)
val elems = List("Hello", "World")
elems.foreach(queue.put)
queue.containsAll(("Akka" :: elems).asJava) should equal(false)
queue.containsAll(("Pekko" :: elems).asJava) should equal(false)
}
"returns true if the queue contains exactly all elements" in {
@ -590,7 +590,7 @@ class BoundedBlockingQueueSpec
val TestContext(queue, _, _, _, _, _) = newBoundedBlockingQueue(3)
val elems = List("Hello", "World")
elems.foreach(queue.put)
queue.put("Akka")
queue.put("Pekko")
queue.containsAll(elems.asJava) should equal(true)
}
}
@ -600,20 +600,20 @@ class BoundedBlockingQueueSpec
val TestContext(queue, _, _, _, _, _) = newBoundedBlockingQueue(3)
val elems = List("Hello", "World")
elems.foreach(queue.put)
queue.put("Akka")
queue.put("Pekko")
queue.removeAll(elems.asJava) should equal(true)
queue.remainingCapacity() should equal(2)
queue.poll() should equal("Akka")
queue.poll() should equal("Pekko")
}
"return false if no elements were removed" in {
val TestContext(queue, _, _, _, _, _) = newBoundedBlockingQueue(1)
val elems = List("Hello", "World")
queue.put("Akka")
queue.put("Pekko")
queue.removeAll(elems.asJava) should equal(false)
queue.poll() should equal("Akka")
queue.poll() should equal("Pekko")
}
}
@ -622,17 +622,17 @@ class BoundedBlockingQueueSpec
val TestContext(queue, _, _, _, _, _) = newBoundedBlockingQueue(3)
val elems = List("Hello", "World")
elems.foreach(queue.put)
queue.put("Akka")
queue.put("Pekko")
queue.retainAll(elems.asJava) should equal(true)
queue.remainingCapacity() should equal(1)
queue.toArray() shouldNot contain("Akka")
queue.toArray() shouldNot contain("Pekko")
queue.toArray() should contain theSameElementsAs elems
}
"return false if no elements were removed" in {
val TestContext(queue, _, _, _, _, _) = newBoundedBlockingQueue(2)
val elems = List("Hello", "World", "Akka")
val elems = List("Hello", "World", "Pekko")
queue.put("Hello")
queue.put("World")
@ -650,7 +650,7 @@ class BoundedBlockingQueueSpec
"disallow remove() before calling next()" in {
val TestContext(queue, _, _, _, _, _) = newBoundedBlockingQueue(2)
queue.put("Hello")
queue.put("Akka")
queue.put("Pekko")
val iter = queue.iterator()
intercept[IllegalStateException] {
@ -661,7 +661,7 @@ class BoundedBlockingQueueSpec
"disallow two subsequent remove()s" in {
val TestContext(queue, _, _, _, _, _) = newBoundedBlockingQueue(2)
queue.put("Hello")
queue.put("Akka")
queue.put("Pekko")
val iter = queue.iterator()
iter.next()

View file

@ -201,7 +201,7 @@ public interface IntroTest {
ActorSystem.create(HelloWorldMain.create(), "hello");
system.tell(new HelloWorldMain.SayHello("World"));
system.tell(new HelloWorldMain.SayHello("Akka"));
system.tell(new HelloWorldMain.SayHello("Pekko"));
// #hello-world
Thread.sleep(3000);

View file

@ -85,7 +85,7 @@ public interface SpawnProtocolDocTest {
greetedReplyTo.whenComplete(
(greetedReplyToRef, exc2) -> {
if (exc2 == null) {
greeterRef.tell(new HelloWorld.Greet("Akka", greetedReplyToRef));
greeterRef.tell(new HelloWorld.Greet("Pekko", greetedReplyToRef));
}
});
}

View file

@ -103,7 +103,7 @@ object IntroSpec {
ActorSystem(HelloWorldMain(), "hello")
system ! HelloWorldMain.SayHello("World")
system ! HelloWorldMain.SayHello("Akka")
system ! HelloWorldMain.SayHello("Pekko")
}
//#hello-world-main
}
@ -260,7 +260,7 @@ class IntroSpec extends ScalaTestWithActorTestKit with AnyWordSpecLike with LogC
ActorSystem(HelloWorldMain(), "hello")
system ! HelloWorldMain.SayHello("World")
system ! HelloWorldMain.SayHello("Akka")
system ! HelloWorldMain.SayHello("Pekko")
// #hello-world

View file

@ -87,7 +87,7 @@ class SpawnProtocolDocSpec extends ScalaTestWithActorTestKit with AnyWordSpecLik
system.ask(SpawnProtocol.Spawn(greetedBehavior, name = "", props = Props.empty, _))
for (greeterRef <- greeter; replyToRef <- greetedReplyTo) {
greeterRef ! HelloWorld.Greet("Akka", replyToRef)
greeterRef ! HelloWorld.Greet("Pekko", replyToRef)
}
// #system-spawn

View file

@ -28,21 +28,21 @@ import org.apache.pekko.annotation.InternalApi
@InternalApi
@Enabled(true)
@StackTrace(false)
@Category(Array("Akka", "Delivery", "ProducerController")) @Label("Delivery ProducerController created")
@Category(Array("Pekko", "Delivery", "ProducerController")) @Label("Delivery ProducerController created")
final class DeliveryProducerCreated(val producerId: String, val actorPath: String) extends Event
/** INTERNAL API */
@InternalApi
@Enabled(true)
@StackTrace(false)
@Category(Array("Akka", "Delivery", "ProducerController")) @Label("Delivery ProducerController started")
@Category(Array("Pekko", "Delivery", "ProducerController")) @Label("Delivery ProducerController started")
final class DeliveryProducerStarted(val producerId: String, val actorPath: String) extends Event
/** INTERNAL API */
@InternalApi
@Enabled(false) // hi frequency event
@StackTrace(false)
@Category(Array("Akka", "Delivery", "ProducerController")) @Label("Delivery ProducerController sent RequestNext")
@Category(Array("Pekko", "Delivery", "ProducerController")) @Label("Delivery ProducerController sent RequestNext")
final class DeliveryProducerRequestNext(val producerId: String, val currentSeqNr: Long, val confirmedSeqNr: Long)
extends Event
@ -50,21 +50,21 @@ final class DeliveryProducerRequestNext(val producerId: String, val currentSeqNr
@InternalApi
@Enabled(false) // hi frequency event
@StackTrace(false)
@Category(Array("Akka", "Delivery", "ProducerController")) @Label("Delivery ProducerController sent SequencedMessage")
@Category(Array("Pekko", "Delivery", "ProducerController")) @Label("Delivery ProducerController sent SequencedMessage")
final class DeliveryProducerSent(val producerId: String, val seqNr: Long) extends Event
/** INTERNAL API */
@InternalApi
@Enabled(true)
@StackTrace(false)
@Category(Array("Akka", "Delivery", "ProducerController")) @Label("Delivery ProducerController waiting for demand")
@Category(Array("Pekko", "Delivery", "ProducerController")) @Label("Delivery ProducerController waiting for demand")
final class DeliveryProducerWaitingForRequest(val producerId: String, val currentSeqNr: Long) extends Event
/** INTERNAL API */
@InternalApi
@Enabled(true)
@StackTrace(false)
@Category(Array("Akka", "Delivery", "ProducerController")) @Label("Delivery ProducerController resent unconfirmed")
@Category(Array("Pekko", "Delivery", "ProducerController")) @Label("Delivery ProducerController resent unconfirmed")
final class DeliveryProducerResentUnconfirmed(val producerId: String, val fromSeqNr: Long, val toSeqNr: Long)
extends Event
@ -72,14 +72,14 @@ final class DeliveryProducerResentUnconfirmed(val producerId: String, val fromSe
@InternalApi
@Enabled(true)
@StackTrace(false)
@Category(Array("Akka", "Delivery", "ProducerController")) @Label("Delivery ProducerController resent first")
@Category(Array("Pekko", "Delivery", "ProducerController")) @Label("Delivery ProducerController resent first")
final class DeliveryProducerResentFirst(val producerId: String, val firstSeqNr: Long) extends Event
/** INTERNAL API */
@InternalApi
@Enabled(true)
@StackTrace(false)
@Category(Array("Akka", "Delivery", "ProducerController")) @Label(
@Category(Array("Pekko", "Delivery", "ProducerController")) @Label(
"Delivery ProducerController resent first unconfirmed")
final class DeliveryProducerResentFirstUnconfirmed(val producerId: String, val seqNr: Long) extends Event
@ -87,14 +87,15 @@ final class DeliveryProducerResentFirstUnconfirmed(val producerId: String, val s
@InternalApi
@Enabled(false) // hi frequency event
@StackTrace(false)
@Category(Array("Akka", "Delivery", "ProducerController")) @Label("Delivery ProducerController received message")
@Category(Array("Pekko", "Delivery", "ProducerController")) @Label("Delivery ProducerController received message")
final class DeliveryProducerReceived(val producerId: String, val currentSeqNr: Long) extends Event
/** INTERNAL API */
@InternalApi
@Enabled(true)
@StackTrace(false)
@Category(Array("Akka", "Delivery", "ProducerController")) @Label("Delivery ProducerController received demand request")
@Category(Array("Pekko", "Delivery", "ProducerController")) @Label(
"Delivery ProducerController received demand request")
final class DeliveryProducerReceivedRequest(val producerId: String, val requestedSeqNr: Long, confirmedSeqNr: Long)
extends Event
@ -102,35 +103,36 @@ final class DeliveryProducerReceivedRequest(val producerId: String, val requeste
@InternalApi
@Enabled(true)
@StackTrace(false)
@Category(Array("Akka", "Delivery", "ProducerController")) @Label("Delivery ProducerController received resend request")
@Category(Array("Pekko", "Delivery", "ProducerController")) @Label(
"Delivery ProducerController received resend request")
final class DeliveryProducerReceivedResend(val producerId: String, val fromSeqNr: Long) extends Event
/** INTERNAL API */
@InternalApi
@Enabled(true)
@StackTrace(false)
@Category(Array("Akka", "Delivery", "ConsumerController")) @Label("Delivery ConsumerController created")
@Category(Array("Pekko", "Delivery", "ConsumerController")) @Label("Delivery ConsumerController created")
final class DeliveryConsumerCreated(val actorPath: String) extends Event
/** INTERNAL API */
@InternalApi
@Enabled(true)
@StackTrace(false)
@Category(Array("Akka", "Delivery", "ConsumerController")) @Label("Delivery ConsumerController started")
@Category(Array("Pekko", "Delivery", "ConsumerController")) @Label("Delivery ConsumerController started")
final class DeliveryConsumerStarted(val actorPath: String) extends Event
/** INTERNAL API */
@InternalApi
@Enabled(false) // hi frequency event
@StackTrace(false)
@Category(Array("Akka", "Delivery", "ConsumerController")) @Label("Delivery ConsumerController received")
@Category(Array("Pekko", "Delivery", "ConsumerController")) @Label("Delivery ConsumerController received")
final class DeliveryConsumerReceived(val producerId: String, val seqNr: Long) extends Event
/** INTERNAL API */
@InternalApi
@Enabled(false) // hi frequency event
@StackTrace(false)
@Category(Array("Akka", "Delivery", "ConsumerController")) @Label(
@Category(Array("Pekko", "Delivery", "ConsumerController")) @Label(
"Delivery ConsumerController received, previous in progress")
final class DeliveryConsumerReceivedPreviousInProgress(val producerId: String, val seqNr: Long, val stashed: Int)
extends Event
@ -139,21 +141,21 @@ final class DeliveryConsumerReceivedPreviousInProgress(val producerId: String, v
@InternalApi
@Enabled(true)
@StackTrace(false)
@Category(Array("Akka", "Delivery", "ConsumerController")) @Label("Delivery ConsumerController received duplicate")
@Category(Array("Pekko", "Delivery", "ConsumerController")) @Label("Delivery ConsumerController received duplicate")
final class DeliveryConsumerDuplicate(val producerId: String, val expectedSeqNr: Long, val seqNr: Long) extends Event
/** INTERNAL API */
@InternalApi
@Enabled(true)
@StackTrace(false)
@Category(Array("Akka", "Delivery", "ConsumerController")) @Label("Delivery ConsumerController received missing")
@Category(Array("Pekko", "Delivery", "ConsumerController")) @Label("Delivery ConsumerController received missing")
final class DeliveryConsumerMissing(val producerId: String, val expectedSeqNr: Long, val seqNr: Long) extends Event
/** INTERNAL API */
@InternalApi
@Enabled(true)
@StackTrace(false)
@Category(Array("Akka", "Delivery", "ConsumerController")) @Label(
@Category(Array("Pekko", "Delivery", "ConsumerController")) @Label(
"Delivery ConsumerController received expected resend")
final class DeliveryConsumerReceivedResend(val seqNr: Long) extends Event
@ -161,19 +163,19 @@ final class DeliveryConsumerReceivedResend(val seqNr: Long) extends Event
@InternalApi
@Enabled(true)
@StackTrace(false)
@Category(Array("Akka", "Delivery", "ConsumerController")) @Label("Delivery ConsumerController sent demand Request")
@Category(Array("Pekko", "Delivery", "ConsumerController")) @Label("Delivery ConsumerController sent demand Request")
final class DeliveryConsumerSentRequest(val producerId: String, val requestedSeqNr: Long) extends Event
/** INTERNAL API */
@InternalApi
@Enabled(true)
@StackTrace(false)
@Category(Array("Akka", "Delivery", "ConsumerController")) @Label("Delivery ConsumerController producer changed")
@Category(Array("Pekko", "Delivery", "ConsumerController")) @Label("Delivery ConsumerController producer changed")
final class DeliveryConsumerChangedProducer(val producerId: String) extends Event
/** INTERNAL API */
@InternalApi
@Enabled(true)
@StackTrace(false)
@Category(Array("Akka", "Delivery", "ConsumerController")) @Label("Delivery ConsumerController stash is full")
@Category(Array("Pekko", "Delivery", "ConsumerController")) @Label("Delivery ConsumerController stash is full")
final class DeliveryConsumerStashFull(val producerId: String, val seqNr: Long) extends Event

View file

@ -68,7 +68,7 @@ import pekko.util.JavaDurationConverters._
* To make sure the messages can be delivered also in that scenario the `ProducerController` can be
* used with a [[DurableProducerQueue]]. Then the unconfirmed messages are stored in a durable way so
* that they can be redelivered when the producer is started again. An implementation of the
* `DurableProducerQueue` is provided by `EventSourcedProducerQueue` in `akka-persistence-typed`.
* `DurableProducerQueue` is provided by `EventSourcedProducerQueue` in `pekko-persistence-typed`.
*
* Instead of using `tell` with the `sendNextTo` in the `RequestNext` the producer can use `context.ask`
* with the `askNextTo` in the `RequestNext`. The difference is that a reply is sent back when the

View file

@ -83,7 +83,7 @@ import pekko.util.JavaDurationConverters._
* To make sure the messages can be delivered also in that scenario the `WorkPullingProducerController` can be
* used with a [[DurableProducerQueue]]. Then the unconfirmed messages are stored in a durable way so
* that they can be redelivered when the producer is started again. An implementation of the
* `DurableProducerQueue` is provided by `EventSourcedProducerQueue` in `akka-persistence-typed`.
* `DurableProducerQueue` is provided by `EventSourcedProducerQueue` in `pekko-persistence-typed`.
*
* Instead of using `tell` with the `sendNextTo` in the `RequestNext` the producer can use `context.ask`
* with the `askNextTo` in the `RequestNext`. The difference is that a reply is sent back when the

View file

@ -45,7 +45,7 @@ private[pekko] object TopicImpl {
final case class Subscribe[T](subscriber: ActorRef[T]) extends Topic.Command[T]
final case class Unsubscribe[T](subscriber: ActorRef[T]) extends Topic.Command[T]
// internal messages, note that the protobuf serializer for those sent remotely is defined in akka-cluster-typed
// internal messages, note that the protobuf serializer for those sent remotely is defined in pekko-cluster-typed
final case class GetTopicStats[T](replyTo: ActorRef[TopicStats]) extends Topic.Command[T]
final case class TopicStats(localSubscriberCount: Int, topicInstanceCount: Int) extends Topic.TopicStats
final case class TopicInstancesUpdated[T](topics: Set[ActorRef[TopicImpl.Command[T]]]) extends Command[T]

View file

@ -334,8 +334,8 @@ object ActorSystem {
* INTERNAL API
*
* When using Akka Typed the Slf4jLogger should be used by default.
* Looking for config property `pekko.use-slf4j` (defined in akka-actor-typed) and
* that `Slf4jLogger` (akka-slf4j) is in classpath.
* Looking for config property `pekko.use-slf4j` (defined in pekko-actor-typed) and
* that `Slf4jLogger` (pekko-slf4j) is in classpath.
* Then adds `Slf4jLogger` to configured loggers and removes `DefaultLogger`.
*/
@InternalApi private[pekko] def amendSlf4jConfig(config: Config, dynamicAccess: DynamicAccess): Config = {

View file

@ -1683,7 +1683,7 @@ trait DiagnosticLoggingAdapter extends LoggingAdapter {
def clearMDC(): Unit = mdc(emptyMDC)
}
/** DO NOT INHERIT: Class is open only for use by akka-slf4j */
/** DO NOT INHERIT: Class is open only for use by pekko-slf4j */
@DoNotInherit
class LogMarker(val name: String, val properties: Map[String, Any]) {

View file

@ -369,7 +369,7 @@ class Serialization(val system: ExtendedActorSystem) extends Extension {
clazz.getName)
}
if (!warnUnexpectedNonAkkaSerializer(clazz, ser))
if (!warnUnexpectedNonPekkoSerializer(clazz, ser))
log.debug("Using serializer [{}] for message [{}]", ser.getClass.getName, clazz.getName)
ser
@ -447,7 +447,7 @@ class Serialization(val system: ExtendedActorSystem) extends Extension {
private[pekko] val bindings: immutable.Seq[ClassSerializer] = {
val fromConfig = for {
(className: String, alias: String) <- settings.SerializationBindings
if alias != "none" && checkGoogleProtobuf(className) && checkAkkaProtobuf(className)
if alias != "none" && checkGoogleProtobuf(className) && checkPekkoProtobuf(className)
} yield (system.dynamicAccess.getClassFor[Any](className).get, serializers(alias))
val fromSettings = serializerDetails.flatMap { detail =>
@ -456,7 +456,7 @@ class Serialization(val system: ExtendedActorSystem) extends Extension {
val result = sort(fromConfig ++ fromSettings)
ensureOnlyAllowedSerializers(result.iterator.map { case (_, ser) => ser })
result.foreach { case (clazz, ser) => warnUnexpectedNonAkkaSerializer(clazz, ser) }
result.foreach { case (clazz, ser) => warnUnexpectedNonPekkoSerializer(clazz, ser) }
result
}
@ -466,12 +466,12 @@ class Serialization(val system: ExtendedActorSystem) extends Extension {
}
}
private def warnUnexpectedNonAkkaSerializer(clazz: Class[_], ser: Serializer): Boolean = {
private def warnUnexpectedNonPekkoSerializer(clazz: Class[_], ser: Serializer): Boolean = {
if (clazz.getName.startsWith("org.apache.pekko.") && !ser.getClass.getName.startsWith("org.apache.pekko.")) {
log.warning(
"Using serializer [{}] for message [{}]. Note that this serializer " +
"is not implemented by Pekko. It's not recommended to replace serializers for messages " +
"provided by Pekko.",
"is not implemented by Apache Pekko. It's not recommended to replace serializers for messages " +
"provided by Apache Pekko.",
ser.getClass.getName,
clazz.getName)
true
@ -484,9 +484,9 @@ class Serialization(val system: ExtendedActorSystem) extends Extension {
// include "com.google.protobuf.GeneratedMessage" = proto in configured serialization-bindings.
private def checkGoogleProtobuf(className: String): Boolean = checkClass("com.google.protobuf", className)
// akka-protobuf is now not a dependency of remote so only load if user has explicitly added it
// remove in 2.7
private def checkAkkaProtobuf(className: String): Boolean = checkClass("org.apache.pekko.protobuf", className)
// pekko-protobuf is now not a dependency of remote so only load if user has explicitly added it
// remove in v1.1
private def checkPekkoProtobuf(className: String): Boolean = checkClass("org.apache.pekko.protobuf", className)
private def checkClass(prefix: String, className: String): Boolean =
!className.startsWith(prefix) || system.dynamicAccess.getClassFor[Any](className).isSuccess

View file

@ -26,7 +26,7 @@ class ByteString_apply_Benchmark {
val bss = ByteStrings(Vector.fill(1024)(ByteString1(Array(0.toByte))))
/*
akka-bench-jmh/jmh:run -f 1 -wi 3 -i 3 .*ByteString_apply_Benchmark.*
bench-jmh/jmh:run -f 1 -wi 3 -i 3 .*ByteString_apply_Benchmark.*
2.12 original
ByteString_apply_Benchmark.bss_apply_best_case thrpt 3 204261596.303 ± 94507102.894 ops/s

View file

@ -28,7 +28,7 @@ class ByteString_grouped_Benchmark {
private val bsLarge = ByteString(Array.ofDim[Byte](1000 * 1000))
/*
> akka-bench-jmh/jmh:run -f1 .*ByteString_grouped_Benchmark
> bench-jmh/jmh:run -f1 .*ByteString_grouped_Benchmark
[info] Benchmark Mode Cnt Score Error Units
[info] ByteString_grouped_Benchmark.grouped avgt 10 59386.328 ± 1466.045 ns/op
*/

View file

@ -30,7 +30,7 @@ class ByteString_toArray_Benchmark {
var kb = 0
/*
akka-bench-jmh/jmh:run -f 1 -wi 5 -i 5 .*ByteString_toArray_Benchmark.*
bench-jmh/jmh:run -f 1 -wi 5 -i 5 .*ByteString_toArray_Benchmark.*
Benchmark (kb) Mode Cnt Score Error Units

View file

@ -78,7 +78,7 @@ import pekko.util.JavaDurationConverters._
* To make sure the messages can be delivered also in that scenario the `ShardingProducerController` can be
* used with a [[DurableProducerQueue]]. Then the unconfirmed messages are stored in a durable way so
* that they can be redelivered when the producer is started again. An implementation of the
* `DurableProducerQueue` is provided by `EventSourcedProducerQueue` in `akka-persistence-typed`.
* `DurableProducerQueue` is provided by `EventSourcedProducerQueue` in `pekko-persistence-typed`.
*
* Instead of using `tell` with the `sendNextTo` in the `RequestNext` the producer can use `context.ask`
* with the `askNextTo` in the `RequestNext`. The difference is that a reply is sent back when the

View file

@ -125,7 +125,7 @@ object ClusterSharding {
* the shard and thereby allocate a new home for the shard and then buffered messages in the
* `ShardRegion` actors are delivered to the new location. This means that the state of the entities
* are not transferred or migrated. If the state of the entities are of importance it should be
* persistent (durable), e.g. with `akka-persistence`, so that it can be recovered at the new
* persistent (durable), e.g. with `pekko-persistence`, so that it can be recovered at the new
* location.
*
* The logic that decides which shards to rebalance is defined in a plugable shard
@ -136,7 +136,7 @@ object ClusterSharding {
* implementation.
*
* The state of shard locations in the `ShardCoordinator` is stored with `akka-distributed-data` or
* `akka-persistence` to survive failures. When a crashed or unreachable coordinator
* `pekko-persistence` to survive failures. When a crashed or unreachable coordinator
* node has been removed (via down) from the cluster a new `ShardCoordinator` singleton
* actor will take over and the state is recovered. During such a failure period shards
* with known location are still available, while messages for new (unknown) shards
@ -146,7 +146,7 @@ object ClusterSharding {
* actor the order of the messages is preserved. As long as the buffer limit is not reached
* messages are delivered on a best effort basis, with at-most once delivery semantics,
* in the same way as ordinary message sending. Reliable end-to-end messaging, with
* at-least-once semantics can be added by using `AtLeastOnceDelivery` in `akka-persistence`.
* at-least-once semantics can be added by using `AtLeastOnceDelivery` in `pekko-persistence`.
*
* Some additional latency is introduced for messages targeted to new or previously
* unused shards due to the round-trip to the coordinator. Rebalancing of shards may

View file

@ -125,7 +125,7 @@ object ClusterSharding extends ExtensionId[ClusterSharding] {
* the shard and thereby allocate a new home for the shard and then buffered messages in the
* `ShardRegion` actors are delivered to the new location. This means that the state of the entities
* are not transferred or migrated. If the state of the entities are of importance it should be
* persistent (durable), e.g. with `akka-persistence`, so that it can be recovered at the new
* persistent (durable), e.g. with `pekko-persistence`, so that it can be recovered at the new
* location.
*
* The logic that decides which shards to rebalance is defined in a plugable shard
@ -137,7 +137,7 @@ object ClusterSharding extends ExtensionId[ClusterSharding] {
* implementation.
*
* The state of shard locations in the `ShardCoordinator` is stored with `akka-distributed-data` or
* `akka-persistence` to survive failures. When a crashed or unreachable coordinator
* `pekko-persistence` to survive failures. When a crashed or unreachable coordinator
* node has been removed (via down) from the cluster a new `ShardCoordinator` singleton
* actor will take over and the state is recovered. During such a failure period shards
* with known location are still available, while messages for new (unknown) shards
@ -147,7 +147,7 @@ object ClusterSharding extends ExtensionId[ClusterSharding] {
* actor the order of the messages is preserved. As long as the buffer limit is not reached
* messages are delivered on a best effort basis, with at-most once delivery semantics,
* in the same way as ordinary message sending. Reliable end-to-end messaging, with
* at-least-once semantics can be added by using `AtLeastOnceDelivery` in `akka-persistence`.
* at-least-once semantics can be added by using `AtLeastOnceDelivery` in `pekko-persistence`.
*
* Some additional latency is introduced for messages targeted to new or previously
* unused shards due to the round-trip to the coordinator. Rebalancing of shards may

View file

@ -22,25 +22,25 @@ import jdk.jfr.{ Category, Enabled, Event, Label, StackTrace, Timespan }
@InternalApi
@StackTrace(false)
@Category(Array("Akka", "Sharding", "Shard")) @Label("Remember Entity Operation")
@Category(Array("Pekko", "Sharding", "Shard")) @Label("Remember Entity Operation")
final class RememberEntityWrite(@Timespan(Timespan.NANOSECONDS) val timeTaken: Long) extends Event
@InternalApi
@StackTrace(false)
@Category(Array("Akka", "Sharding", "Shard")) @Label("Remember Entity Add")
@Category(Array("Pekko", "Sharding", "Shard")) @Label("Remember Entity Add")
final class RememberEntityAdd(val entityId: String) extends Event
@InternalApi
@StackTrace(false)
@Category(Array("Akka", "Sharding", "Shard")) @Label("Remember Entity Remove")
@Category(Array("Pekko", "Sharding", "Shard")) @Label("Remember Entity Remove")
final class RememberEntityRemove(val entityId: String) extends Event
@InternalApi
@StackTrace(false)
@Category(Array("Akka", "Sharding", "Shard")) @Label("Passivate")
@Category(Array("Pekko", "Sharding", "Shard")) @Label("Passivate")
final class Passivate(val entityId: String) extends Event
@InternalApi
@StackTrace(false)
@Category(Array("Akka", "Sharding", "Shard")) @Label("Passivate Restart")
@Category(Array("Pekko", "Sharding", "Shard")) @Label("Passivate Restart")
final class PassivateRestart(val entityId: String) extends Event

View file

@ -118,7 +118,7 @@ import pekko.util.ccompat.JavaConverters._
* the shard, allocate a new home for the shard and then buffered messages in the
* `ShardRegion` actors are delivered to the new location. This means that the state of the entities
* are not transferred or migrated. If the state of the entities are of importance it should be
* persistent (durable), e.g. with `akka-persistence` so that it can be recovered at the new
* persistent (durable), e.g. with `pekko-persistence` so that it can be recovered at the new
* location.
*
* '''Shard Allocation''':
@ -131,7 +131,7 @@ import pekko.util.ccompat.JavaConverters._
*
* '''Recovery''':
* The state of shard locations in the `ShardCoordinator` is stored with `akka-distributed-data` or
* `akka-persistence` to survive failures. When a crashed or unreachable coordinator
* `pekko-persistence` to survive failures. When a crashed or unreachable coordinator
* node has been removed (via down) from the cluster a new `ShardCoordinator` singleton
* actor will take over and the state is recovered. During such a failure period shards
* with known location are still available, while messages for new (unknown) shards
@ -142,7 +142,7 @@ import pekko.util.ccompat.JavaConverters._
* actor the order of the messages is preserved. As long as the buffer limit is not reached
* messages are delivered on a best effort basis, with at-most once delivery semantics,
* in the same way as ordinary message sending. Reliable end-to-end messaging, with
* at-least-once semantics can be added by using `AtLeastOnceDelivery` in `akka-persistence`.
* at-least-once semantics can be added by using `AtLeastOnceDelivery` in `pekko-persistence`.
*
* Some additional latency is introduced for messages targeted to new or previously
* unused shards due to the round-trip to the coordinator. Rebalancing of shards may

View file

@ -127,7 +127,7 @@ object ClusterClientSettings {
* external service registry
*/
@deprecated(
"Use Akka gRPC instead, see https://pekko.apache.org/docs/pekko/current/cluster-client.html#migration-to-akka-grpc",
"Use Apache Pekko gRPC instead, see https://pekko.apache.org/docs/pekko/current/cluster-client.html#migration-to-pekko-grpc",
since = "Akka 2.6.0")
final class ClusterClientSettings(
val initialContacts: Set[ActorPath],
@ -292,7 +292,7 @@ final case class ContactPoints(contactPoints: Set[ActorPath]) {
}
@deprecated(
"Use Akka gRPC instead, see https://pekko.apache.org/docs/pekko/current/cluster-client.html#migration-to-akka-grpc",
"Use Apache Pekko gRPC instead, see https://pekko.apache.org/docs/pekko/current/cluster-client.html#migration-to-pekko-grpc",
since = "Akka 2.6.0")
object ClusterClient {
@ -300,14 +300,14 @@ object ClusterClient {
* Scala API: Factory method for `ClusterClient` [[pekko.actor.Props]].
*/
@deprecated(
"Use Akka gRPC instead, see https://pekko.apache.org/docs/pekko/current/cluster-client.html#migration-to-akka-grpc",
"Use Apache Pekko gRPC instead, see https://pekko.apache.org/docs/pekko/current/cluster-client.html#migration-to-pekko-grpc",
since = "Akka 2.6.0")
def props(settings: ClusterClientSettings): Props =
Props(new ClusterClient(settings)).withDeploy(Deploy.local)
@SerialVersionUID(1L)
@deprecated(
"Use Akka gRPC instead, see https://pekko.apache.org/docs/pekko/current/cluster-client.html#migration-to-akka-grpc",
"Use Apache Pekko gRPC instead, see https://pekko.apache.org/docs/pekko/current/cluster-client.html#migration-to-pekko-grpc",
since = "Akka 2.6.0")
final case class Send(path: String, msg: Any, localAffinity: Boolean) {
@ -318,13 +318,13 @@ object ClusterClient {
}
@SerialVersionUID(1L)
@deprecated(
"Use Akka gRPC instead, see https://pekko.apache.org/docs/pekko/current/cluster-client.html#migration-to-akka-grpc",
"Use Apache Pekko gRPC instead, see https://pekko.apache.org/docs/pekko/current/cluster-client.html#migration-to-pekko-grpc",
since = "Akka 2.6.0")
final case class SendToAll(path: String, msg: Any)
@SerialVersionUID(1L)
@deprecated(
"Use Akka gRPC instead, see https://pekko.apache.org/docs/pekko/current/cluster-client.html#migration-to-akka-grpc",
"Use Apache Pekko gRPC instead, see https://pekko.apache.org/docs/pekko/current/cluster-client.html#migration-to-pekko-grpc",
since = "Akka 2.6.0")
final case class Publish(topic: String, msg: Any)
@ -381,7 +381,7 @@ object ClusterClient {
* nature of the actors involved.
*/
@deprecated(
"Use Akka gRPC instead, see https://pekko.apache.org/docs/pekko/current/cluster-client.html#migration-to-akka-grpc",
"Use Apache Pekko gRPC instead, see https://pekko.apache.org/docs/pekko/current/cluster-client.html#migration-to-pekko-grpc",
since = "Akka 2.6.0")
final class ClusterClient(settings: ClusterClientSettings) extends Actor with ActorLogging {
@ -565,7 +565,7 @@ final class ClusterClient(settings: ClusterClientSettings) extends Actor with Ac
}
@deprecated(
"Use Akka gRPC instead, see https://pekko.apache.org/docs/pekko/current/cluster-client.html#migration-to-akka-grpc",
"Use Apache Pekko gRPC instead, see https://pekko.apache.org/docs/pekko/current/cluster-client.html#migration-to-pekko-grpc",
since = "Akka 2.6.0")
object ClusterClientReceptionist extends ExtensionId[ClusterClientReceptionist] with ExtensionIdProvider {
override def get(system: ActorSystem): ClusterClientReceptionist = super.get(system)
@ -583,7 +583,7 @@ object ClusterClientReceptionist extends ExtensionId[ClusterClientReceptionist]
* The [[pekko.cluster.pubsub.DistributedPubSubMediator]] is started by the [[pekko.cluster.pubsub.DistributedPubSub]] extension.
*/
@deprecated(
"Use Akka gRPC instead, see https://pekko.apache.org/docs/pekko/current/cluster-client.html#migration-to-akka-grpc",
"Use Apache Pekko gRPC instead, see https://pekko.apache.org/docs/pekko/current/cluster-client.html#migration-to-pekko-grpc",
since = "Akka 2.6.0")
final class ClusterClientReceptionist(system: ExtendedActorSystem) extends Extension {
@ -661,7 +661,7 @@ final class ClusterClientReceptionist(system: ExtendedActorSystem) extends Exten
}
@deprecated(
"Use Akka gRPC instead, see https://pekko.apache.org/docs/pekko/current/cluster-client.html#migration-to-akka-grpc",
"Use Apache Pekko gRPC instead, see https://pekko.apache.org/docs/pekko/current/cluster-client.html#migration-to-pekko-grpc",
since = "Akka 2.6.0")
object ClusterReceptionistSettings {
@ -713,7 +713,7 @@ object ClusterReceptionistSettings {
* client will be stopped after this time of inactivity.
*/
@deprecated(
"Use Akka gRPC instead, see https://pekko.apache.org/docs/pekko/current/cluster-client.html#migration-to-akka-grpc",
"Use Apache Pekko gRPC instead, see https://pekko.apache.org/docs/pekko/current/cluster-client.html#migration-to-pekko-grpc",
since = "Akka 2.6.0")
final class ClusterReceptionistSettings(
val role: Option[String],
@ -875,7 +875,7 @@ final case class ClusterClients(clusterClients: Set[ActorRef]) {
}
@deprecated(
"Use Akka gRPC instead, see https://pekko.apache.org/docs/pekko/current/cluster-client.html#migration-to-akka-grpc",
"Use Apache Pekko gRPC instead, see https://pekko.apache.org/docs/pekko/current/cluster-client.html#migration-to-pekko-grpc",
since = "Akka 2.6.0")
object ClusterReceptionist {
@ -952,7 +952,7 @@ object ClusterReceptionist {
* the client is supposed to communicate directly to the actor in the cluster.
*/
@deprecated(
"Use Akka gRPC instead, see https://pekko.apache.org/docs/pekko/current/cluster-client.html#migration-to-akka-grpc",
"Use Apache Pekko gRPC instead, see https://pekko.apache.org/docs/pekko/current/cluster-client.html#migration-to-pekko-grpc",
since = "Akka 2.6.0")
final class ClusterReceptionist(pubSubMediator: ActorRef, settings: ClusterReceptionistSettings)
extends Actor

View file

@ -146,5 +146,5 @@ import pekko.util.Helpers.Requiring
releaseAfter: FiniteDuration,
role: Option[String],
leaseName: Option[String]) {
def safeLeaseName(systemName: String) = leaseName.getOrElse(s"$systemName-akka-sbr")
def safeLeaseName(systemName: String) = leaseName.getOrElse(s"$systemName-pekko-sbr")
}

View file

@ -82,7 +82,7 @@ class LeaseMajority5NodeSpecMultiJvmNode5 extends LeaseMajority5NodeSpec
class LeaseMajority5NodeSpec extends MultiNodeClusterSpec(LeaseMajority5NodeSpec) {
import LeaseMajority5NodeSpec._
private val testLeaseName = "LeaseMajority5NodeSpec-akka-sbr"
private val testLeaseName = "LeaseMajority5NodeSpec-pekko-sbr"
def sortByAddress(roles: RoleName*): List[RoleName] = {

View file

@ -31,7 +31,7 @@ class LeaseMajoritySpec extends PekkoSpec() with Eventually {
}""").withFallback(default)
val named = ConfigFactory.parseString("""
pekko.cluster.split-brain-resolver.lease-majority {
lease-name = "shopping-cart-akka-sbr"
lease-name = "shopping-cart-pekko-sbr"
}""").withFallback(default)
"Split Brain Resolver Lease Majority provider" must {
@ -39,14 +39,15 @@ class LeaseMajoritySpec extends PekkoSpec() with Eventually {
"read the configured name" in {
new SplitBrainResolverSettings(default).leaseMajoritySettings.leaseName shouldBe None
new SplitBrainResolverSettings(blank).leaseMajoritySettings.leaseName shouldBe None
new SplitBrainResolverSettings(named).leaseMajoritySettings.leaseName shouldBe Some("shopping-cart-akka-sbr")
new SplitBrainResolverSettings(named).leaseMajoritySettings.leaseName shouldBe Some("shopping-cart-pekko-sbr")
}
"use a safe name" in {
new SplitBrainResolverSettings(default).leaseMajoritySettings.safeLeaseName("sysName") shouldBe "sysName-akka-sbr"
new SplitBrainResolverSettings(blank).leaseMajoritySettings.safeLeaseName("sysName") shouldBe "sysName-akka-sbr"
new SplitBrainResolverSettings(default).leaseMajoritySettings.safeLeaseName(
"sysName") shouldBe "sysName-pekko-sbr"
new SplitBrainResolverSettings(blank).leaseMajoritySettings.safeLeaseName("sysName") shouldBe "sysName-pekko-sbr"
new SplitBrainResolverSettings(named).leaseMajoritySettings
.safeLeaseName("sysName") shouldBe "shopping-cart-akka-sbr"
.safeLeaseName("sysName") shouldBe "shopping-cart-pekko-sbr"
}
}

View file

@ -126,7 +126,7 @@ class SplitBrainResolverSpec
private lazy val selfUniqueAddress = Cluster(system).selfUniqueAddress
private val testLeaseSettings =
new LeaseSettings("akka-sbr", "test", new TimeoutSettings(1.second, 2.minutes, 3.seconds), ConfigFactory.empty)
new LeaseSettings("pekko-sbr", "test", new TimeoutSettings(1.second, 2.minutes, 3.seconds), ConfigFactory.empty)
def createReachability(unreachability: Seq[(Member, Member)]): Reachability = {
Reachability(

View file

@ -274,7 +274,7 @@ used the next time this projection is started. This pattern is implemented in th
## Query plugins
Query plugins are various (mostly community driven) @apidoc[query.*.ReadJournal] implementations for all kinds
of available datastores. The complete list of available plugins is maintained on the Pekko Persistence Query [Community Plugins](https://akka.io/community/#plugins-to-akka-persistence-query) page.
of available datastores.
This section aims to provide tips and guide plugin developers through implementing a custom query plugin.
Most users will not need to implement journals themselves, except if targeting a not yet supported datastore.

View file

@ -43,9 +43,9 @@ Artifact names:
| pekko-distributed-data | pekko-cluster-typed |
| pekko-persistence | pekko-persistence-typed |
| pekko-stream | pekko-stream-typed |
| pekko-testkit | akka-actor-testkit-typed |
| pekko-testkit | pekko-actor-testkit-typed |
Cluster Singleton and Distributed Data are included in `akka-cluster-typed`.
Cluster Singleton and Distributed Data are included in `pekko-cluster-typed`.
Artifacts not listed in above table don't have a specific API for Pekko Typed.

View file

@ -49,8 +49,8 @@ class OsgiActorSystemFactory(
/**
* Strategy method to create the Config for the ActorSystem
* ensuring that the default/reference configuration is loaded from the akka-actor bundle.
* Configuration files found in akka-actor bundle
* ensuring that the default/reference configuration is loaded from the pekko-actor bundle.
* Configuration files found in pekko-actor bundle
*/
def actorSystemConfig(@unused context: BundleContext): Config = {
config.withFallback(
@ -71,7 +71,7 @@ class OsgiActorSystemFactory(
object OsgiActorSystemFactory {
/**
* Class loader of akka-actor bundle.
* Class loader of pekko-actor bundle.
*/
def akkaActorClassLoader = classOf[ActorSystemActivator].getClassLoader

View file

@ -31,7 +31,7 @@ import org.scalatest.{ BeforeAndAfterAll, Suite }
import org.apache.pekko.util.ccompat.JavaConverters._
/**
* Trait that provides support for building akka-osgi tests using PojoSR
* Trait that provides support for building pekko-osgi tests using PojoSR
*/
trait PojoSRTestSupport extends Suite with BeforeAndAfterAll {

View file

@ -40,7 +40,7 @@ abstract class PluginSpec(val config: Config)
private var _writerUuid: String = _
// used to avoid messages be delivered to a restarted actor,
// this is akka-persistence internals and journals themselves don't really care
// this is pekko-persistence internals and journals themselves don't really care
protected val actorInstanceId = 1
override protected def beforeEach(): Unit = {

View file

@ -24,7 +24,7 @@ import org.scalatest.{ Args, ConfigMap, Filter, Status, Suite, TestData }
/**
* JAVA API
*
* This spec aims to verify custom akka-persistence [[pekko.persistence.snapshot.SnapshotStore]] implementations.
* This spec aims to verify custom pekko-persistence [[pekko.persistence.snapshot.SnapshotStore]] implementations.
* Plugin authors are highly encouraged to include it in their plugin's test suites.
*
* In case your snapshot-store plugin needs some kind of setup or teardown, override the `beforeAll` or `afterAll`

View file

@ -40,7 +40,7 @@ object JournalSpec {
}
/**
* This spec aims to verify custom akka-persistence Journal implementations.
* This spec aims to verify custom pekko-persistence Journal implementations.
* Plugin authors are highly encouraged to include it in their plugin's test suites.
*
* In case your journal plugin needs some kind of setup or teardown, override the `beforeAll` or `afterAll`

View file

@ -40,7 +40,7 @@ object SnapshotStoreSpec {
}
/**
* This spec aims to verify custom akka-persistence [[SnapshotStore]] implementations.
* This spec aims to verify custom pekko-persistence [[SnapshotStore]] implementations.
* Plugin authors are highly encouraged to include it in their plugin's test suites.
*
* In case your journal plugin needs some kind of setup or teardown, override the `beforeAll` or `afterAll`

View file

@ -24,7 +24,7 @@ object JavaFormatter extends AutoPlugin {
private val ignoreConfigFileName: String = ".sbt-java-formatter.conf"
private val descriptor: String = "sbt-java-formatter"
private val formatOnCompile = !sys.props.contains("akka.no.discipline")
private val formatOnCompile = !sys.props.contains("pekko.no.discipline")
import JavaFormatterPlugin.autoImport._
import sbt.Keys._

View file

@ -47,7 +47,7 @@ object OSGi {
val actor = osgiSettings ++ Seq(
OsgiKeys.exportPackage := Seq("org.apache.pekko*"),
OsgiKeys.privatePackage := Seq("org.apache.pekko.osgi.impl"),
// akka-actor packages are not imported, as contained in the CP
// pekko-actor packages are not imported, as contained in the CP
OsgiKeys.importPackage := (osgiOptionalImports.map(optionalResolution)) ++ Seq(
"!sun.misc",
scalaJava8CompatImport(),

View file

@ -23,8 +23,8 @@ object PekkoDisciplinePlugin extends AutoPlugin {
override def requires: Plugins = JvmPlugin
override lazy val projectSettings = disciplineSettings
// allow toggling for pocs/exploration of ideas without discpline
val enabled = !sys.props.contains("akka.no.discipline")
// allow toggling for pocs/exploration of ideas without discipline
val enabled = !sys.props.contains("pekko.no.discipline")
// We allow warnings in docs to get the 'snippets' right
val nonFatalJavaWarningsFor = Set(
@ -140,7 +140,7 @@ object PekkoDisciplinePlugin extends AutoPlugin {
else Seq.empty),
// Discipline is not needed for the docs compilation run (which uses
// different compiler phases from the regular run), and in particular
// '-Ywarn-unused:explicits' breaks 'sbt ++2.13.0-M5 akka-actor/doc'
// '-Ywarn-unused:explicits' breaks 'sbt ++2.13.0-M5 actor/doc'
// https://github.com/akka/akka/issues/26119
Compile / doc / scalacOptions --= disciplineScalacOptions.toSeq :+ "-Xfatal-warnings",
// having discipline warnings in console is just an annoyance

View file

@ -5,7 +5,7 @@ addSbtPlugin("com.lightbend.sbt" % "sbt-bill-of-materials" % "1.0.2")
addSbtPlugin("org.scalameta" % "sbt-scalafmt" % "2.4.6")
addSbtPlugin("ch.epfl.scala" % "sbt-scalafix" % "0.9.33")
// sbt-osgi 0.9.5 is available but breaks including jdk9-only classes
// sbt-osgi 0.9.6 is available but breaks populating akka-protobuf-v3
// sbt-osgi 0.9.6 is available but breaks populating pekko-protobuf-v3
addSbtPlugin("com.typesafe.sbt" % "sbt-osgi" % "0.9.4")
addSbtPlugin("com.typesafe" % "sbt-mima-plugin" % "0.8.1")
addSbtPlugin("com.github.sbt" % "sbt-unidoc" % "0.5.0")

View file

@ -49,7 +49,7 @@ private[pekko] object JFREventUtils {
*/
@InternalApi
@StackTrace(false)
@Category(Array("Akka", "Remoting", "Aeron")) @Label("Media driver started")
@Category(Array("Pekko", "Remoting", "Aeron")) @Label("Media driver started")
final class TransportMediaDriverStarted(val directoryName: String) extends Event
/**
@ -57,7 +57,7 @@ final class TransportMediaDriverStarted(val directoryName: String) extends Event
*/
@InternalApi
@StackTrace(false)
@Category(Array("Akka", "Remoting", "Transport")) @Label("Transport started")
@Category(Array("Pekko", "Remoting", "Transport")) @Label("Transport started")
final class TransportStarted() extends Event
/**
@ -65,7 +65,7 @@ final class TransportStarted() extends Event
*/
@InternalApi
@StackTrace(false)
@Category(Array("Akka", "Remoting", "Aeron")) @Label("Aeron error log started")
@Category(Array("Pekko", "Remoting", "Aeron")) @Label("Aeron error log started")
final class TransportAeronErrorLogStarted() extends Event
/**
@ -73,7 +73,7 @@ final class TransportAeronErrorLogStarted() extends Event
*/
@InternalApi
@StackTrace(false)
@Category(Array("Akka", "Remoting", "Transport")) @Label("Task runner started")
@Category(Array("Pekko", "Remoting", "Transport")) @Label("Task runner started")
final class TransportTaskRunnerStarted() extends Event
/**
@ -81,7 +81,7 @@ final class TransportTaskRunnerStarted() extends Event
*/
@InternalApi
@StackTrace(false)
@Category(Array("Akka", "Remoting", "Transport")) @Label("Unique address set")
@Category(Array("Pekko", "Remoting", "Transport")) @Label("Unique address set")
final class TransportUniqueAddressSet(_uniqueAddress: UniqueAddress) extends Event {
val uniqueAddress = _uniqueAddress.toString()
}
@ -91,7 +91,7 @@ final class TransportUniqueAddressSet(_uniqueAddress: UniqueAddress) extends Eve
*/
@InternalApi
@StackTrace(false)
@Category(Array("Akka", "Remoting", "Transport")) @Label("Materializer started")
@Category(Array("Pekko", "Remoting", "Transport")) @Label("Materializer started")
final class TransportMaterializerStarted() extends Event
/**
@ -99,7 +99,7 @@ final class TransportMaterializerStarted() extends Event
*/
@InternalApi
@StackTrace(false)
@Category(Array("Akka", "Remoting", "Transport")) @Label("Startup finished")
@Category(Array("Pekko", "Remoting", "Transport")) @Label("Startup finished")
final class TransportStartupFinished() extends Event
/**
@ -107,7 +107,7 @@ final class TransportStartupFinished() extends Event
*/
@InternalApi
@StackTrace(false)
@Category(Array("Akka", "Remoting", "Transport")) @Label("Kill switch pulled")
@Category(Array("Pekko", "Remoting", "Transport")) @Label("Kill switch pulled")
final class TransportKillSwitchPulled() extends Event
/**
@ -115,7 +115,7 @@ final class TransportKillSwitchPulled() extends Event
*/
@InternalApi
@StackTrace(false)
@Category(Array("Akka", "Remoting", "Transport")) @Label("Stopped")
@Category(Array("Pekko", "Remoting", "Transport")) @Label("Stopped")
final class TransportStopped() extends Event
/**
@ -123,7 +123,7 @@ final class TransportStopped() extends Event
*/
@InternalApi
@StackTrace(false)
@Category(Array("Akka", "Remoting", "Aeron")) @Label("Aeron log task stopped")
@Category(Array("Pekko", "Remoting", "Aeron")) @Label("Aeron log task stopped")
final class TransportAeronErrorLogTaskStopped() extends Event
/**
@ -131,7 +131,7 @@ final class TransportAeronErrorLogTaskStopped() extends Event
*/
@InternalApi
@StackTrace(false)
@Category(Array("Akka", "Remoting", "Transport")) @Label("Media file deleted")
@Category(Array("Pekko", "Remoting", "Transport")) @Label("Media file deleted")
final class TransportMediaFileDeleted() extends Event
/**
@ -139,7 +139,7 @@ final class TransportMediaFileDeleted() extends Event
*/
@InternalApi
@StackTrace(false)
@Category(Array("Akka", "Remoting", "Transport")) @Label("Send queue overflow")
@Category(Array("Pekko", "Remoting", "Transport")) @Label("Send queue overflow")
final class TransportSendQueueOverflow(val queueIndex: Int) extends Event
/**
@ -147,7 +147,7 @@ final class TransportSendQueueOverflow(val queueIndex: Int) extends Event
*/
@InternalApi
@StackTrace(false)
@Category(Array("Akka", "Remoting", "Transport")) @Label("Stop idle outbound")
@Category(Array("Pekko", "Remoting", "Transport")) @Label("Stop idle outbound")
final class TransportStopIdleOutbound(_remoteAddress: Address, val queueIndex: Int) extends Event {
val remoteAddress = _remoteAddress.toString
}
@ -157,7 +157,7 @@ final class TransportStopIdleOutbound(_remoteAddress: Address, val queueIndex: I
*/
@InternalApi
@StackTrace(false)
@Category(Array("Akka", "Remoting", "Transport")) @Label("Quarantined")
@Category(Array("Pekko", "Remoting", "Transport")) @Label("Quarantined")
final class TransportQuarantined(_remoteAddress: Address, val uid: Long) extends Event {
val remoteAddress = _remoteAddress.toString
}
@ -167,7 +167,7 @@ final class TransportQuarantined(_remoteAddress: Address, val uid: Long) extends
*/
@InternalApi
@StackTrace(false)
@Category(Array("Akka", "Remoting", "Transport")) @Label("Remove quarantined")
@Category(Array("Pekko", "Remoting", "Transport")) @Label("Remove quarantined")
final class TransportRemoveQuarantined(_remoteAddress: Address) extends Event {
val remoteAddress = _remoteAddress.toString
}
@ -177,7 +177,7 @@ final class TransportRemoveQuarantined(_remoteAddress: Address) extends Event {
*/
@InternalApi
@StackTrace(false)
@Category(Array("Akka", "Remoting", "Transport")) @Label("Restart outbound")
@Category(Array("Pekko", "Remoting", "Transport")) @Label("Restart outbound")
final class TransportRestartOutbound(_remoteAddress: Address, val streamName: String) extends Event {
val remoteAddress = _remoteAddress.toString
}
@ -187,7 +187,7 @@ final class TransportRestartOutbound(_remoteAddress: Address, val streamName: St
*/
@InternalApi
@StackTrace(false)
@Category(Array("Akka", "Remoting", "Transport")) @Label("Restart inbound")
@Category(Array("Pekko", "Remoting", "Transport")) @Label("Restart inbound")
final class TransportRestartInbound(_remoteAddress: UniqueAddress, val streamName: String) extends Event {
val remoteAddress = _remoteAddress.toString()
}
@ -199,7 +199,7 @@ final class TransportRestartInbound(_remoteAddress: UniqueAddress, val streamNam
*/
@InternalApi
@StackTrace(false)
@Category(Array("Akka", "Remoting", "Aeron", "Sink")) @Label("Started")
@Category(Array("Pekko", "Remoting", "Aeron", "Sink")) @Label("Started")
final class AeronSinkStarted(val channel: String, val streamId: Int) extends Event {}
/**
@ -207,7 +207,7 @@ final class AeronSinkStarted(val channel: String, val streamId: Int) extends Eve
*/
@InternalApi
@StackTrace(false)
@Category(Array("Akka", "Remoting", "Aeron", "Sink")) @Label("Task runner removed")
@Category(Array("Pekko", "Remoting", "Aeron", "Sink")) @Label("Task runner removed")
final class AeronSinkTaskRunnerRemoved(val channel: String, val streamId: Int) extends Event
/**
@ -215,7 +215,7 @@ final class AeronSinkTaskRunnerRemoved(val channel: String, val streamId: Int) e
*/
@InternalApi
@StackTrace(false)
@Category(Array("Akka", "Remoting", "Aeron", "Sink")) @Label("Publication closed")
@Category(Array("Pekko", "Remoting", "Aeron", "Sink")) @Label("Publication closed")
final class AeronSinkPublicationClosed(val channel: String, val streamId: Int) extends Event
/**
@ -223,7 +223,7 @@ final class AeronSinkPublicationClosed(val channel: String, val streamId: Int) e
*/
@InternalApi
@StackTrace(false)
@Category(Array("Akka", "Remoting", "Aeron", "Sink")) @Label("Publication closed unexpectedly")
@Category(Array("Pekko", "Remoting", "Aeron", "Sink")) @Label("Publication closed unexpectedly")
final class AeronSinkPublicationClosedUnexpectedly(val channel: String, val streamId: Int) extends Event
/**
@ -231,7 +231,7 @@ final class AeronSinkPublicationClosedUnexpectedly(val channel: String, val stre
*/
@InternalApi
@StackTrace(false)
@Category(Array("Akka", "Remoting", "Aeron", "Sink")) @Label("Stopped")
@Category(Array("Pekko", "Remoting", "Aeron", "Sink")) @Label("Stopped")
final class AeronSinkStopped(val channel: String, val streamId: Int) extends Event
/**
@ -240,7 +240,7 @@ final class AeronSinkStopped(val channel: String, val streamId: Int) extends Eve
@InternalApi
@Enabled(false) // hi frequency event
@StackTrace(false)
@Category(Array("Akka", "Remoting", "Aeron", "Sink")) @Label("Envelope grabbed")
@Category(Array("Pekko", "Remoting", "Aeron", "Sink")) @Label("Envelope grabbed")
final class AeronSinkEnvelopeGrabbed(@DataAmount() val lastMessageSize: Int) extends Event
/**
@ -249,7 +249,7 @@ final class AeronSinkEnvelopeGrabbed(@DataAmount() val lastMessageSize: Int) ext
@InternalApi
@Enabled(false) // hi frequency event
@StackTrace(false)
@Category(Array("Akka", "Remoting", "Aeron", "Sink")) @Label("Envelope offered")
@Category(Array("Pekko", "Remoting", "Aeron", "Sink")) @Label("Envelope offered")
final class AeronSinkEnvelopeOffered(@DataAmount() val lastMessageSize: Int) extends Event
/**
@ -257,7 +257,7 @@ final class AeronSinkEnvelopeOffered(@DataAmount() val lastMessageSize: Int) ext
*/
@InternalApi
@StackTrace(false)
@Category(Array("Akka", "Remoting", "Aeron", "Sink")) @Label("Gave up envelope")
@Category(Array("Pekko", "Remoting", "Aeron", "Sink")) @Label("Gave up envelope")
final class AeronSinkGaveUpEnvelope(val cause: String) extends Event
/**
@ -266,7 +266,7 @@ final class AeronSinkGaveUpEnvelope(val cause: String) extends Event
@InternalApi
@Enabled(false) // hi frequency event
@StackTrace(false)
@Category(Array("Akka", "Remoting", "Aeron", "Sink")) @Label("Delegate to task runner")
@Category(Array("Pekko", "Remoting", "Aeron", "Sink")) @Label("Delegate to task runner")
final class AeronSinkDelegateToTaskRunner(val countBeforeDelegate: Long) extends Event
/**
@ -275,7 +275,7 @@ final class AeronSinkDelegateToTaskRunner(val countBeforeDelegate: Long) extends
@InternalApi
@Enabled(false) // hi frequency event
@StackTrace(false)
@Category(Array("Akka", "Remoting", "Aeron", "Sink")) @Label("Return from task runner")
@Category(Array("Pekko", "Remoting", "Aeron", "Sink")) @Label("Return from task runner")
final class AeronSinkReturnFromTaskRunner(@Timespan(Timespan.NANOSECONDS) val nanosSinceTaskStartTime: Long)
extends Event
@ -286,7 +286,7 @@ final class AeronSinkReturnFromTaskRunner(@Timespan(Timespan.NANOSECONDS) val na
*/
@InternalApi
@StackTrace(false)
@Category(Array("Akka", "Remoting", "Aeron", "Source")) @Label("Started")
@Category(Array("Pekko", "Remoting", "Aeron", "Source")) @Label("Started")
final class AeronSourceStarted(val channel: String, val streamId: Int) extends Event
/**
@ -294,7 +294,7 @@ final class AeronSourceStarted(val channel: String, val streamId: Int) extends E
*/
@InternalApi
@StackTrace(false)
@Category(Array("Akka", "Remoting", "Aeron", "Source")) @Label("Stopped")
@Category(Array("Pekko", "Remoting", "Aeron", "Source")) @Label("Stopped")
final class AeronSourceStopped(val channel: String, val streamId: Int) extends Event
/**
@ -303,7 +303,7 @@ final class AeronSourceStopped(val channel: String, val streamId: Int) extends E
@InternalApi
@Enabled(false) // hi frequency event
@StackTrace(false)
@Category(Array("Akka", "Remoting", "Aeron", "Source")) @Label("Received")
@Category(Array("Pekko", "Remoting", "Aeron", "Source")) @Label("Received")
final class AeronSourceReceived(@DataAmount() val size: Int) extends Event
/**
@ -312,7 +312,7 @@ final class AeronSourceReceived(@DataAmount() val size: Int) extends Event
@InternalApi
@Enabled(false) // hi frequency event
@StackTrace(false)
@Category(Array("Akka", "Remoting", "Aeron", "Source")) @Label("Delegate to task runner")
@Category(Array("Pekko", "Remoting", "Aeron", "Source")) @Label("Delegate to task runner")
final class AeronSourceDelegateToTaskRunner(val countBeforeDelegate: Long) extends Event
/**
@ -321,7 +321,7 @@ final class AeronSourceDelegateToTaskRunner(val countBeforeDelegate: Long) exten
@InternalApi
@Enabled(false) // hi frequency event
@StackTrace(false)
@Category(Array("Akka", "Remoting", "Aeron", "Source")) @Label("Return from task runner")
@Category(Array("Pekko", "Remoting", "Aeron", "Source")) @Label("Return from task runner")
final class AeronSourceReturnFromTaskRunner(@Timespan(Timespan.NANOSECONDS) val nanosSinceTaskStartTime: Long)
extends Event
@ -332,7 +332,7 @@ final class AeronSourceReturnFromTaskRunner(@Timespan(Timespan.NANOSECONDS) val
*/
@InternalApi
@StackTrace(false)
@Category(Array("Akka", "Remoting", "Compression")) @Label("ActorRef advertisement")
@Category(Array("Pekko", "Remoting", "Compression")) @Label("ActorRef advertisement")
final class CompressionActorRefAdvertisement(val uid: Long) extends Event
/**
@ -340,7 +340,7 @@ final class CompressionActorRefAdvertisement(val uid: Long) extends Event
*/
@InternalApi
@StackTrace(false)
@Category(Array("Akka", "Remoting", "Compression")) @Label("ClassManifest advertisement")
@Category(Array("Pekko", "Remoting", "Compression")) @Label("ClassManifest advertisement")
final class CompressionClassManifestAdvertisement(val uid: Long) extends Event
// tcp outbound events
@ -350,7 +350,7 @@ final class CompressionClassManifestAdvertisement(val uid: Long) extends Event
*/
@InternalApi
@StackTrace(false)
@Category(Array("Akka", "Remoting", "Tcp", "Outbound")) @Label("Connected")
@Category(Array("Pekko", "Remoting", "Tcp", "Outbound")) @Label("Connected")
final class TcpOutboundConnected(_remoteAddress: Address, val streamName: String) extends Event {
val remoteAddress = _remoteAddress.toString
}
@ -361,7 +361,7 @@ final class TcpOutboundConnected(_remoteAddress: Address, val streamName: String
@InternalApi
@Enabled(false) // hi frequency event
@StackTrace(false)
@Category(Array("Akka", "Remoting", "Tcp", "Outbound")) @Label("Sent")
@Category(Array("Pekko", "Remoting", "Tcp", "Outbound")) @Label("Sent")
final class TcpOutboundSent(@DataAmount() val size: Int) extends Event
// tcp inbound events
@ -371,7 +371,7 @@ final class TcpOutboundSent(@DataAmount() val size: Int) extends Event
*/
@InternalApi
@StackTrace(false)
@Category(Array("Akka", "Remoting", "Tcp", "Inbound")) @Label("Bound")
@Category(Array("Pekko", "Remoting", "Tcp", "Inbound")) @Label("Bound")
final class TcpInboundBound(val bindHost: String, _address: InetSocketAddress) extends Event {
val address = JFREventUtils.stringOf(_address)
}
@ -381,7 +381,7 @@ final class TcpInboundBound(val bindHost: String, _address: InetSocketAddress) e
*/
@InternalApi
@StackTrace(false)
@Category(Array("Akka", "Remoting", "Tcp", "Inbound")) @Label("Unbound")
@Category(Array("Pekko", "Remoting", "Tcp", "Inbound")) @Label("Unbound")
final class TcpInboundUnbound(_localAddress: UniqueAddress) extends Event {
val localAddress = _localAddress.toString()
}
@ -391,7 +391,7 @@ final class TcpInboundUnbound(_localAddress: UniqueAddress) extends Event {
*/
@InternalApi
@StackTrace(false)
@Category(Array("Akka", "Remoting", "Tcp", "Inbound")) @Label("Connected")
@Category(Array("Pekko", "Remoting", "Tcp", "Inbound")) @Label("Connected")
final class TcpInboundConnected(_remoteAddress: InetSocketAddress) extends Event {
val remoteAddress = JFREventUtils.stringOf(_remoteAddress)
}
@ -402,5 +402,5 @@ final class TcpInboundConnected(_remoteAddress: InetSocketAddress) extends Event
@InternalApi
@Enabled(false) // hi frequency event
@StackTrace(false)
@Category(Array("Akka", "Remoting", "Tcp", "Inbound")) @Label("Received")
@Category(Array("Pekko", "Remoting", "Tcp", "Inbound")) @Label("Received")
final class TcpInboundReceived(@DataAmount() val size: Int) extends Event

View file

@ -34,7 +34,7 @@ private[pekko] object MessageSerializer {
class SerializationException(msg: String, cause: Throwable) extends RuntimeException(msg, cause)
/**
* Uses Akka Serialization for the specified ActorSystem to transform the given MessageProtocol to a message
* Uses Pekko Serialization for the specified ActorSystem to transform the given MessageProtocol to a message
*/
def deserialize(system: ExtendedActorSystem, messageProtocol: SerializedMessage): AnyRef = {
SerializationExtension(system)
@ -46,7 +46,7 @@ private[pekko] object MessageSerializer {
}
/**
* Uses Akka Serialization for the specified ActorSystem to transform the given message to a MessageProtocol
* Uses Pekko Serialization for the specified ActorSystem to transform the given message to a MessageProtocol
* Throws `NotSerializableException` if serializer was not configured for the message type.
* Throws `MessageSerializer.SerializationException` if exception was thrown from `toBinary` of the
* serializer.

View file

@ -26,7 +26,7 @@ class PemManagersProviderSpec extends AnyWordSpec with Matchers {
"A PemManagersProvider" must {
"load stores reading files setup in config (akka-pki samples)" in {
"load stores reading files setup in config (pekko-pki samples)" in {
// These set of certificates are valid PEMs but are invalid for pekko-remote
// use. Either the key length, certificate usage limitations (via the UsageKeyExtensions),
// or the fact that the key's certificate is self-signed cause one of the following

View file

@ -97,7 +97,7 @@ class PrimitivesSerializationSpec extends PekkoSpec(PrimitivesSerializationSpec.
}
"have right serializer id" in {
// checking because moved to akka-actor
// checking because moved to pekko-actor
serializerFor(1L.asInstanceOf[AnyRef]).identifier === 18
}
@ -120,7 +120,7 @@ class PrimitivesSerializationSpec extends PekkoSpec(PrimitivesSerializationSpec.
}
"have right serializer id" in {
// checking because moved to akka-actor
// checking because moved to pekko-actor
serializerFor(1L.asInstanceOf[AnyRef]).identifier === 19
}
}
@ -144,7 +144,7 @@ class PrimitivesSerializationSpec extends PekkoSpec(PrimitivesSerializationSpec.
}
"have right serializer id" in {
// checking because moved to akka-actor
// checking because moved to pekko-actor
serializerFor(1L.asInstanceOf[AnyRef]).identifier === 20
}
@ -173,7 +173,7 @@ class PrimitivesSerializationSpec extends PekkoSpec(PrimitivesSerializationSpec.
}
"have right serializer id" in {
// checking because moved to akka-actor
// checking because moved to pekko-actor
serializerFor(1L.asInstanceOf[AnyRef]).identifier === 21
}

View file

@ -37,7 +37,7 @@ import pekko.annotation.InternalApi
@InternalApi private[pekko] object JacksonModule {
lazy val version: Version = {
val groupId = "com.typesafe.akka"
val groupId = "org.apache.pekko"
val artifactId = "serialization-jackson"
val version = pekko.Version.current
VersionUtil.parseVersion(version, groupId, artifactId)

View file

@ -281,10 +281,10 @@ object JacksonObjectMapperProvider extends ExtensionId[JacksonObjectMapperProvid
private def isModuleEnabled(fqcn: String, dynamicAccess: DynamicAccess): Boolean =
fqcn match {
case "org.apache.pekko.serialization.jackson.PekkoTypedJacksonModule" =>
// akka-actor-typed dependency is "provided" and may not be included
// pekko-actor-typed dependency is "provided" and may not be included
dynamicAccess.classIsOnClasspath("org.apache.pekko.actor.typed.ActorRef")
case "org.apache.pekko.serialization.jackson.PekkoStreamJacksonModule" =>
// akka-stream dependency is "provided" and may not be included
// pekko-stream dependency is "provided" and may not be included
dynamicAccess.classIsOnClasspath("org.apache.pekko.stream.Graph")
case _ => true
}

View file

@ -142,7 +142,7 @@ import pekko.util.OptionVal
/**
* INTERNAL API: only public by configuration
*
* Akka serializer for Jackson with JSON.
* Pekko serializer for Jackson with JSON.
*/
@InternalApi private[pekko] final class JacksonJsonSerializer(system: ExtendedActorSystem, bindingName: String)
extends JacksonSerializer(
@ -153,7 +153,7 @@ import pekko.util.OptionVal
/**
* INTERNAL API: only public by configuration
*
* Akka serializer for Jackson with CBOR.
* Pekko serializer for Jackson with CBOR.
*/
@InternalApi private[pekko] final class JacksonCborSerializer(system: ExtendedActorSystem, bindingName: String)
extends JacksonSerializer(

View file

@ -129,7 +129,7 @@ object ScalaTestMessages {
extends TestMessage
// #jackson-scala-enumeration
// delegate to AkkaSerialization
// delegate to PekkoSerialization
object HasPekkoSerializer {
def apply(description: String): HasPekkoSerializer = new HasPekkoSerializer(description)
}
@ -163,7 +163,7 @@ object ScalaTestMessages {
final case class WithPekkoSerializer(
@JsonDeserialize(`using` = classOf[PekkoSerializationDeserializer])
@JsonSerialize(`using` = classOf[PekkoSerializationSerializer])
akkaSerializer: HasPekkoSerializer)
pekkoSerializer: HasPekkoSerializer)
extends TestMessage
}
@ -1278,7 +1278,7 @@ abstract class JacksonSerializerSpec(serializerName: String)
}
}
"delegate to akka serialization" in {
"delegate to pekko serialization" in {
checkSerialization(WithPekkoSerializer(HasPekkoSerializer("cat")))
}

View file

@ -26,10 +26,10 @@ class IterablePublisherViaJavaFlowPublisherTest extends PekkoPublisherVerificati
val sourceViaJavaFlowPublisher: JavaFlow.Publisher[Int] = Source(iterable(elements))
.runWith(JavaFlowSupport.Sink.asPublisher(fanout = false))
val javaFlowPublisherIntoAkkaSource: Source[Int, NotUsed] =
val javaFlowPublisherIntoPekkoSource: Source[Int, NotUsed] =
JavaFlowSupport.Source.fromPublisher(sourceViaJavaFlowPublisher)
javaFlowPublisherIntoAkkaSource
javaFlowPublisherIntoPekkoSource
.runWith(Sink.asPublisher(false)) // back as RS Publisher
}

View file

@ -50,7 +50,7 @@ abstract class PekkoIdentityProcessorVerification[T](env: TestEnvironment, publi
}
}
/** By default Akka Publishers do not support Fanout! */
/** By default, Pekko Publishers do not support Fanout! */
override def maxSupportedSubscribers: Long = 1L
override lazy val publisherExecutorService: ExecutorService =