Breaks binary compatibility because adding new methods to Eventsourced trait. Since akka-persistence is experimental this is ok, yet source-level compatibility has been perserved thankfuly :-) Deprecates: * Rename of EventsourcedProcessor -> PersistentActor * Processor -> suggest using PersistentActor * Migration guide for akka-persistence is separate, as wel'll deprecate in minor versions (its experimental) * Persistent as well as ConfirmablePersistent - since Processor, their main user will be removed soon. Other changes: * persistAsync works as expected when mixed with persist * A counter must be kept for pending stashing invocations * Uses only 1 shared list buffer for persit / persistAsync * Includes small benchmark * Docs also include info about not using Persistent() wrapper * uses java LinkedList, for best performance of append / head on persistInvocations; the get(0) is safe, because these msgs only come in response to persistInvocations * Renamed internal *MessagesSuccess/Failure messages because we kept small mistakes seeing the class "with s" and "without s" as the same * Updated everything that refered to EventsourcedProcessor to PersistentActor, including samples Refs #15227 Conflicts: akka-docs/rst/project/migration-guides.rst akka-persistence/src/main/scala/akka/persistence/JournalProtocol.scala akka-persistence/src/main/scala/akka/persistence/Persistent.scala akka-persistence/src/test/scala/akka/persistence/PersistentActorSpec.scala project/AkkaBuild.scala
101 lines
4 KiB
Scala
101 lines
4 KiB
Scala
/**
|
|
* Copyright (C) 2009-2014 Typesafe Inc. <http://www.typesafe.com>
|
|
* Copyright (C) 2012-2013 Eligotech BV.
|
|
*/
|
|
|
|
package akka.persistence.journal
|
|
|
|
import scala.collection.immutable
|
|
import scala.util._
|
|
|
|
import akka.actor.Actor
|
|
import akka.pattern.pipe
|
|
import akka.persistence._
|
|
|
|
/**
|
|
* Abstract journal, optimized for synchronous writes.
|
|
*/
|
|
trait SyncWriteJournal extends Actor with AsyncRecovery {
|
|
import JournalProtocol._
|
|
import context.dispatcher
|
|
|
|
private val extension = Persistence(context.system)
|
|
private val publish = extension.settings.internal.publishPluginCommands
|
|
|
|
final def receive = {
|
|
case WriteMessages(persistentBatch, processor) ⇒
|
|
Try(writeMessages(persistentBatch.map(_.prepareWrite()))) match {
|
|
case Success(_) ⇒
|
|
processor ! WriteMessagesSuccessful
|
|
persistentBatch.foreach(p ⇒ processor.tell(WriteMessageSuccess(p), p.sender))
|
|
case Failure(e) ⇒
|
|
processor ! WriteMessagesFailed(e)
|
|
persistentBatch.foreach(p ⇒ processor tell (WriteMessageFailure(p, e), p.sender))
|
|
throw e
|
|
}
|
|
case r @ ReplayMessages(fromSequenceNr, toSequenceNr, max, processorId, processor, replayDeleted) ⇒
|
|
asyncReplayMessages(processorId, fromSequenceNr, toSequenceNr, max) { p ⇒
|
|
if (!p.deleted || replayDeleted) processor.tell(ReplayedMessage(p), p.sender)
|
|
} map {
|
|
case _ ⇒ ReplayMessagesSuccess
|
|
} recover {
|
|
case e ⇒ ReplayMessagesFailure(e)
|
|
} pipeTo (processor) onSuccess {
|
|
case _ if publish ⇒ context.system.eventStream.publish(r)
|
|
}
|
|
case ReadHighestSequenceNr(fromSequenceNr, processorId, processor) ⇒
|
|
asyncReadHighestSequenceNr(processorId, fromSequenceNr).map {
|
|
highest ⇒ ReadHighestSequenceNrSuccess(highest)
|
|
} recover {
|
|
case e ⇒ ReadHighestSequenceNrFailure(e)
|
|
} pipeTo (processor)
|
|
case WriteConfirmations(confirmationsBatch, requestor) ⇒
|
|
Try(writeConfirmations(confirmationsBatch)) match {
|
|
case Success(_) ⇒ requestor ! WriteConfirmationsSuccess(confirmationsBatch)
|
|
case Failure(e) ⇒ requestor ! WriteConfirmationsFailure(e)
|
|
}
|
|
case d @ DeleteMessages(messageIds, permanent, requestorOption) ⇒
|
|
Try(deleteMessages(messageIds, permanent)) match {
|
|
case Success(_) ⇒
|
|
requestorOption.foreach(_ ! DeleteMessagesSuccess(messageIds))
|
|
if (publish) context.system.eventStream.publish(d)
|
|
case Failure(e) ⇒
|
|
requestorOption.foreach(_ ! DeleteMessagesFailure(e))
|
|
}
|
|
case d @ DeleteMessagesTo(processorId, toSequenceNr, permanent) ⇒
|
|
Try(deleteMessagesTo(processorId, toSequenceNr, permanent)) match {
|
|
case Success(_) ⇒ if (publish) context.system.eventStream.publish(d)
|
|
case Failure(e) ⇒
|
|
}
|
|
case LoopMessage(message, processor) ⇒
|
|
processor forward LoopMessageSuccess(message)
|
|
}
|
|
|
|
//#journal-plugin-api
|
|
/**
|
|
* Plugin API: synchronously writes a batch of persistent messages to the journal.
|
|
* The batch write must be atomic i.e. either all persistent messages in the batch
|
|
* are written or none.
|
|
*/
|
|
def writeMessages(messages: immutable.Seq[PersistentRepr]): Unit
|
|
|
|
/**
|
|
* Plugin API: synchronously writes a batch of delivery confirmations to the journal.
|
|
*/
|
|
def writeConfirmations(confirmations: immutable.Seq[PersistentConfirmation]): Unit
|
|
|
|
/**
|
|
* Plugin API: synchronously deletes messages identified by `messageIds` from the
|
|
* journal. If `permanent` is set to `false`, the persistent messages are marked as
|
|
* deleted, otherwise they are permanently deleted.
|
|
*/
|
|
def deleteMessages(messageIds: immutable.Seq[PersistentId], permanent: Boolean): Unit
|
|
|
|
/**
|
|
* Plugin API: synchronously deletes all persistent messages up to `toSequenceNr`
|
|
* (inclusive). If `permanent` is set to `false`, the persistent messages are marked
|
|
* as deleted, otherwise they are permanently deleted.
|
|
*/
|
|
def deleteMessagesTo(processorId: String, toSequenceNr: Long, permanent: Boolean): Unit
|
|
//#journal-plugin-api
|
|
}
|