diff --git a/akka-core/src/main/scala/actor/ActiveObject.scala b/akka-core/src/main/scala/actor/ActiveObject.scala index 9ec943cfc6..0780e355dd 100644 --- a/akka-core/src/main/scala/actor/ActiveObject.scala +++ b/akka-core/src/main/scala/actor/ActiveObject.scala @@ -66,23 +66,23 @@ final class ActiveObjectConfiguration { } /** - * Holds RTTI (runtime type information) for the Active Object, f.e. current 'sender' + * Holds RTTI (runtime type information) for the Active Object, f.e. current 'sender' * reference, the 'senderFuture' reference etc. *

- * In order to make use of this context you have to create a member field in your - * Active Object that has the type 'ActiveObjectContext', then an instance will - * be injected for you to use. + * In order to make use of this context you have to create a member field in your + * Active Object that has the type 'ActiveObjectContext', then an instance will + * be injected for you to use. *

- * This class does not contain static information but is updated by the runtime system - * at runtime. + * This class does not contain static information but is updated by the runtime system + * at runtime. *

- * Here is an example of usage: + * Here is an example of usage: *

  * class Ping {
- *   // This context will be injected, holds RTTI (runtime type information) 
- *   // for the current message send 
+ *   // This context will be injected, holds RTTI (runtime type information)
+ *   // for the current message send
  *   private ActiveObjectContext context = null;
- *   
+ *
  *   public void hit(int count) {
  *     Pong pong = (Pong) context.getSender();
  *     pong.hit(count++)
@@ -100,19 +100,19 @@ final class ActiveObjectContext {
    * Returns the current sender Active Object reference.
    * Scala style getter.
    */
-  def sender: AnyRef = { 
+  def sender: AnyRef = {
     if (_sender eq null) throw new IllegalStateException("Sender reference should not be null.")
     else _sender
-  } 
+  }
 
   /**
    * Returns the current sender Active Object reference.
    * Java style getter.
    */
-   def getSender: AnyRef = { 
+   def getSender: AnyRef = {
      if (_sender eq null) throw new IllegalStateException("Sender reference should not be null.")
      else _sender
-   } 
+   }
 
   /**
    * Returns the current sender future Active Object reference.
@@ -364,7 +364,7 @@ object ActiveObject extends Logging {
     proxy.asInstanceOf[T]
   }
 
-  private[akka] def newInstance[T](intf: Class[T], target: AnyRef, actorRef: ActorRef, 
+  private[akka] def newInstance[T](intf: Class[T], target: AnyRef, actorRef: ActorRef,
                                    remoteAddress: Option[InetSocketAddress], timeout: Long): T = {
     val context = injectActiveObjectContext(target)
     val proxy = Proxy.newInstance(Array(intf), Array(target), false, false)
@@ -462,7 +462,7 @@ object ActiveObject extends Logging {
         if (parent != null) injectActiveObjectContext0(activeObject, parent)
         else {
           log.warning(
-          "Can't set 'ActiveObjectContext' for ActiveObject [%s] since no field of this type could be found.", 
+          "Can't set 'ActiveObjectContext' for ActiveObject [%s] since no field of this type could be found.",
           activeObject.getClass.getName)
           None
         }
@@ -522,7 +522,7 @@ private[akka] sealed class ActiveObjectAspect {
       remoteAddress = init.remoteAddress
       timeout = init.timeout
       isInitialized = true
-      
+
     }
     dispatch(joinPoint)
   }
@@ -583,7 +583,7 @@ private[akka] sealed class ActiveObjectAspect {
     } else future.result
 
   private def isVoid(rtti: MethodRtti) = rtti.getMethod.getReturnType == java.lang.Void.TYPE
-  
+
   private def escapeArguments(args: Array[AnyRef]): Tuple2[Array[AnyRef], Boolean] = {
     var isEscaped = false
     val escapedArgs = for (arg <- args) yield {
@@ -606,11 +606,11 @@ private[akka] sealed class ActiveObjectAspect {
   joinPoint: JoinPoint, isOneWay: Boolean, isVoid: Boolean, sender: AnyRef, senderFuture: CompletableFuture[Any]) {
 
   override def toString: String = synchronized {
-    "Invocation [joinPoint: " + joinPoint.toString + 
-    ", isOneWay: " + isOneWay + 
+    "Invocation [joinPoint: " + joinPoint.toString +
+    ", isOneWay: " + isOneWay +
     ", isVoid: " + isVoid +
-    ", sender: " + sender + 
-    ", senderFuture: " + senderFuture + 
+    ", sender: " + sender +
+    ", senderFuture: " + senderFuture +
     "]"
   }
 
@@ -653,11 +653,11 @@ private[akka] class Dispatcher(transactionalRequired: Boolean, val callbacks: Op
   private var postRestart: Option[Method] = None
   private var initTxState: Option[Method] = None
   private var context: Option[ActiveObjectContext] = None
-  
+
   def this(transactionalRequired: Boolean) = this(transactionalRequired,None)
 
   private[actor] def initialize(targetClass: Class[_], targetInstance: AnyRef, ctx: Option[ActiveObjectContext]) = {
-    if (transactionalRequired || targetClass.isAnnotationPresent(Annotations.transactionrequired)) 
+    if (transactionalRequired || targetClass.isAnnotationPresent(Annotations.transactionrequired))
       self.makeTransactionRequired
     self.id = targetClass.getName
     target = Some(targetInstance)
@@ -705,7 +705,7 @@ private[akka] class Dispatcher(transactionalRequired: Boolean, val callbacks: Op
 
   def receive = {
     case Invocation(joinPoint, isOneWay, _, sender, senderFuture) =>
-      context.foreach { ctx => 
+      context.foreach { ctx =>
         if (sender ne null) ctx._sender = sender
         if (senderFuture ne null) ctx._senderFuture = senderFuture
       }
diff --git a/akka-core/src/main/scala/actor/Actor.scala b/akka-core/src/main/scala/actor/Actor.scala
index 120bea199d..3b7d62f97f 100644
--- a/akka-core/src/main/scala/actor/Actor.scala
+++ b/akka-core/src/main/scala/actor/Actor.scala
@@ -329,7 +329,7 @@ trait Actor extends Logging {
    * 
    * self ! message
    * 
- * Here you also find most of the Actor API. + * Here you also find most of the Actor API. *

* For example fields like: *

@@ -384,7 +384,7 @@ trait Actor extends Logging {
    * Is called when an Actor is started by invoking 'actor.start'.
    */
   def init {}
-  
+
   /**
    * User overridable callback.
    * 

diff --git a/akka-core/src/main/scala/actor/ActorRef.scala b/akka-core/src/main/scala/actor/ActorRef.scala index 200997d1d2..682e85655c 100644 --- a/akka-core/src/main/scala/actor/ActorRef.scala +++ b/akka-core/src/main/scala/actor/ActorRef.scala @@ -222,7 +222,7 @@ trait ActorRef extends TransactionManagement { * Is defined if the message was sent with sent with '!!' or '!!!', else None. */ def senderFuture: Option[CompletableFuture[Any]] = guard.withGuard { _senderFuture } - + /** * Is the actor being restarted? */ @@ -356,7 +356,7 @@ trait ActorRef extends TransactionManagement { "\n\tYou have probably: " + "\n\t\t1. Sent a message to an Actor from an instance that is NOT an Actor." + "\n\t\t2. Invoked a method on an Active Object from an instance NOT an Active Object.") - + /** * Use reply_?(..) to reply with a message to the original sender of the message currently * being processed. @@ -1224,7 +1224,7 @@ private[akka] case class RemoteActorRef private[akka] ( extends ActorRef { _uuid = uuuid timeout = _timeout - + start lazy val remoteClient = RemoteClient.clientFor(hostname, port, loader) diff --git a/akka-core/src/main/scala/remote/RemoteClient.scala b/akka-core/src/main/scala/remote/RemoteClient.scala index f93c5dc345..15ed0409ed 100644 --- a/akka-core/src/main/scala/remote/RemoteClient.scala +++ b/akka-core/src/main/scala/remote/RemoteClient.scala @@ -82,19 +82,19 @@ object RemoteClient extends Logging { private[akka] def actorFor(uuid: String, className: String, timeout: Long, hostname: String, port: Int, loader: Option[ClassLoader]): ActorRef = RemoteActorRef(uuid, className, hostname, port, timeout, loader) - def clientFor(hostname: String, port: Int): RemoteClient = + def clientFor(hostname: String, port: Int): RemoteClient = clientFor(new InetSocketAddress(hostname, port), None) - def clientFor(hostname: String, port: Int, loader: ClassLoader): RemoteClient = + def clientFor(hostname: String, port: Int, loader: ClassLoader): RemoteClient = clientFor(new InetSocketAddress(hostname, port), Some(loader)) - def clientFor(address: InetSocketAddress): RemoteClient = + def clientFor(address: InetSocketAddress): RemoteClient = clientFor(address, None) - def clientFor(address: InetSocketAddress, loader: ClassLoader): RemoteClient = + def clientFor(address: InetSocketAddress, loader: ClassLoader): RemoteClient = clientFor(address, Some(loader)) - private[akka] def clientFor(hostname: String, port: Int, loader: Option[ClassLoader]): RemoteClient = + private[akka] def clientFor(hostname: String, port: Int, loader: Option[ClassLoader]): RemoteClient = clientFor(new InetSocketAddress(hostname, port), loader) private[akka] def clientFor(address: InetSocketAddress, loader: Option[ClassLoader]): RemoteClient = synchronized { @@ -330,7 +330,7 @@ class RemoteClientHandler(val name: String, client.connection = bootstrap.connect(remoteAddress) client.connection.awaitUninterruptibly // Wait until the connection attempt succeeds or fails. if (!client.connection.isSuccess) { - client.listeners.toArray.foreach(l => + client.listeners.toArray.foreach(l => l.asInstanceOf[ActorRef] ! RemoteClientError(client.connection.getCause)) log.error(client.connection.getCause, "Reconnection to [%s] has failed", remoteAddress) } @@ -339,13 +339,13 @@ class RemoteClientHandler(val name: String, } override def channelConnected(ctx: ChannelHandlerContext, event: ChannelStateEvent) = { - client.listeners.toArray.foreach(l => + client.listeners.toArray.foreach(l => l.asInstanceOf[ActorRef] ! RemoteClientConnected(client.hostname, client.port)) log.debug("Remote client connected to [%s]", ctx.getChannel.getRemoteAddress) } override def channelDisconnected(ctx: ChannelHandlerContext, event: ChannelStateEvent) = { - client.listeners.toArray.foreach(l => + client.listeners.toArray.foreach(l => l.asInstanceOf[ActorRef] ! RemoteClientDisconnected(client.hostname, client.port)) log.debug("Remote client disconnected from [%s]", ctx.getChannel.getRemoteAddress) } diff --git a/akka-core/src/main/scala/remote/RemoteServer.scala b/akka-core/src/main/scala/remote/RemoteServer.scala index 3b02a3a850..8f6cd40684 100644 --- a/akka-core/src/main/scala/remote/RemoteServer.scala +++ b/akka-core/src/main/scala/remote/RemoteServer.scala @@ -184,7 +184,7 @@ class RemoteServer extends Logging { def start(_hostname: String, _port: Int): RemoteServer = start(_hostname, _port, None) - private def start(_hostname: String, _port: Int, loader: ClassLoader): RemoteServer = + private def start(_hostname: String, _port: Int, loader: ClassLoader): RemoteServer = start(_hostname, _port, Some(loader)) private def start(_hostname: String, _port: Int, loader: Option[ClassLoader]): RemoteServer = synchronized { diff --git a/akka-core/src/main/scala/routing/Listeners.scala b/akka-core/src/main/scala/routing/Listeners.scala index 4bfc3f30b6..7563e0800b 100644 --- a/akka-core/src/main/scala/routing/Listeners.scala +++ b/akka-core/src/main/scala/routing/Listeners.scala @@ -13,7 +13,7 @@ case class Listen(listener: ActorRef) extends ListenerMessage case class Deafen(listener: ActorRef) extends ListenerMessage case class WithListeners(f: List[ActorRef] => Unit) extends ListenerMessage -/** +/** * Listeners is a generic trait to implement listening capability on an Actor. *

* Use the gossip(msg) method to have it sent to the listeners. @@ -34,6 +34,6 @@ trait Listeners { self: Actor => } protected def gossip(msg: Any) = listenersAsList foreach (_ ! msg) - + private def listenersAsList: List[ActorRef] = listeners.toArray.toList.asInstanceOf[List[ActorRef]] } diff --git a/akka-core/src/main/scala/stm/Transaction.scala b/akka-core/src/main/scala/stm/Transaction.scala index a7870cec93..439093dec6 100644 --- a/akka-core/src/main/scala/stm/Transaction.scala +++ b/akka-core/src/main/scala/stm/Transaction.scala @@ -282,10 +282,10 @@ object Transaction { setTransaction(Some(tx)) mtx.registerLifecycleListener(new TransactionLifecycleListener() { def notify(mtx: MultiverseTransaction, event: TransactionLifecycleEvent) = event.name match { - case "postCommit" => + case "postCommit" => log.trace("Committing transaction [%s]", mtx) tx.commit - case "postAbort" => + case "postAbort" => log.trace("Aborting transaction [%s]", mtx) tx.abort case _ => {} diff --git a/akka-core/src/test/scala/ActorPatternsTest.scala b/akka-core/src/test/scala/ActorPatternsTest.scala index 028ea80ea3..0d4e9b6b08 100644 --- a/akka-core/src/test/scala/ActorPatternsTest.scala +++ b/akka-core/src/test/scala/ActorPatternsTest.scala @@ -43,7 +43,7 @@ class ActorPatternsTest extends junit.framework.TestCase with Suite with MustMat b <- (d.!![Int](testMsg2,5000)) c <- (d.!![Int](testMsg3,5000)) } yield a + b + c - + result.get must be(21) for(a <- List(t1,t2,d)) a.stop } diff --git a/akka-core/src/test/scala/StmSpec.scala b/akka-core/src/test/scala/StmSpec.scala index 58b9b6805f..17d4be32bd 100644 --- a/akka-core/src/test/scala/StmSpec.scala +++ b/akka-core/src/test/scala/StmSpec.scala @@ -95,7 +95,7 @@ class StmSpec extends val size2: Int = (actor !! Size).getOrElse(fail("Could not get Vector::size")) size2 should equal(3) } catch { - case e => + case e => e.printStackTrace fail(e.toString) } @@ -122,7 +122,7 @@ class StmSpec extends val size4: Int = (actor !! Size).getOrElse(fail("Could not get size")) size4 should equal(3) } catch { - case e => + case e => fail(e.toString) } } @@ -130,7 +130,7 @@ class StmSpec extends /* describe("Multiverse API") { it("should blablabla") { - + import org.multiverse.api.programmatic._ // import org.multiverse.api._ import org.multiverse.templates._ @@ -139,13 +139,13 @@ class StmSpec extends import org.multiverse.api.{GlobalStmInstance, ThreadLocalTransaction, Transaction => MultiverseTransaction} import org.multiverse.api.lifecycle.{TransactionLifecycleListener, TransactionLifecycleEvent} import org.multiverse.commitbarriers._ - + def createRef[T]: ProgrammaticReference[T] = GlobalStmInstance .getGlobalStmInstance .getProgrammaticReferenceFactoryBuilder .build .atomicCreateReference(null.asInstanceOf[T]) - + val ref1 = Ref(0)//createRef[Int] val ref2 = Ref(0)//createRef[Int] @@ -185,13 +185,13 @@ class GlobalTransactionVectorTestActor extends Actor { import se.scalablesolutions.akka.stm.Transaction.Global private val vector: TransactionalVector[Int] = Global.atomic { TransactionalVector(1) } - + def receive = { - case Add(value) => + case Add(value) => Global.atomic { vector + value} self.reply(Success) - case Size => + case Size => val size = Global.atomic { vector.size } self.reply(size) } @@ -200,12 +200,12 @@ class GlobalTransactionVectorTestActor extends Actor { class NestedTransactorLevelOneActor extends Actor { import GlobalTransactionVectorTestActor._ private val nested = actorOf[NestedTransactorLevelTwoActor].start - + def receive = { - case add @ Add(_) => + case add @ Add(_) => self.reply((nested !! add).get) - case Size => + case Size => self.reply((nested !! Size).get) case "HiLevelOne" => println("HiLevelOne") @@ -216,15 +216,15 @@ class NestedTransactorLevelOneActor extends Actor { class NestedTransactorLevelTwoActor extends Actor { import GlobalTransactionVectorTestActor._ private val ref = Ref(0) - + def receive = { - case Add(value) => + case Add(value) => ref.swap(value) self.reply(Success) - case Size => + case Size => self.reply(ref.getOrElse(-1)) - + case "HiLevelTwo" => println("HiLevelTwo") } } diff --git a/akka-http/src/main/scala/AkkaLoader.scala b/akka-http/src/main/scala/AkkaLoader.scala index 431758e2c7..886c2c0688 100644 --- a/akka-http/src/main/scala/AkkaLoader.scala +++ b/akka-http/src/main/scala/AkkaLoader.scala @@ -47,29 +47,29 @@ class AkkaLoader extends Logging { private def printBanner = { log.info( """ - t - t t t - t t tt t - tt t t tt t - t ttttttt t ttt t - t tt ttt t ttt t - t t ttt t ttt t t - tt t ttt ttt ttt t - t t ttt ttt t tt t - t ttt ttt t t - tt ttt ttt t - ttt ttt - tttttttt ttt ttt ttt ttt tttttttt - ttt tt ttt ttt ttt ttt ttt ttt - ttt ttt ttt ttt ttt ttt ttt ttt - ttt ttt ttt ttt ttt tt ttt ttt - tttt ttttttttt tttttttt tttt - ttttttttt ttt ttt ttt ttt ttttttttt - ttt ttt ttt ttt ttt ttt ttt ttt - ttt ttt ttt ttt ttt ttt ttt ttt - ttt tt ttt ttt ttt ttt ttt ttt + t + t t t + t t tt t + tt t t tt t + t ttttttt t ttt t + t tt ttt t ttt t + t t ttt t ttt t t + tt t ttt ttt ttt t + t t ttt ttt t tt t + t ttt ttt t t + tt ttt ttt t + ttt ttt tttttttt ttt ttt ttt ttt tttttttt - + ttt tt ttt ttt ttt ttt ttt ttt + ttt ttt ttt ttt ttt ttt ttt ttt + ttt ttt ttt ttt ttt tt ttt ttt + tttt ttttttttt tttttttt tttt + ttttttttt ttt ttt ttt ttt ttttttttt + ttt ttt ttt ttt ttt ttt ttt ttt + ttt ttt ttt ttt ttt ttt ttt ttt + ttt tt ttt ttt ttt ttt ttt ttt + tttttttt ttt ttt ttt ttt tttttttt + ================================================== """) log.info(" Running version %s", Config.VERSION) diff --git a/akka-kernel/src/main/scala/EmbeddedAppServer.scala b/akka-kernel/src/main/scala/EmbeddedAppServer.scala index 5d3001edbd..8d9982c7e2 100644 --- a/akka-kernel/src/main/scala/EmbeddedAppServer.scala +++ b/akka-kernel/src/main/scala/EmbeddedAppServer.scala @@ -50,14 +50,14 @@ trait EmbeddedAppServer extends Bootable with Logging { Thread.currentThread.setContextClassLoader(applicationLoader.get) super.init(sc) } - finally { + finally { Thread.currentThread.setContextClassLoader(cl) } } }) adapter.setContextPath(uri.getPath) - adapter.addInitParameter("cometSupport", + adapter.addInitParameter("cometSupport", "org.atmosphere.container.GrizzlyCometSupport") adapter.addInitParameter("com.sun.jersey.config.property.resourceConfigClass", "com.sun.jersey.api.core.PackagesResourceConfig") diff --git a/akka-persistence/akka-persistence-cassandra/src/main/scala/CassandraStorageBackend.scala b/akka-persistence/akka-persistence-cassandra/src/main/scala/CassandraStorageBackend.scala index b67a0e9256..f393f4a162 100644 --- a/akka-persistence/akka-persistence-cassandra/src/main/scala/CassandraStorageBackend.scala +++ b/akka-persistence/akka-persistence-cassandra/src/main/scala/CassandraStorageBackend.scala @@ -42,7 +42,7 @@ private[akka] object CassandraStorageBackend extends case "ALL" => ConsistencyLevel.ALL case "ANY" => ConsistencyLevel.ANY case unknown => throw new IllegalArgumentException( - "Cassandra consistency level [" + unknown + "] is not supported." + + "Cassandra consistency level [" + unknown + "] is not supported." + "\n\tExpected one of [ZERO, ONE, QUORUM, DCQUORUM, DCQUORUMSYNC, ALL, ANY] in the akka.conf configuration file.") } } @@ -105,9 +105,9 @@ private[akka] object CassandraStorageBackend extends } } - def insertVectorStorageEntriesFor(name: String, elements: List[Array[Byte]]) = + def insertVectorStorageEntriesFor(name: String, elements: List[Array[Byte]]) = elements.foreach(insertVectorStorageEntryFor(name, _)) - + def updateVectorStorageEntryFor(name: String, index: Int, elem: Array[Byte]) = { val columnPath = new ColumnPath(VECTOR_COLUMN_PARENT.getColumn_family) columnPath.setColumn(intToBytes(index)) diff --git a/akka-persistence/akka-persistence-cassandra/src/test/scala/CassandraPersistentActorSpec.scala b/akka-persistence/akka-persistence-cassandra/src/test/scala/CassandraPersistentActorSpec.scala index 6746fca529..648dec2be3 100644 --- a/akka-persistence/akka-persistence-cassandra/src/test/scala/CassandraPersistentActorSpec.scala +++ b/akka-persistence/akka-persistence-cassandra/src/test/scala/CassandraPersistentActorSpec.scala @@ -171,4 +171,4 @@ object EmbeddedCassandraService { def start: Unit = {} } -*/ \ No newline at end of file +*/ diff --git a/akka-persistence/akka-persistence-common/src/main/scala/Storage.scala b/akka-persistence/akka-persistence-common/src/main/scala/Storage.scala index bcfb2c1025..6a0eb9a8d8 100644 --- a/akka-persistence/akka-persistence-common/src/main/scala/Storage.scala +++ b/akka-persistence/akka-persistence-common/src/main/scala/Storage.scala @@ -30,7 +30,7 @@ class StorageException(message: String) extends RuntimeException(message) *

  * val myMap = CassandraStorage.getMap(id)
  * 
- * + * * Example Java usage: *
  * PersistentMap myMap = MongoStorage.newMap();
@@ -72,7 +72,7 @@ trait Storage {
 }
 
 /**
- * Implementation of PersistentMap for every concrete 
+ * Implementation of PersistentMap for every concrete
  * storage will have the same workflow. This abstracts the workflow.
  *
  * Subclasses just need to provide the actual concrete instance for the
@@ -117,23 +117,23 @@ trait PersistentMap[K, V] extends scala.collection.mutable.Map[K, V]
     put(key, value)
     this
   }
-  
+
   override def put(key: K, value: V): Option[V] = {
     register
     newAndUpdatedEntries.put(key, value)
   }
- 
-  override def update(key: K, value: V) = { 
+
+  override def update(key: K, value: V) = {
     register
     newAndUpdatedEntries.update(key, value)
   }
-  
+
   override def remove(key: K) = {
     register
     removedEntries.add(key)
     newAndUpdatedEntries.get(key)
   }
-  
+
   def slice(start: Option[K], count: Int): List[Tuple2[K, V]] =
     slice(start, None, count)
 
@@ -141,11 +141,11 @@ trait PersistentMap[K, V] extends scala.collection.mutable.Map[K, V]
     storage.getMapStorageRangeFor(uuid, start, finish, count)
   } catch { case e: Exception => Nil }
 
-  override def clear = { 
+  override def clear = {
     register
     shouldClearOnCommit.swap(true)
   }
-  
+
   override def contains(key: K): Boolean = try {
     newAndUpdatedEntries.contains(key) ||
     storage.getMapStorageEntryFor(uuid, key).isDefined
@@ -163,9 +163,9 @@ trait PersistentMap[K, V] extends scala.collection.mutable.Map[K, V]
       storage.getMapStorageEntryFor(uuid, key)
     } catch { case e: Exception => None }
   }
-  
+
   def iterator = elements
-  
+
   override def elements: Iterator[Tuple2[K, V]]  = {
     new Iterator[Tuple2[K, V]] {
       private val originalList: List[Tuple2[K, V]] = try {
@@ -173,10 +173,10 @@ trait PersistentMap[K, V] extends scala.collection.mutable.Map[K, V]
       } catch {
         case e: Throwable => Nil
       }
-      private var elements = newAndUpdatedEntries.toList union originalList.reverse 
+      private var elements = newAndUpdatedEntries.toList union originalList.reverse
       override def next: Tuple2[K, V]= synchronized {
         val element = elements.head
-        elements = elements.tail        
+        elements = elements.tail
         element
       }
       override def hasNext: Boolean = synchronized { !elements.isEmpty }
@@ -217,12 +217,12 @@ trait PersistentVector[T] extends IndexedSeq[T] with Transactional with Committa
   }
 
   def +(elem: T) = add(elem)
-  
+
   def add(elem: T) = {
     register
     newElems + elem
   }
- 
+
   def apply(index: Int): T = get(index)
 
   def get(index: Int): T = {
@@ -231,7 +231,7 @@ trait PersistentVector[T] extends IndexedSeq[T] with Transactional with Committa
   }
 
   override def slice(start: Int, finish: Int): IndexedSeq[T] = slice(Some(start), Some(finish))
-  
+
   def slice(start: Option[Int], finish: Option[Int], count: Int = 0): IndexedSeq[T] = {
     val buffer = new scala.collection.mutable.ArrayBuffer[T]
     storage.getVectorStorageRangeFor(uuid, start, finish, count).foreach(buffer.append(_))
@@ -277,21 +277,21 @@ trait PersistentVector[T] extends IndexedSeq[T] with Transactional with Committa
  */
 trait PersistentRef[T] extends Transactional with Committable with Abortable {
   protected val ref = new TransactionalRef[T]
-  
+
   val storage: RefStorageBackend[T]
 
   def commit = if (ref.isDefined) {
     storage.insertRefStorageFor(uuid, ref.get.get)
-    ref.swap(null.asInstanceOf[T]) 
+    ref.swap(null.asInstanceOf[T])
   }
 
-  def abort = ref.swap(null.asInstanceOf[T]) 
+  def abort = ref.swap(null.asInstanceOf[T])
 
   def swap(elem: T) = {
     register
     ref.swap(elem)
   }
-  
+
   def get: Option[T] = if (ref.isDefined) ref.get else storage.getRefStorageFor(uuid)
 
   def isDefined: Boolean = ref.isDefined || storage.getRefStorageFor(uuid).isDefined
@@ -309,7 +309,7 @@ trait PersistentRef[T] extends Transactional with Committable with Abortable {
 }
 
 /**
- * Implementation of PersistentQueue for every concrete 
+ * Implementation of PersistentQueue for every concrete
  * storage will have the same workflow. This abstracts the workflow.
  * 

* Enqueue is simpler, we just have to record the operation in a local @@ -410,13 +410,13 @@ trait PersistentQueue[A] extends scala.collection.mutable.Queue[A] } } - override def clear = { + override def clear = { register shouldClearOnCommit.swap(true) localQ.swap(Queue.empty) pickMeForDQ.swap(0) } - + override def size: Int = try { storage.size(uuid) + localQ.get.get.length } catch { case e: Exception => 0 } @@ -424,11 +424,11 @@ trait PersistentQueue[A] extends scala.collection.mutable.Queue[A] override def isEmpty: Boolean = size == 0 - override def +=(elem: A) = { + override def +=(elem: A) = { enqueue(elem) this } - def ++=(elems: Iterator[A]) = { + def ++=(elems: Iterator[A]) = { enqueue(elems.toList: _*) this } @@ -450,7 +450,7 @@ trait PersistentQueue[A] extends scala.collection.mutable.Queue[A] * Implements a template for a concrete persistent transactional sorted set based storage. *

* Sorting is done based on a zscore. But the computation of zscore has been kept - * outside the abstraction. + * outside the abstraction. *

* zscore can be implemented in a variety of ways by the calling class: *

@@ -467,7 +467,7 @@ trait PersistentQueue[A] extends scala.collection.mutable.Queue[A]
  * class Foo {
  *   //..
  * }
- * 
+ *
  * implicit def Foo2Scorable(foo: Foo): ZScorable = new ZScorable {
  *   def toZScore = {
  *     //..
@@ -526,7 +526,7 @@ trait PersistentSortedSet[A] extends Transactional with Committable with Abortab
       }
     }
   }
- 
+
   def size: Int = newElems.size + storage.zcard(uuid) - removedElems.size
 
   def zscore(elem: A): Float = {
@@ -541,9 +541,9 @@ trait PersistentSortedSet[A] extends Transactional with Committable with Abortab
   implicit def order(x: (A, Float)) = new Ordered[(A, Float)] {
     def compare(that: (A, Float)) = x._2 compare that._2
   }
-  
+
   implicit def ordering = new scala.math.Ordering[(A,Float)] {
-    def compare(x: (A, Float),y : (A,Float)) = x._2 compare y._2   
+    def compare(x: (A, Float),y : (A,Float)) = x._2 compare y._2
   }
 
 
@@ -556,7 +556,7 @@ trait PersistentSortedSet[A] extends Transactional with Committable with Abortab
 
     // -1 means the last element, -2 means the second last
     val s = if (start < 0) start + l else start
-    val e = 
+    val e =
       if (end < 0) end + l
       else if (end >= l) (l - 1)
       else end
diff --git a/akka-persistence/akka-persistence-common/src/main/scala/StorageBackend.scala b/akka-persistence/akka-persistence-common/src/main/scala/StorageBackend.scala
index ab0cfaf4d3..a5226eb1a4 100644
--- a/akka-persistence/akka-persistence-common/src/main/scala/StorageBackend.scala
+++ b/akka-persistence/akka-persistence-common/src/main/scala/StorageBackend.scala
@@ -26,7 +26,7 @@ trait VectorStorageBackend[T] extends StorageBackend {
   def updateVectorStorageEntryFor(name: String, index: Int, elem: T)
   def getVectorStorageEntryFor(name: String, index: Int): T
   def getVectorStorageRangeFor(name: String, start: Option[Int], finish: Option[Int], count: Int): List[T]
-  def getVectorStorageSizeFor(name: String): Int 
+  def getVectorStorageSizeFor(name: String): Int
 }
 
 // for Ref
@@ -47,17 +47,17 @@ trait RefStorageBackend[T] extends StorageBackend {
 trait QueueStorageBackend[T] extends StorageBackend {
   // add to the end of the queue
   def enqueue(name: String, item: T): Boolean
-  
+
   // pop from the front of the queue
   def dequeue(name: String): Option[T]
-  
+
   // get the size of the queue
   def size(name: String): Int
-  
+
   // return an array of items currently stored in the queue
   // start is the item to begin, count is how many items to return
   def peek(name: String, start: Int, count: Int): List[T]
-  
+
   // completely delete the queue
   def remove(name: String): Boolean
 }
@@ -65,19 +65,19 @@ trait QueueStorageBackend[T] extends StorageBackend {
 trait SortedSetStorageBackend[T] extends StorageBackend {
   // add item to sorted set identified by name
   def zadd(name: String, zscore: String, item: T): Boolean
-  
+
   // remove item from sorted set identified by name
   def zrem(name: String, item: T): Boolean
-  
+
   // cardinality of the set identified by name
   def zcard(name: String): Int
-  
+
   // zscore of the item from sorted set identified by name
   def zscore(name: String, item: T): Option[Float]
-  
+
   // zrange from the sorted set identified by name
   def zrange(name: String, start: Int, end: Int): List[T]
 
   // zrange with score from the sorted set identified by name
-  def zrangeWithScore(name: String, start: Int, end: Int): List[(T, Float)] 
+  def zrangeWithScore(name: String, start: Int, end: Int): List[(T, Float)]
 }
diff --git a/akka-persistence/akka-persistence-mongo/src/main/scala/MongoStorageBackend.scala b/akka-persistence/akka-persistence-mongo/src/main/scala/MongoStorageBackend.scala
index 49f46db134..d5581b373b 100644
--- a/akka-persistence/akka-persistence-mongo/src/main/scala/MongoStorageBackend.scala
+++ b/akka-persistence/akka-persistence-mongo/src/main/scala/MongoStorageBackend.scala
@@ -280,7 +280,7 @@ private[akka] object MongoStorageBackend extends
       }
     val currentList = dbobj.get(VALUE).asInstanceOf[JArrayList[AnyRef]]
     currentList.set(index, serializer.out(elem))
-    coll.update(q, 
+    coll.update(q,
       new BasicDBObject().append(KEY, name).append(VALUE, currentList))
   }
 
diff --git a/akka-samples/akka-sample-ants/src/main/scala/Ants.scala b/akka-samples/akka-sample-ants/src/main/scala/Ants.scala
index 85757f25e8..05fe245b10 100644
--- a/akka-samples/akka-sample-ants/src/main/scala/Ants.scala
+++ b/akka-samples/akka-sample-ants/src/main/scala/Ants.scala
@@ -87,7 +87,7 @@ object World {
     pingEvery(evaporator, EvapMillis)
   }
 
-  private def pingEvery(actor: ActorRef, millis: Long) = 
+  private def pingEvery(actor: ActorRef, millis: Long) =
     Scheduler.schedule(actor, "ping", Config.StartDelay, millis, TimeUnit.MILLISECONDS)
 }
 
diff --git a/akka-samples/akka-sample-camel/src/main/scala/Application1.scala b/akka-samples/akka-sample-camel/src/main/scala/Application1.scala
index dfff0e0539..6dcb437992 100644
--- a/akka-samples/akka-sample-camel/src/main/scala/Application1.scala
+++ b/akka-samples/akka-sample-camel/src/main/scala/Application1.scala
@@ -26,4 +26,4 @@ object Application1 {
     println(actor2 !! Message("actor2"))
   }
 
-}
\ No newline at end of file
+}
diff --git a/akka-samples/akka-sample-camel/src/main/scala/Application2.scala b/akka-samples/akka-sample-camel/src/main/scala/Application2.scala
index e01b510a71..8a789d13bf 100644
--- a/akka-samples/akka-sample-camel/src/main/scala/Application2.scala
+++ b/akka-samples/akka-sample-camel/src/main/scala/Application2.scala
@@ -19,4 +19,4 @@ object Application2 {
     RemoteNode.start("localhost", 7777)
     RemoteNode.register("remote2", actorOf[RemoteActor2].start)
   }
-}
\ No newline at end of file
+}
diff --git a/akka-samples/akka-sample-camel/src/main/scala/Boot.scala b/akka-samples/akka-sample-camel/src/main/scala/Boot.scala
index 5b08e15a1a..206661c8fd 100644
--- a/akka-samples/akka-sample-camel/src/main/scala/Boot.scala
+++ b/akka-samples/akka-sample-camel/src/main/scala/Boot.scala
@@ -74,4 +74,4 @@ class CustomRouteBuilder extends RouteBuilder {
       }
     })
   }
-}
\ No newline at end of file
+}
diff --git a/akka-samples/akka-sample-chat/src/main/scala/ChatServer.scala b/akka-samples/akka-sample-chat/src/main/scala/ChatServer.scala
index f193efd217..f244f8eeef 100644
--- a/akka-samples/akka-sample-chat/src/main/scala/ChatServer.scala
+++ b/akka-samples/akka-sample-chat/src/main/scala/ChatServer.scala
@@ -27,17 +27,17 @@ First we need to download, build and start up Redis:
 4. Run: ‘./redis-server’.
 For details on how to set up Redis server have a look at http://code.google.com/p/redis/wiki/QuickStart.
 
-Then to run the sample: 
+Then to run the sample:
 
 1. Fire up two shells. For each of them:
   - Step down into to the root of the Akka distribution.
   - Set 'export AKKA_HOME=.
   - Run 'sbt console' to start up a REPL (interpreter).
-2. In the first REPL you get execute: 
+2. In the first REPL you get execute:
   - scala> import sample.chat._
   - scala> import se.scalablesolutions.akka.actor.Actor._
   - scala> val chatService = actorOf[ChatService].start
-3. In the second REPL you get execute: 
+3. In the second REPL you get execute:
     - scala> import sample.chat._
     - scala> Runner.run
 4. See the chat simulation run.
@@ -60,12 +60,12 @@ case class ChatMessage(from: String, message: String) extends Event
 /**
  * Chat client.
  */
-class ChatClient(val name: String) { 
+class ChatClient(val name: String) {
   val chat = RemoteClient.actorFor("chat:service", "localhost", 9999)
 
-  def login =                 chat ! Login(name) 
-  def logout =                chat ! Logout(name)  
-  def post(message: String) = chat ! ChatMessage(name, name + ": " + message)  
+  def login =                 chat ! Login(name)
+  def logout =                chat ! Logout(name)
+  def post(message: String) = chat ! ChatMessage(name, name + ": " + message)
   def chatLog: ChatLog =     (chat !! GetChatLog(name)).getOrElse(throw new Exception("Couldn't get the chat log from ChatServer"))
 }
 
@@ -75,15 +75,15 @@ class ChatClient(val name: String) {
 class Session(user: String, storage: ActorRef) extends Actor {
   private val loginTime = System.currentTimeMillis
   private var userLog: List[String] = Nil
-  
+
   log.info("New session for user [%s] has been created at [%s]", user, loginTime)
 
   def receive = {
-    case msg @ ChatMessage(from, message) => 
+    case msg @ ChatMessage(from, message) =>
       userLog ::= message
       storage ! msg
-      
-    case msg @ GetChatLog(_) => 
+
+    case msg @ GetChatLog(_) =>
       storage forward msg
   }
 }
@@ -97,24 +97,24 @@ trait ChatStorage extends Actor
  * Redis-backed chat storage implementation.
  */
 class RedisChatStorage extends ChatStorage {
-  self.lifeCycle = Some(LifeCycle(Permanent))    
+  self.lifeCycle = Some(LifeCycle(Permanent))
   val CHAT_LOG = "akka.chat.log"
-  
+
   private var chatLog = atomic { RedisStorage.getVector(CHAT_LOG) }
 
   log.info("Redis-based chat storage is starting up...")
 
   def receive = {
-    case msg @ ChatMessage(from, message) => 
+    case msg @ ChatMessage(from, message) =>
       log.debug("New chat message [%s]", message)
       atomic { chatLog + message.getBytes("UTF-8") }
 
-    case GetChatLog(_) => 
+    case GetChatLog(_) =>
       val messageList = atomic { chatLog.map(bytes => new String(bytes, "UTF-8")).toList }
       self.reply(ChatLog(messageList))
   }
-  
-  override def postRestart(reason: Throwable) = chatLog = RedisStorage.getVector(CHAT_LOG)  
+
+  override def postRestart(reason: Throwable) = chatLog = RedisStorage.getVector(CHAT_LOG)
 }
 
 /**
@@ -122,27 +122,27 @@ class RedisChatStorage extends ChatStorage {
  * 

* Uses self-type annotation (this: Actor =>) to declare that it needs to be mixed in with an Actor. */ -trait SessionManagement { this: Actor => - +trait SessionManagement { this: Actor => + val storage: ActorRef // needs someone to provide the ChatStorage val sessions = new HashMap[String, ActorRef] - + protected def sessionManagement: Receive = { - case Login(username) => + case Login(username) => log.info("User [%s] has logged in", username) val session = actorOf(new Session(username, storage)) session.start sessions += (username -> session) - - case Logout(username) => + + case Logout(username) => log.info("User [%s] has logged out", username) val session = sessions(username) session.stop - sessions -= username - } - - protected def shutdownSessions = - sessions.foreach { case (_, session) => session.stop } + sessions -= username + } + + protected def shutdownSessions = + sessions.foreach { case (_, session) => session.stop } } /** @@ -152,7 +152,7 @@ trait SessionManagement { this: Actor => */ trait ChatManagement { this: Actor => val sessions: HashMap[String, ActorRef] // needs someone to provide the Session map - + protected def chatManagement: Receive = { case msg @ ChatMessage(from, _) => sessions(from) ! msg case msg @ GetChatLog(from) => sessions(from) forward msg @@ -172,20 +172,20 @@ trait RedisChatStorageFactory { this: Actor => trait ChatServer extends Actor { self.faultHandler = Some(OneForOneStrategy(5, 5000)) self.trapExit = List(classOf[Exception]) - + val storage: ActorRef log.info("Chat server is starting up...") // actor message handler def receive = sessionManagement orElse chatManagement - + // abstract methods to be defined somewhere else protected def chatManagement: Receive - protected def sessionManagement: Receive + protected def sessionManagement: Receive protected def shutdownSessions: Unit - override def shutdown = { + override def shutdown = { log.info("Chat server is shutting down...") shutdownSessions self.unlink(storage) @@ -200,10 +200,10 @@ trait ChatServer extends Actor { * val chatService = Actor.actorOf[ChatService].start *

*/ -class ChatService extends - ChatServer with - SessionManagement with - ChatManagement with +class ChatService extends + ChatServer with + SessionManagement with + ChatManagement with RedisChatStorageFactory { override def init = { RemoteNode.start("localhost", 9999) @@ -217,7 +217,7 @@ class ChatService extends object Runner { def run = { val client = new ChatClient("jonas") - + client.login client.post("Hi there") @@ -228,4 +228,4 @@ object Runner { client.logout } -} \ No newline at end of file +} diff --git a/akka-samples/akka-sample-lift/src/main/scala/bootstrap/liftweb/Boot.scala b/akka-samples/akka-sample-lift/src/main/scala/bootstrap/liftweb/Boot.scala index f8e4f15bd9..0f4a0e9020 100644 --- a/akka-samples/akka-sample-lift/src/main/scala/bootstrap/liftweb/Boot.scala +++ b/akka-samples/akka-sample-lift/src/main/scala/bootstrap/liftweb/Boot.scala @@ -23,7 +23,7 @@ class Boot extends Logging { def boot { // where to search snippet LiftRules.addToPackages("sample.lift") - + LiftRules.httpAuthProtectedResource.prepend { case (Req("liftcount" :: Nil, _, _)) => Full(AuthRole("admin")) } @@ -35,9 +35,9 @@ class Boot extends Logging { true } } - + LiftRules.passNotFoundToChain = true - + val factory = SupervisorFactory( SupervisorConfig( RestartStrategy(OneForOne, 3, 100, List(classOf[Exception])), @@ -49,7 +49,7 @@ class Boot extends Logging { LifeCycle(Permanent)) :: Nil)) factory.newInstance.start - + // Build SiteMap // val entries = Menu(Loc("Home", List("index"), "Home")) :: Nil // LiftRules.setSiteMap(SiteMap(entries:_*)) diff --git a/akka-samples/akka-sample-lift/src/test/scala/LiftConsole.scala b/akka-samples/akka-sample-lift/src/test/scala/LiftConsole.scala index a5aa1698e8..43296bc1f4 100644 --- a/akka-samples/akka-sample-lift/src/test/scala/LiftConsole.scala +++ b/akka-samples/akka-sample-lift/src/test/scala/LiftConsole.scala @@ -13,4 +13,4 @@ object LiftConsole { exit(0) } } -*/ \ No newline at end of file +*/ diff --git a/akka-samples/akka-sample-pubsub/src/main/scala/RedisPubSub.scala b/akka-samples/akka-sample-pubsub/src/main/scala/RedisPubSub.scala index 44366d31a4..ca3148aab7 100644 --- a/akka-samples/akka-sample-pubsub/src/main/scala/RedisPubSub.scala +++ b/akka-samples/akka-sample-pubsub/src/main/scala/RedisPubSub.scala @@ -10,7 +10,7 @@ import se.scalablesolutions.akka.actor.Actor._ /** * Sample Akka application for Redis PubSub - * + * * Prerequisite: Need Redis Server running (the version that supports pubsub) *
  * 1. Download redis from http://github.com/antirez/redis
@@ -65,7 +65,7 @@ object Sub {
   val r = new RedisClient("localhost", 6379)
   val s = actorOf(new Subscriber(r))
   s.start
-  s ! Register(callback) 
+  s ! Register(callback)
 
   def sub(channels: String*) = {
     s ! Subscribe(channels.toArray)
@@ -78,29 +78,29 @@ object Sub {
   def callback(pubsub: PubSubMessage) = pubsub match {
     case S(channel, no) => println("subscribed to " + channel + " and count = " + no)
     case U(channel, no) => println("unsubscribed from " + channel + " and count = " + no)
-    case M(channel, msg) => 
+    case M(channel, msg) =>
       msg match {
         // exit will unsubscribe from all channels and stop subscription service
-        case "exit" => 
+        case "exit" =>
           println("unsubscribe all ..")
           r.unsubscribe
 
         // message "+x" will subscribe to channel x
-        case x if x startsWith "+" => 
+        case x if x startsWith "+" =>
           val s: Seq[Char] = x
           s match {
             case Seq('+', rest @ _*) => r.subscribe(rest.toString){ m => }
           }
 
         // message "-x" will unsubscribe from channel x
-        case x if x startsWith "-" => 
+        case x if x startsWith "-" =>
           val s: Seq[Char] = x
           s match {
             case Seq('-', rest @ _*) => r.unsubscribe(rest.toString)
           }
 
         // other message receive
-        case x => 
+        case x =>
           println("received message on channel " + channel + " as : " + x)
       }
   }
diff --git a/akka-samples/akka-sample-remote/src/main/scala/ClientManagedRemoteActorSample.scala b/akka-samples/akka-sample-remote/src/main/scala/ClientManagedRemoteActorSample.scala
index 9070d4a7f8..24f81872f7 100644
--- a/akka-samples/akka-sample-remote/src/main/scala/ClientManagedRemoteActorSample.scala
+++ b/akka-samples/akka-sample-remote/src/main/scala/ClientManagedRemoteActorSample.scala
@@ -11,7 +11,7 @@ import se.scalablesolutions.akka.util.Logging
 
 class RemoteHelloWorldActor extends RemoteActor("localhost", 9999) {
   def receive = {
-    case "Hello" => 
+    case "Hello" =>
       log.info("Received 'Hello'")
       self.reply("World")
   }
@@ -27,7 +27,7 @@ object ClientManagedRemoteActorServer extends Logging {
 }
 
 object ClientManagedRemoteActorClient extends Logging {
-  
+
   def run = {
     val actor = actorOf[RemoteHelloWorldActor].start
     log.info("Remote actor created, moved to the server")
diff --git a/akka-samples/akka-sample-remote/src/main/scala/ServerManagedRemoteActorSample.scala b/akka-samples/akka-sample-remote/src/main/scala/ServerManagedRemoteActorSample.scala
index 87253af24b..2671d9e25e 100644
--- a/akka-samples/akka-sample-remote/src/main/scala/ServerManagedRemoteActorSample.scala
+++ b/akka-samples/akka-sample-remote/src/main/scala/ServerManagedRemoteActorSample.scala
@@ -11,7 +11,7 @@ import se.scalablesolutions.akka.util.Logging
 
 class HelloWorldActor extends Actor {
   def receive = {
-    case "Hello" => 
+    case "Hello" =>
       log.info("Received 'Hello'")
       self.reply("World")
   }
@@ -30,7 +30,7 @@ object ServerManagedRemoteActorServer extends Logging {
 }
 
 object ServerManagedRemoteActorClient extends Logging {
-  
+
   def run = {
     val actor = RemoteClient.actorFor("hello-service", "localhost", 9999)
     log.info("Remote client created")
diff --git a/akka-samples/akka-sample-rest-scala/src/main/scala/SimpleService.scala b/akka-samples/akka-sample-rest-scala/src/main/scala/SimpleService.scala
index 18b8bda6e6..eb2d07340e 100644
--- a/akka-samples/akka-sample-rest-scala/src/main/scala/SimpleService.scala
+++ b/akka-samples/akka-sample-rest-scala/src/main/scala/SimpleService.scala
@@ -158,7 +158,7 @@ class Chat {
 }
 
 object ChatActor {
-  case class ChatMsg(val who: String, val what: String, val msg: String)        
+  case class ChatMsg(val who: String, val what: String, val msg: String)
 }
 
 class ChatActor extends Actor with Logging {
diff --git a/akka-samples/akka-sample-security/src/main/scala/SimpleService.scala b/akka-samples/akka-sample-security/src/main/scala/SimpleService.scala
index 800fdca7ff..e6892c7b62 100644
--- a/akka-samples/akka-sample-security/src/main/scala/SimpleService.scala
+++ b/akka-samples/akka-sample-security/src/main/scala/SimpleService.scala
@@ -147,4 +147,4 @@ class SecureTickActor extends Transactor with Logging {
       self.reply(new Integer(0))
     }
   }
-}
\ No newline at end of file
+}
diff --git a/config/akka-reference.conf b/config/akka-reference.conf
index 83c35d143f..81f2ee4580 100644
--- a/config/akka-reference.conf
+++ b/config/akka-reference.conf
@@ -1,10 +1,10 @@
 ####################
 # Akka Config File #
 ####################
- 
+
 # This file has all the default settings, so all these could be removed with no visible effect.
 # Modify as needed.
- 
+
 
   filename = "./logs/akka.log"
   roll = "daily"  # Options: never, hourly, daily, sunday/monday/...
@@ -13,14 +13,14 @@
   # syslog_host = ""
   # syslog_server_name = ""
 
- 
+
 
   version = "0.9"
- 
+
   # FQN (Fully Qualified Name) to the class doing initial active object/actor
   # supervisor bootstrap, should be defined in default constructor
   boot = ["sample.camel.Boot",
-          "sample.rest.java.Boot", 
+          "sample.rest.java.Boot",
           "sample.rest.scala.Boot",
           "sample.security.Boot"]
 
@@ -41,10 +41,10 @@
     jta-aware = off          # 'on' means that if there JTA Transaction Manager available then the STM will
                              # begin (or join), commit or rollback the JTA transaction. Default is 'off'.
   
-  
+
   
-    provider = "from-jndi"   # Options: "from-jndi" (means that Akka will try to detect a TransactionManager in the JNDI) 
-                             #          "atomikos" (means that Akka will use the Atomikos based JTA impl in 'akka-jta', 
+    provider = "from-jndi"   # Options: "from-jndi" (means that Akka will try to detect a TransactionManager in the JNDI)
+                             #          "atomikos" (means that Akka will use the Atomikos based JTA impl in 'akka-jta',
                              #          e.g. you need the akka-jta JARs on classpath).
     timeout = 60000
   
@@ -56,7 +56,7 @@
     filters = ["se.scalablesolutions.akka.security.AkkaSecurityFilterFactory"] # List with all jersey filters to use
     resource_packages = ["sample.rest.scala","sample.rest.java","sample.security"] # List with all resource packages for your Jersey services
     authenticator = "sample.security.BasicAuthenticationService" # The authentication service to use. Need to be overridden (uses sample now)
-    
+
     #IF you are using a KerberosAuthenticationActor
     #   
     #     servicePrincipal = "HTTP/localhost@EXAMPLE.COM"
@@ -75,8 +75,8 @@
       name = "default"                                                        # The name of the cluster
       serializer = "se.scalablesolutions.akka.serialization.Serializer$Java$" # FQN of the serializer class
     
-    
-      
+
+    
       service = on
       hostname = "localhost"
       port = 9999
@@ -88,14 +88,14 @@
       read-timeout = 10000      # in millis (10 sec default)
     
   
- 
+
   
     
       hostname = "127.0.0.1"        # IP address or hostname of one of the Cassandra cluster's seeds
       port = 9160
       consistency-level = "QUORUM"  # Options: ZERO, ONE, QUORUM, DCQUORUM, DCQUORUMSYNC, ALL, ANY
     
-    
+
     
       hostname = "127.0.0.1"        # IP address or hostname of the MongoDB DB instance
       port = 27017
diff --git a/config/log4j.properties b/config/log4j.properties
index 903ccd16af..76d3661163 100755
--- a/config/log4j.properties
+++ b/config/log4j.properties
@@ -1,20 +1,20 @@
-# for production, you should probably set the root to INFO
-# and the pattern to %c instead of %l.  (%l is slower.)
-
-# output messages into a rolling log file as well as stdout
-log4j.rootLogger=INFO,stdout,R
-
-# stdout
-log4j.appender.stdout=org.apache.log4j.ConsoleAppender
-log4j.appender.stdout.layout=org.apache.log4j.SimpleLayout
-
-# rolling log file ("system.log
-log4j.appender.R=org.apache.log4j.DailyRollingFileAppender
-log4j.appender.R.DatePattern='.'yyyy-MM-dd-HH
-log4j.appender.R.layout=org.apache.log4j.PatternLayout
-log4j.appender.R.layout.ConversionPattern=%5p [%t] %d{ISO8601} %F (line %L) %m%n
-
-# Edit the next line to point to your logs directory
-log4j.appender.R.File=./logs/akka.log
-
-log4j.logger.org.atmosphere=DEBUG
+# for production, you should probably set the root to INFO
+# and the pattern to %c instead of %l.  (%l is slower.)
+
+# output messages into a rolling log file as well as stdout
+log4j.rootLogger=INFO,stdout,R
+
+# stdout
+log4j.appender.stdout=org.apache.log4j.ConsoleAppender
+log4j.appender.stdout.layout=org.apache.log4j.SimpleLayout
+
+# rolling log file ("system.log
+log4j.appender.R=org.apache.log4j.DailyRollingFileAppender
+log4j.appender.R.DatePattern='.'yyyy-MM-dd-HH
+log4j.appender.R.layout=org.apache.log4j.PatternLayout
+log4j.appender.R.layout.ConversionPattern=%5p [%t] %d{ISO8601} %F (line %L) %m%n
+
+# Edit the next line to point to your logs directory
+log4j.appender.R.File=./logs/akka.log
+
+log4j.logger.org.atmosphere=DEBUG
diff --git a/config/storage-conf.xml b/config/storage-conf.xml
index 06ba8007a2..2647fdcd0e 100644
--- a/config/storage-conf.xml
+++ b/config/storage-conf.xml
@@ -21,15 +21,15 @@
   
   
 
-  
   akka
 
   
-      
       
-      
-      
 
       
       org.apache.cassandra.locator.EndPointSnitch
-        
+
     
   
 
@@ -158,7 +158,7 @@
    ~ Authenticator: any IAuthenticator may be used, including your own as long
    ~ as it is on the classpath.  Out of the box, Cassandra provides
    ~ org.apache.cassandra.auth.AllowAllAuthenticator and,
-   ~ org.apache.cassandra.auth.SimpleAuthenticator 
+   ~ org.apache.cassandra.auth.SimpleAuthenticator
    ~ (SimpleAuthenticator uses access.properties and passwd.properties by
    ~ default).
    ~
@@ -188,7 +188,7 @@
    ~ are sent to the node with the "closest" token, so distributing your
    ~ tokens equally along the key distribution space will spread keys
    ~ evenly across your cluster.)  This setting is only checked the first
-   ~ time a node is started. 
+   ~ time a node is started.
 
    ~ This can also be useful with RandomPartitioner to force equal spacing
    ~ of tokens around the hash space, especially for clusters with a small
@@ -227,9 +227,9 @@
 
   
 
-  
   9160
-  
   false
@@ -285,16 +285,16 @@
 
   
   64
 
   
@@ -314,7 +314,7 @@
 
   
-   
+