diff --git a/akka-amqp/src/main/scala/AMQP.scala b/akka-amqp/src/main/scala/AMQP.scala
index 81c6e5a64a..a03a2cb056 100644
--- a/akka-amqp/src/main/scala/AMQP.scala
+++ b/akka-amqp/src/main/scala/AMQP.scala
@@ -13,8 +13,7 @@ import se.scalablesolutions.akka.config.ScalaConfig._
import se.scalablesolutions.akka.util.{HashCode, Logging}
import scala.collection.mutable.HashMap
-
-import org.scala_tools.javautils.Imports._
+import scala.collection.JavaConversions._
import java.util.concurrent.ConcurrentHashMap
import java.util.{Timer, TimerTask}
@@ -411,7 +410,7 @@ object AMQP {
protected def setupChannel = {
connection = connectionFactory.newConnection(hostname, port)
channel = connection.createChannel
- channel.exchangeDeclare(exchangeName.toString, exchangeType.toString, passive, durable, autoDelete, configurationArguments.asJava)
+ channel.exchangeDeclare(exchangeName.toString, exchangeType.toString, passive, durable, autoDelete, configurationArguments)
listeners.elements.toList.map(_._2).foreach(registerListener)
if (shutdownListener.isDefined) connection.addShutdownListener(shutdownListener.get)
}
@@ -426,7 +425,7 @@ object AMQP {
listener.queueName,
passive, durable,
listener.exclusive, listener.autoDelete,
- configurationArguments.asJava)
+ configurationArguments)
}
log.debug("Binding new queue for MessageConsumerListener [%s]", listener.queueName)
diff --git a/akka-cluster/akka-cluster-jgroups/src/main/scala/JGroupsClusterActor.scala b/akka-cluster/akka-cluster-jgroups/src/main/scala/JGroupsClusterActor.scala
index 2b6730dd60..afc917260e 100644
--- a/akka-cluster/akka-cluster-jgroups/src/main/scala/JGroupsClusterActor.scala
+++ b/akka-cluster/akka-cluster-jgroups/src/main/scala/JGroupsClusterActor.scala
@@ -8,7 +8,7 @@ import org.jgroups.{JChannel, View => JG_VIEW, Address, Message => JG_MSG, Exten
*/
class JGroupsClusterActor extends BasicClusterActor {
import ClusterActor._
- import org.scala_tools.javautils.Imports._
+ import scala.collection.JavaConversions._
type ADDR_T = Address
@@ -32,7 +32,7 @@ class JGroupsClusterActor extends BasicClusterActor {
if (isActive && m.getSrc != channel.map(_.getAddress).getOrElse(m.getSrc)) me send Message(m.getSrc,m.getRawBuffer)
def viewAccepted(view: JG_VIEW): Unit =
- if (isActive) me send View(Set[ADDR_T]() ++ view.getMembers.asScala - channel.get.getAddress)
+ if (isActive) me send View(Set[ADDR_T]() ++ view.getMembers - channel.get.getAddress)
def suspect(a: Address): Unit =
if (isActive) me send Zombie(a)
diff --git a/akka-cluster/akka-cluster-shoal/src/main/scala/ShoalClusterActor.scala b/akka-cluster/akka-cluster-shoal/src/main/scala/ShoalClusterActor.scala
index c656bd4b81..007cc96b9a 100644
--- a/akka-cluster/akka-cluster-shoal/src/main/scala/ShoalClusterActor.scala
+++ b/akka-cluster/akka-cluster-shoal/src/main/scala/ShoalClusterActor.scala
@@ -84,7 +84,7 @@ class ShoalClusterActor extends BasicClusterActor {
* Creates a CallBack instance that deals with the cluster signalling
*/
protected def createCallback : CallBack = {
- import org.scala_tools.javautils.Imports._
+ import scala.collection.JavaConversions._
val me = this
new CallBack {
def processNotification(signal : Signal) {
@@ -93,7 +93,7 @@ class ShoalClusterActor extends BasicClusterActor {
if(isActive) {
signal match {
case ms : MessageSignal => me send Message(ms.getMemberToken,ms.getMessage)
- case jns : JoinNotificationSignal => me send View(Set[ADDR_T]() ++ jns.getCurrentCoreMembers.asScala - serverName)
+ case jns : JoinNotificationSignal => me send View(Set[ADDR_T]() ++ jns.getCurrentCoreMembers - serverName)
case fss : FailureSuspectedSignal => me send Zombie(fss.getMemberToken)
case fns : FailureNotificationSignal => me send Zombie(fns.getMemberToken)
case _ => log.debug("Unhandled signal: [%s]",signal)
diff --git a/akka-comet/src/main/scala/AkkaServlet.scala b/akka-comet/src/main/scala/AkkaServlet.scala
index 6e37d56c49..cae6dd3000 100644
--- a/akka-comet/src/main/scala/AkkaServlet.scala
+++ b/akka-comet/src/main/scala/AkkaServlet.scala
@@ -59,12 +59,12 @@ class AkkaServlet extends org.atmosphere.cpr.AtmosphereServlet with Logging {
* we need to handle that.
*/
override def createCometSupportResolver() : CometSupportResolver = {
- import org.scala_tools.javautils.Imports._
+ import scala.collection.JavaConversions._
new DefaultCometSupportResolver(config) {
type CS = CometSupport[_ <: AtmosphereResource[_,_]]
override def resolveMultipleNativeSupportConflict(available : JList[Class[_ <: CS]]) : CS = {
- available.asScala.filter(_ != classOf[GrizzlyCometSupport]).toList match {
+ available.filter(_ != classOf[GrizzlyCometSupport]).toList match {
case Nil => new GrizzlyCometSupport(config)
case x :: Nil => newCometSupport(x.asInstanceOf[Class[_ <: CS]])
case _ => super.resolveMultipleNativeSupportConflict(available)
diff --git a/akka-core/pom.xml b/akka-core/pom.xml
index 14ebdcfd14..10e05fe701 100644
--- a/akka-core/pom.xml
+++ b/akka-core/pom.xml
@@ -46,11 +46,6 @@
netty
3.2.0.ALPHA3
-
- org.scala-tools
- javautils
- 2.7.4-0.1
-
@@ -68,11 +63,11 @@
jackson-mapper-asl
1.2.1
-
+
net.databinder
dispatch-json_2.8.0.Beta1
@@ -91,7 +86,7 @@
sjson.json
sjson
- 0.4
+ 0.5-SNAPSHOT-2.8.Beta1
diff --git a/akka-core/src/main/scala/actor/ActiveObject.scala b/akka-core/src/main/scala/actor/ActiveObject.scala
index 1858952f40..56d78268f5 100644
--- a/akka-core/src/main/scala/actor/ActiveObject.scala
+++ b/akka-core/src/main/scala/actor/ActiveObject.scala
@@ -260,9 +260,10 @@ private[akka] sealed class ActiveObjectAspect {
}
private def localDispatch(joinPoint: JoinPoint): AnyRef = {
- import Actor.Sender.Self
val rtti = joinPoint.getRtti.asInstanceOf[MethodRtti]
- if (isOneWay(rtti)) actor ! Invocation(joinPoint, true, true)
+ if (isOneWay(rtti)) {
+ (actor ! Invocation(joinPoint, true, true) ).asInstanceOf[AnyRef]
+ }
else {
val result = actor !! Invocation(joinPoint, false, isVoid(rtti))
if (result.isDefined) result.get
@@ -357,7 +358,7 @@ private[akka] sealed class ActiveObjectAspect {
*/
private[akka] class Dispatcher(transactionalRequired: Boolean, val callbacks: Option[RestartCallbacks]) extends Actor {
private val ZERO_ITEM_CLASS_ARRAY = Array[Class[_]]()
- private val ZERO_ITEM_OBJECT_ARRAY = Array[Object[_]]()
+ private val ZERO_ITEM_OBJECT_ARRAY = Array[Object]()
private[actor] var target: Option[AnyRef] = None
private var preRestart: Option[Method] = None
diff --git a/akka-core/src/main/scala/actor/Actor.scala b/akka-core/src/main/scala/actor/Actor.scala
index f8eb5e0347..b4e10057b4 100644
--- a/akka-core/src/main/scala/actor/Actor.scala
+++ b/akka-core/src/main/scala/actor/Actor.scala
@@ -73,7 +73,7 @@ object Actor extends Logging {
val PORT = config.getInt("akka.remote.server.port", 9999)
object Sender{
- implicit val Self: Option[Actor] = None
+ object Self
}
/**
@@ -474,7 +474,7 @@ trait Actor extends TransactionManagement {
* actor.send(message)
*
*/
- def !(message: Any)(implicit sender: Option[Actor]) = {
+ def !(message: Any)(implicit sender: Option[Actor] = None) = {
//FIXME 2.8 def !(message: Any)(implicit sender: Option[Actor] = None) = {
if (_isKilled) throw new ActorKilledException("Actor [" + toString + "] has been killed, can't respond to messages")
if (_isRunning) postMessageToMailbox(message, sender)
@@ -539,8 +539,7 @@ trait Actor extends TransactionManagement {
/*
- //FIXME 2.8 def !!!(message: Any)(implicit sender: AnyRef = None): FutureResult = {
- def !!!(message: Any)(implicit sender: AnyRef): FutureResult = {
+ def !!!(message: Any)(implicit sender: Option[Actor] = None): FutureResult = {
if (_isKilled) throw new ActorKilledException("Actor [" + toString + "] has been killed, can't respond to messages")
if (_isRunning) {
val from = if (sender != null && sender.isInstanceOf[Actor]) Some(sender.asInstanceOf[Actor])
@@ -562,7 +561,7 @@ trait Actor extends TransactionManagement {
*
* Works with both '!' and '!!'.
*/
- def forward(message: Any)(implicit sender: Option[Actor]) = {
+ def forward(message: Any)(implicit sender: Option[Actor] = None) = {
if (_isKilled) throw new ActorKilledException("Actor [" + toString + "] has been killed, can't respond to messages")
if (_isRunning) {
val forwarder = sender.getOrElse(throw new IllegalStateException("Can't forward message when the forwarder/mediator is not an actor"))
@@ -1027,7 +1026,8 @@ trait Actor extends TransactionManagement {
!message.isInstanceOf[List[_]] &&
!message.isInstanceOf[scala.collection.immutable.Map[_, _]] &&
!message.isInstanceOf[scala.collection.immutable.Set[_]] &&
- !message.isInstanceOf[scala.collection.immutable.Tree[_, _]] &&
+ //Removed in Scala 2.8
+ //!message.isInstanceOf[scala.collection.immutable.Tree[_, _]] &&
!message.getClass.isAnnotationPresent(Annotations.immutable)) {
Serializer.Java.deepClone(message)
} else message
diff --git a/akka-core/src/main/scala/actor/Scheduler.scala b/akka-core/src/main/scala/actor/Scheduler.scala
index 8205db5843..7217af7574 100644
--- a/akka-core/src/main/scala/actor/Scheduler.scala
+++ b/akka-core/src/main/scala/actor/Scheduler.scala
@@ -19,8 +19,6 @@ import se.scalablesolutions.akka.config.ScalaConfig._
import se.scalablesolutions.akka.config.{AllForOneStrategy, OneForOneStrategy, FaultHandlingStrategy}
import se.scalablesolutions.akka.util.{Logging}
-import org.scala_tools.javautils.Imports._
-
case object UnSchedule
case class SchedulerException(msg: String, e: Throwable) extends RuntimeException(msg, e)
@@ -66,7 +64,8 @@ object Scheduler extends Actor {
}
override def shutdown = {
- schedulers.values.asScala.foreach(_ ! UnSchedule)
+ import scala.collection.JavaConversions._
+ schedulers.values.foreach(_ ! UnSchedule)
service.shutdown
}
diff --git a/akka-core/src/main/scala/config/ConfiguratorRepository.scala b/akka-core/src/main/scala/config/ConfiguratorRepository.scala
index 9c12bf4b32..097259164b 100644
--- a/akka-core/src/main/scala/config/ConfiguratorRepository.scala
+++ b/akka-core/src/main/scala/config/ConfiguratorRepository.scala
@@ -6,7 +6,7 @@ package se.scalablesolutions.akka.config
import scala.collection.mutable.HashSet
-import util.Logging
+import se.scalablesolutions.akka.util.Logging
object ConfiguratorRepository extends Logging {
diff --git a/akka-core/src/main/scala/remote/Cluster.scala b/akka-core/src/main/scala/remote/Cluster.scala
index c2e9069a01..cb68b271ed 100644
--- a/akka-core/src/main/scala/remote/Cluster.scala
+++ b/akka-core/src/main/scala/remote/Cluster.scala
@@ -133,7 +133,7 @@ abstract class BasicClusterActor extends ClusterActor {
case RegisterLocalNode(s) => {
log debug ("RegisterLocalNode: %s", s)
- local = Node(local.endpoints + s)
+ local = Node(s :: local.endpoints)
broadcast(Papers(local.endpoints))
}
diff --git a/akka-core/src/main/scala/remote/RemoteClient.scala b/akka-core/src/main/scala/remote/RemoteClient.scala
index f97f014f06..7d22f4289b 100644
--- a/akka-core/src/main/scala/remote/RemoteClient.scala
+++ b/akka-core/src/main/scala/remote/RemoteClient.scala
@@ -267,7 +267,7 @@ class RemoteClientPipelineFactory(name: String,
/**
* @author Jonas Bonér
*/
-@ChannelPipelineCoverage {val value = "all"}
+@ChannelPipelineCoverage(value = "all")
class RemoteClientHandler(val name: String,
val futures: ConcurrentMap[Long, CompletableFutureResult],
val supervisors: ConcurrentMap[String, Actor],
diff --git a/akka-core/src/main/scala/remote/RemoteProtocolBuilder.scala b/akka-core/src/main/scala/remote/RemoteProtocolBuilder.scala
index 287168140a..ac0a3986c6 100644
--- a/akka-core/src/main/scala/remote/RemoteProtocolBuilder.scala
+++ b/akka-core/src/main/scala/remote/RemoteProtocolBuilder.scala
@@ -4,7 +4,7 @@
package se.scalablesolutions.akka.remote
-import se.scalablesolutions.akka.serialization.Serializable.SBinary
+//import se.scalablesolutions.akka.serialization.Serializable.SBinary
import se.scalablesolutions.akka.serialization.{Serializer, Serializable, SerializationProtocol}
import se.scalablesolutions.akka.remote.protobuf.RemoteProtocol.{RemoteRequest, RemoteReply}
@@ -14,7 +14,7 @@ object RemoteProtocolBuilder {
private var SERIALIZER_JAVA: Serializer.Java = Serializer.Java
private var SERIALIZER_JAVA_JSON: Serializer.JavaJSON = Serializer.JavaJSON
private var SERIALIZER_SCALA_JSON: Serializer.ScalaJSON = Serializer.ScalaJSON
- private var SERIALIZER_SBINARY: Serializer.SBinary = Serializer.SBinary
+ //private var SERIALIZER_SBINARY: Serializer.SBinary = Serializer.SBinary
private var SERIALIZER_PROTOBUF: Serializer.Protobuf = Serializer.Protobuf
@@ -29,9 +29,9 @@ object RemoteProtocolBuilder {
def getMessage(request: RemoteRequest): Any = {
request.getProtocol match {
- case SerializationProtocol.SBINARY =>
- val renderer = Class.forName(new String(request.getMessageManifest.toByteArray)).newInstance.asInstanceOf[SBinary[_ <: AnyRef]]
- renderer.fromBytes(request.getMessage.toByteArray)
+ //case SerializationProtocol.SBINARY =>
+ // val renderer = Class.forName(new String(request.getMessageManifest.toByteArray)).newInstance.asInstanceOf[SBinary[_ <: AnyRef]]
+ // renderer.fromBytes(request.getMessage.toByteArray)
case SerializationProtocol.SCALA_JSON =>
val manifest = SERIALIZER_JAVA.in(request.getMessageManifest.toByteArray, None).asInstanceOf[String]
SERIALIZER_SCALA_JSON.in(request.getMessage.toByteArray, Some(Class.forName(manifest)))
@@ -50,9 +50,9 @@ object RemoteProtocolBuilder {
def getMessage(reply: RemoteReply): Any = {
reply.getProtocol match {
- case SerializationProtocol.SBINARY =>
- val renderer = Class.forName(new String(reply.getMessageManifest.toByteArray)).newInstance.asInstanceOf[SBinary[_ <: AnyRef]]
- renderer.fromBytes(reply.getMessage.toByteArray)
+ //case SerializationProtocol.SBINARY =>
+ // val renderer = Class.forName(new String(reply.getMessageManifest.toByteArray)).newInstance.asInstanceOf[SBinary[_ <: AnyRef]]
+ // renderer.fromBytes(reply.getMessage.toByteArray)
case SerializationProtocol.SCALA_JSON =>
val manifest = SERIALIZER_JAVA.in(reply.getMessageManifest.toByteArray, None).asInstanceOf[String]
SERIALIZER_SCALA_JSON.in(reply.getMessage.toByteArray, Some(Class.forName(manifest)))
@@ -70,12 +70,12 @@ object RemoteProtocolBuilder {
}
def setMessage(message: Any, builder: RemoteRequest.Builder) = {
- if (message.isInstanceOf[Serializable.SBinary[_]]) {
+ /*if (message.isInstanceOf[Serializable.SBinary[_]]) {
val serializable = message.asInstanceOf[Serializable.SBinary[_ <: Any]]
builder.setProtocol(SerializationProtocol.SBINARY)
builder.setMessage(ByteString.copyFrom(serializable.toBytes))
builder.setMessageManifest(ByteString.copyFrom(serializable.getClass.getName.getBytes))
- } else if (message.isInstanceOf[Message]) {
+ } else*/ if (message.isInstanceOf[Message]) {
val serializable = message.asInstanceOf[Message]
builder.setProtocol(SerializationProtocol.PROTOBUF)
builder.setMessage(ByteString.copyFrom(serializable.toByteArray))
@@ -98,12 +98,12 @@ object RemoteProtocolBuilder {
}
def setMessage(message: Any, builder: RemoteReply.Builder) = {
- if (message.isInstanceOf[Serializable.SBinary[_]]) {
+ /*if (message.isInstanceOf[Serializable.SBinary[_]]) {
val serializable = message.asInstanceOf[Serializable.SBinary[_ <: Any]]
builder.setProtocol(SerializationProtocol.SBINARY)
builder.setMessage(ByteString.copyFrom(serializable.toBytes))
builder.setMessageManifest(ByteString.copyFrom(serializable.getClass.getName.getBytes))
- } else if (message.isInstanceOf[Message]) {
+ } else*/ if (message.isInstanceOf[Message]) {
val serializable = message.asInstanceOf[Message]
builder.setProtocol(SerializationProtocol.PROTOBUF)
builder.setMessage(ByteString.copyFrom(serializable.toByteArray))
diff --git a/akka-core/src/main/scala/remote/RemoteServer.scala b/akka-core/src/main/scala/remote/RemoteServer.scala
index 5288bd7bd4..7564fe757c 100644
--- a/akka-core/src/main/scala/remote/RemoteServer.scala
+++ b/akka-core/src/main/scala/remote/RemoteServer.scala
@@ -244,7 +244,7 @@ class RemoteServerPipelineFactory(
/**
* @author Jonas Bonér
*/
-@ChannelPipelineCoverage {val value = "all"}
+@ChannelPipelineCoverage(value = "all")
class RemoteServerHandler(
val name: String,
val openChannels: ChannelGroup,
diff --git a/akka-core/src/main/scala/serialization/Binary.scala b/akka-core/src/main/scala/serialization/Binary.scala
index 01a530b215..818131cbf8 100644
--- a/akka-core/src/main/scala/serialization/Binary.scala
+++ b/akka-core/src/main/scala/serialization/Binary.scala
@@ -1,7 +1,7 @@
/**
* Copyright (C) 2009-2010 Scalable Solutions AB
*/
-
+/*
package se.scalablesolutions.akka.serialization
// --- PRIMITIVES ---
@@ -373,3 +373,4 @@ case class BinaryStringByteArrayMap(val value: Map[String, Array[Byte]]) extends
def fromBytes(bytes: Array[Byte]) = BinaryStringByteArrayMap(fromByteArray[Map[String, Array[Byte]]](bytes))
def toBytes: Array[Byte] = toByteArray(value)
}
+*/
\ No newline at end of file
diff --git a/akka-core/src/main/scala/serialization/Serializable.scala b/akka-core/src/main/scala/serialization/Serializable.scala
index b9a3cf5927..0111cdfc8c 100644
--- a/akka-core/src/main/scala/serialization/Serializable.scala
+++ b/akka-core/src/main/scala/serialization/Serializable.scala
@@ -7,12 +7,12 @@ package se.scalablesolutions.akka.serialization
import org.codehaus.jackson.map.ObjectMapper
import com.google.protobuf.Message
import reflect.Manifest
-import sbinary.DefaultProtocol
+//import sbinary.DefaultProtocol
import java.io.{StringWriter, ByteArrayOutputStream, ObjectOutputStream}
import sjson.json.{Serializer=>SJSONSerializer}
object SerializationProtocol {
- val SBINARY = 1
+ //val SBINARY = 1
val SCALA_JSON = 2
val JAVA_JSON = 3
val PROTOBUF = 4
@@ -60,10 +60,10 @@ object Serializable {
*
* @author Jonas Bonér
*/
- trait SBinary[T <: AnyRef] extends Serializable {
+ /*trait SBinary[T <: AnyRef] extends Serializable {
def fromBytes(bytes: Array[Byte]): T
def toBytes: Array[Byte]
- }
+ } */
/**
* @author Jonas Bonér
diff --git a/akka-core/src/main/scala/serialization/Serializer.scala b/akka-core/src/main/scala/serialization/Serializer.scala
index 3eb9315126..55f695a5c4 100644
--- a/akka-core/src/main/scala/serialization/Serializer.scala
+++ b/akka-core/src/main/scala/serialization/Serializer.scala
@@ -29,7 +29,7 @@ class SerializerFactory {
def getJava: Java.type = Java
def getJavaJSON: JavaJSON.type = JavaJSON
def getScalaJSON: ScalaJSON.type = ScalaJSON
- def getSBinary: SBinary.type = SBinary
+ //def getSBinary: SBinary.type = SBinary
def getProtobuf: Protobuf.type = Protobuf
}
@@ -165,7 +165,7 @@ object Serializer {
/**
* @author Jonas Bonér
*/
- object SBinary extends SBinary
+ /*object SBinary extends SBinary
class SBinary {
import sbinary.DefaultProtocol._
@@ -176,6 +176,6 @@ object Serializer {
def in[T](array : Array[Byte], clazz: Option[Class[T]])(implicit bin : Reads[T]): T = fromByteArray[T](array)
def in[T](array : Array[Byte])(implicit bin : Reads[T]): T = fromByteArray[T](array)
- }
+ } */
}
diff --git a/akka-core/src/main/scala/stm/HashTrie.scala b/akka-core/src/main/scala/stm/HashTrie.scala
index 02b7ad2145..5f79004ef0 100644
--- a/akka-core/src/main/scala/stm/HashTrie.scala
+++ b/akka-core/src/main/scala/stm/HashTrie.scala
@@ -68,7 +68,7 @@ final class HashTrie[K, +V] private (root: Node[K, V]) extends Map[K, V] with Pe
}
object HashTrie {
- def apply[K, V](pairs: (K, V)*) = pairs.foldLeft(new HashTrie[K, V]) { _ + _ }
+ def apply[K, V](pairs: (K, V)*) = pairs.foldLeft((new HashTrie[K, V]).asInstanceOf[Map[K,V]]) { _ + _ }
def unapplySeq[K, V](map: HashTrie[K, V]) = map.toSeq
}
diff --git a/akka-core/src/main/scala/stm/ResultOrFailure.scala b/akka-core/src/main/scala/stm/ResultOrFailure.scala
index 51ce6ddf68..a8b5090a68 100644
--- a/akka-core/src/main/scala/stm/ResultOrFailure.scala
+++ b/akka-core/src/main/scala/stm/ResultOrFailure.scala
@@ -4,7 +4,7 @@
package se.scalablesolutions.akka.util
-import stm.Transaction
+import se.scalablesolutions.akka.stm.Transaction
/**
* Reference that can hold either a typed value or an exception.
diff --git a/akka-core/src/main/scala/stm/TransactionalState.scala b/akka-core/src/main/scala/stm/TransactionalState.scala
index 6003a89f89..3898409e67 100644
--- a/akka-core/src/main/scala/stm/TransactionalState.scala
+++ b/akka-core/src/main/scala/stm/TransactionalState.scala
@@ -173,11 +173,19 @@ class TransactionalMap[K, V] extends Transactional with scala.collection.mutable
ref.swap(new HashTrie[K, V])
- def -=(key: K) = remove(key)
+ def -=(key: K) = {
+ remove(key)
+ this
+ }
def +=(key: K, value: V) = put(key, value)
- def remove(key: K) = ref.swap(ref.get.get - key)
+ def remove(key: K) = {
+ val map = ref.get.get
+ val oldValue = map.get(key)
+ ref.swap(ref.get.get - key)
+ oldValue
+ }
def get(key: K): Option[V] = ref.get.get.get(key)
diff --git a/akka-persistence/akka-persistence-cassandra/src/main/scala/CassandraSession.scala b/akka-persistence/akka-persistence-cassandra/src/main/scala/CassandraSession.scala
index 0cbf9fa04b..2db1f7a2ba 100644
--- a/akka-persistence/akka-persistence-cassandra/src/main/scala/CassandraSession.scala
+++ b/akka-persistence/akka-persistence-cassandra/src/main/scala/CassandraSession.scala
@@ -23,8 +23,7 @@ import org.apache.thrift.protocol._
* @author Jonas Bonér
*/
trait CassandraSession extends Closeable with Flushable {
- import scala.collection.jcl.Conversions._
- import org.scala_tools.javautils.Imports._
+ import scala.collection.JavaConversions._
import java.util.{Map => JMap, List => JList}
protected val client: Cassandra.Client
@@ -91,7 +90,7 @@ trait CassandraSession extends Closeable with Flushable {
def ++|(key: String, batch: Map[String, List[ColumnOrSuperColumn]], consistencyLevel: Int): Unit = {
val jmap = new java.util.HashMap[String, JList[ColumnOrSuperColumn]]
- for (entry <- batch; (key, value) = entry) jmap.put(key, value.asJava)
+ for (entry <- batch; (key, value) = entry) jmap.put(key, value)
client.batch_insert(keyspace, key, jmap, consistencyLevel)
}
diff --git a/akka-rest/src/main/scala/AkkaServlet.scala b/akka-rest/src/main/scala/AkkaServlet.scala
index 2aa2a6b2c5..c79f4b17d1 100644
--- a/akka-rest/src/main/scala/AkkaServlet.scala
+++ b/akka-rest/src/main/scala/AkkaServlet.scala
@@ -17,14 +17,14 @@ import com.sun.jersey.spi.container.WebApplication
* @author Jonas Bonér
*/
class AkkaServlet extends ServletContainer {
- import org.scala_tools.javautils.Imports._
+ import scala.collection.JavaConversions._
override def initiate(resourceConfig: ResourceConfig, webApplication: WebApplication) = {
//Kernel.boot // will boot if not already booted by 'main'
val configurators = ConfiguratorRepository.getConfigurators
- resourceConfig.getClasses.addAll(configurators.flatMap(_.getComponentInterfaces).asJava)
+ resourceConfig.getClasses.addAll(configurators.flatMap(_.getComponentInterfaces))
resourceConfig.getProperties.put(
"com.sun.jersey.spi.container.ResourceFilters",
Config.config.getList("akka.rest.filters").mkString(","))
diff --git a/embedded-repo/sjson/json/sjson/0.5-SNAPSHOT-2.8.Beta1/sjson-0.5-SNAPSHOT-2.8.Beta1.jar b/embedded-repo/sjson/json/sjson/0.5-SNAPSHOT-2.8.Beta1/sjson-0.5-SNAPSHOT-2.8.Beta1.jar
new file mode 100644
index 0000000000..4543815742
Binary files /dev/null and b/embedded-repo/sjson/json/sjson/0.5-SNAPSHOT-2.8.Beta1/sjson-0.5-SNAPSHOT-2.8.Beta1.jar differ
diff --git a/embedded-repo/sjson/json/sjson/0.5-SNAPSHOT-2.8.Beta1/sjson-0.5-SNAPSHOT-2.8.Beta1.pom b/embedded-repo/sjson/json/sjson/0.5-SNAPSHOT-2.8.Beta1/sjson-0.5-SNAPSHOT-2.8.Beta1.pom
new file mode 100755
index 0000000000..9c6db30374
--- /dev/null
+++ b/embedded-repo/sjson/json/sjson/0.5-SNAPSHOT-2.8.Beta1/sjson-0.5-SNAPSHOT-2.8.Beta1.pom
@@ -0,0 +1,8 @@
+
+
+ 4.0.0
+ sjson.json
+ sjson
+ 0.5-SNAPSHOT-2.8.Beta1
+ jar
+