This is an automated email from the ASF dual-hosted git repository.
fanningpj pushed a commit to branch main
in repository https://gitbox.apache.org/repos/asf/pekko-projection.git
commit 5520b6f1246a020be414c308a537f7748ce1acc6
Author: scala-steward-asf[bot]
<147768647+scala-steward-asf[bot]@users.noreply.github.com>
AuthorDate: Sun Nov 2 00:29:48 2025 +0000
Reformat with scalafmt 3.9.10
Executed command: scalafmt --non-interactive
---
.../apache/pekko/projection/cassandra/CassandraProjectionSpec.scala | 2 +-
.../apache/pekko/projection/internal/OffsetSerializationSpec.scala | 2 +-
.../org/apache/pekko/projection/internal/ActorHandlerInit.scala | 2 +-
.../apache/pekko/projection/internal/InternalProjectionState.scala | 4 ++--
.../org/apache/pekko/projection/internal/OffsetSerialization.scala | 6 +++---
.../scala/org/apache/pekko/projection/internal/OffsetStrategy.scala | 2 +-
.../org/apache/pekko/projection/internal/ProjectionSerializer.scala | 6 +++---
.../org/apache/pekko/projection/internal/ProjectionSettings.scala | 2 +-
.../org/apache/pekko/projection/scaladsl/StatefulHandler.scala | 2 +-
.../scala/org/apache/pekko/projection/jdbc/JdbcProjectionSpec.scala | 2 +-
.../org/apache/pekko/projection/jdbc/internal/JdbcSettings.scala | 2 +-
.../main/scala/org/apache/pekko/projection/kafka/KafkaOffsets.scala | 2 +-
project/Common.scala | 2 +-
.../scala/org/apache/pekko/projection/slick/SlickProjection.scala | 4 ++--
.../apache/pekko/projection/slick/internal/SlickOffsetStore.scala | 4 ++--
.../org/apache/pekko/projection/slick/SlickProjectionSpec.scala | 4 ++--
16 files changed, 24 insertions(+), 24 deletions(-)
diff --git
a/cassandra-test/src/test/scala/org/apache/pekko/projection/cassandra/CassandraProjectionSpec.scala
b/cassandra-test/src/test/scala/org/apache/pekko/projection/cassandra/CassandraProjectionSpec.scala
index 09b8ea0..e7d36d6 100644
---
a/cassandra-test/src/test/scala/org/apache/pekko/projection/cassandra/CassandraProjectionSpec.scala
+++
b/cassandra-test/src/test/scala/org/apache/pekko/projection/cassandra/CassandraProjectionSpec.scala
@@ -911,7 +911,7 @@ class CassandraProjectionSpec
}
def handler: LifecycleHandler = _handler match {
case Some(h) => h
- case None =>
+ case None =>
handlerProbe.awaitAssert {
_handler.get
}
diff --git
a/core-test/src/test/scala/org/apache/pekko/projection/internal/OffsetSerializationSpec.scala
b/core-test/src/test/scala/org/apache/pekko/projection/internal/OffsetSerializationSpec.scala
index 7b46161..3368e8e 100644
---
a/core-test/src/test/scala/org/apache/pekko/projection/internal/OffsetSerializationSpec.scala
+++
b/core-test/src/test/scala/org/apache/pekko/projection/internal/OffsetSerializationSpec.scala
@@ -39,7 +39,7 @@ object OffsetSerializationSpec {
def toBinary(o: AnyRef): Array[Byte] = o match {
case OtherOffset(s) => s.getBytes(StandardCharsets.UTF_8)
- case _ =>
+ case _ =>
throw new IllegalArgumentException(s"Can't serialize object of type
${o.getClass} in [${getClass.getName}]")
}
diff --git
a/core/src/main/scala/org/apache/pekko/projection/internal/ActorHandlerInit.scala
b/core/src/main/scala/org/apache/pekko/projection/internal/ActorHandlerInit.scala
index dfe002d..8bb269f 100644
---
a/core/src/main/scala/org/apache/pekko/projection/internal/ActorHandlerInit.scala
+++
b/core/src/main/scala/org/apache/pekko/projection/internal/ActorHandlerInit.scala
@@ -35,7 +35,7 @@ import pekko.annotation.InternalApi
final private[projection] def getActor(): ActorRef[T] = {
actor match {
case Some(ref) => ref
- case None =>
+ case None =>
throw new IllegalStateException(
"Actor not started, please report issue at " +
"https://github.com/apache/pekko-projection/issues")
diff --git
a/core/src/main/scala/org/apache/pekko/projection/internal/InternalProjectionState.scala
b/core/src/main/scala/org/apache/pekko/projection/internal/InternalProjectionState.scala
index cebc2d2..bedc44d 100644
---
a/core/src/main/scala/org/apache/pekko/projection/internal/InternalProjectionState.scala
+++
b/core/src/main/scala/org/apache/pekko/projection/internal/InternalProjectionState.scala
@@ -114,7 +114,7 @@ private[projection] abstract class
InternalProjectionState[Offset, Envelope](
remaining: List[(String, immutable.Seq[ProjectionContextImpl[Offset,
Envelope]])],
n: Int): Future[Done] = {
remaining match {
- case Nil => Future.successful(Done)
+ case Nil => Future.successful(Done)
case (key, batch) :: tail =>
op(key, batch).flatMap { _ =>
if (n % logProgressEvery == 0)
@@ -428,7 +428,7 @@ private[projection] abstract class
InternalProjectionState[Offset, Envelope](
sourceProvider match {
case vsp: VerifiableSourceProvider[Offset, Envelope] =>
vsp.verifyOffset(context.offset) match {
- case VerificationSuccess => true
+ case VerificationSuccess => true
case VerificationFailure(reason) =>
logger.warning(
"Source provider instructed projection to skip offset [{}]
with reason: {}",
diff --git
a/core/src/main/scala/org/apache/pekko/projection/internal/OffsetSerialization.scala
b/core/src/main/scala/org/apache/pekko/projection/internal/OffsetSerialization.scala
index ee37df8..05c943b 100644
---
a/core/src/main/scala/org/apache/pekko/projection/internal/OffsetSerialization.scala
+++
b/core/src/main/scala/org/apache/pekko/projection/internal/OffsetSerialization.scala
@@ -58,7 +58,7 @@ import pekko.serialization.Serializers
def fromStorageRepresentation[Offset, Inner](rep: StorageRepresentation):
Offset = {
val offset: Offset = rep match {
case SingleOffset(_, manifest, offsetStr, _) =>
fromStorageRepresentation[Offset](offsetStr, manifest)
- case MultipleOffsets(reps) =>
+ case MultipleOffsets(reps) =>
val offsets: Map[String, Inner] = reps.map {
case SingleOffset(id, manifest, offsetStr, _) =>
id.key -> fromStorageRepresentation[Inner](offsetStr, manifest)
@@ -79,7 +79,7 @@ import pekko.serialization.Serializers
case IntManifest => offsetStr.toInt
case SequenceManifest => query.Offset.sequence(offsetStr.toLong)
case TimeBasedUUIDManifest =>
query.Offset.timeBasedUUID(UUID.fromString(offsetStr))
- case _ =>
+ case _ =>
val parts = manifest.split(':')
val serializerId = parts(0).toInt
val serializerManifest = parts(1)
@@ -101,7 +101,7 @@ import pekko.serialization.Serializers
case i: Int => SingleOffset(id, IntManifest,
i.toString, mergeable)
case seq: query.Sequence => SingleOffset(id, SequenceManifest,
seq.value.toString, mergeable)
case tbu: query.TimeBasedUUID => SingleOffset(id, TimeBasedUUIDManifest,
tbu.value.toString, mergeable)
- case mrg: MergeableOffset[_] =>
+ case mrg: MergeableOffset[_] =>
val list = mrg.entries.map {
case (key, innerOffset) =>
toStorageRepresentation(ProjectionId(id.name, key), innerOffset,
mergeable = true)
diff --git
a/core/src/main/scala/org/apache/pekko/projection/internal/OffsetStrategy.scala
b/core/src/main/scala/org/apache/pekko/projection/internal/OffsetStrategy.scala
index 082a72a..9c8ff4c 100644
---
a/core/src/main/scala/org/apache/pekko/projection/internal/OffsetStrategy.scala
+++
b/core/src/main/scala/org/apache/pekko/projection/internal/OffsetStrategy.scala
@@ -83,7 +83,7 @@ private[projection] sealed abstract class
FunctionHandlerStrategy[Envelope](hand
def handler(): Handler[Envelope] = {
_handler match {
case Some(h: Handler[Any] @unchecked) if !_recreateHandlerOnNextAccess
=> h
- case _ =>
+ case _ =>
createHandler()
_recreateHandlerOnNextAccess = false
_handler.get
diff --git
a/core/src/main/scala/org/apache/pekko/projection/internal/ProjectionSerializer.scala
b/core/src/main/scala/org/apache/pekko/projection/internal/ProjectionSerializer.scala
index 0f3310c..e848480 100644
---
a/core/src/main/scala/org/apache/pekko/projection/internal/ProjectionSerializer.scala
+++
b/core/src/main/scala/org/apache/pekko/projection/internal/ProjectionSerializer.scala
@@ -56,7 +56,7 @@ import pekko.serialization.SerializerWithStringManifest
case _: SetOffset[_] => SetOffsetManifest
case _: IsPaused => IsPausedManifest
case _: SetPaused => SetPausedManifest
- case _ =>
+ case _ =>
throw new IllegalArgumentException(s"Can't serialize object of type
${o.getClass} in [${getClass.getName}]")
}
@@ -66,7 +66,7 @@ import pekko.serialization.SerializerWithStringManifest
case m: SetOffset[_] => setOffsetToBinary(m)
case m: IsPaused => isPausedToBinary(m)
case m: SetPaused => setPausedToBinary(m)
- case _ =>
+ case _ =>
throw new IllegalArgumentException(s"Cannot serialize object of type
[${o.getClass.getName}]")
}
@@ -133,7 +133,7 @@ import pekko.serialization.SerializerWithStringManifest
case SetOffsetManifest => setOffsetFromBinary(bytes)
case IsPausedManifest => isPausedFromBinary(bytes)
case SetPausedManifest => setPausedFromBinary(bytes)
- case _ =>
+ case _ =>
throw new NotSerializableException(
s"Unimplemented deserialization of message with manifest [$manifest]
in [${getClass.getName}]")
}
diff --git
a/core/src/main/scala/org/apache/pekko/projection/internal/ProjectionSettings.scala
b/core/src/main/scala/org/apache/pekko/projection/internal/ProjectionSettings.scala
index c94c2c8..4efb063 100644
---
a/core/src/main/scala/org/apache/pekko/projection/internal/ProjectionSettings.scala
+++
b/core/src/main/scala/org/apache/pekko/projection/internal/ProjectionSettings.scala
@@ -84,7 +84,7 @@ private object RecoveryStrategyConfig {
case "skip" => HandlerRecoveryStrategy.skip
case "retry-and-fail" => HandlerRecoveryStrategy.retryAndFail(retries,
retryDelay)
case "retry-and-skip" => HandlerRecoveryStrategy.retryAndSkip(retries,
retryDelay)
- case s =>
+ case s =>
throw new IllegalArgumentException(
s"Strategy type [$s] is not supported. Supported options are [fail,
skip, retry-and-fail, retry-and-skip]")
}
diff --git
a/core/src/main/scala/org/apache/pekko/projection/scaladsl/StatefulHandler.scala
b/core/src/main/scala/org/apache/pekko/projection/scaladsl/StatefulHandler.scala
index 4403a6f..66f297e 100644
---
a/core/src/main/scala/org/apache/pekko/projection/scaladsl/StatefulHandler.scala
+++
b/core/src/main/scala/org/apache/pekko/projection/scaladsl/StatefulHandler.scala
@@ -60,7 +60,7 @@ import pekko.annotation.ApiMayChange
val newState = state.value match {
case Some(Success(_)) => state
case Some(Failure(_)) => initialState()
- case None =>
+ case None =>
throw new IllegalStateException(
"Process called before previous Future completed. " +
"Did you share the same handler instance between several Projection
instances? " +
diff --git
a/jdbc-int-test/src/test/scala/org/apache/pekko/projection/jdbc/JdbcProjectionSpec.scala
b/jdbc-int-test/src/test/scala/org/apache/pekko/projection/jdbc/JdbcProjectionSpec.scala
index 9105816..7226b53 100644
---
a/jdbc-int-test/src/test/scala/org/apache/pekko/projection/jdbc/JdbcProjectionSpec.scala
+++
b/jdbc-int-test/src/test/scala/org/apache/pekko/projection/jdbc/JdbcProjectionSpec.scala
@@ -1183,7 +1183,7 @@ class JdbcProjectionSpec
}
def handler: LifecycleHandler = _handler match {
case Some(h) => h
- case None =>
+ case None =>
handlerProbe.awaitAssert {
_handler.get
}
diff --git
a/jdbc/src/main/scala/org/apache/pekko/projection/jdbc/internal/JdbcSettings.scala
b/jdbc/src/main/scala/org/apache/pekko/projection/jdbc/internal/JdbcSettings.scala
index 31db11f..40bd3fa 100644
---
a/jdbc/src/main/scala/org/apache/pekko/projection/jdbc/internal/JdbcSettings.scala
+++
b/jdbc/src/main/scala/org/apache/pekko/projection/jdbc/internal/JdbcSettings.scala
@@ -51,7 +51,7 @@ case class JdbcSettings(config: Config, executionContext:
ExecutionContext) {
case "mysql-dialect" => MySQLDialect(schema, table, managementTable)
case "mssql-dialect" => MSSQLServerDialect(schema, table,
managementTable)
case "oracle-dialect" => OracleDialect(schema, table, managementTable)
- case unknown =>
+ case unknown =>
throw new IllegalArgumentException(
s"Unknown dialect type: [$unknown]. Check settings
'pekko.projection.jdbc.dialect'")
}
diff --git
a/kafka/src/main/scala/org/apache/pekko/projection/kafka/KafkaOffsets.scala
b/kafka/src/main/scala/org/apache/pekko/projection/kafka/KafkaOffsets.scala
index 8072b38..5e3acf1 100644
--- a/kafka/src/main/scala/org/apache/pekko/projection/kafka/KafkaOffsets.scala
+++ b/kafka/src/main/scala/org/apache/pekko/projection/kafka/KafkaOffsets.scala
@@ -37,7 +37,7 @@ object KafkaOffsets {
def keyToPartition(key: String): TopicPartition = key match {
case RegexTp(topic, partition) => new TopicPartition(topic,
partition.toInt)
- case _ =>
+ case _ =>
throw new IllegalArgumentException(s"Row entry name ($key) must match
pattern: ${RegexTp.pattern.toString}")
}
diff --git a/project/Common.scala b/project/Common.scala
index 0d0962b..a4ddf04 100644
--- a/project/Common.scala
+++ b/project/Common.scala
@@ -73,7 +73,7 @@ object Common extends AutoPlugin {
mimaPreviousArtifacts := {
moduleName.value match {
case name if name.endsWith("-tests") => Set.empty
- case _ =>
+ case _ =>
Set(
organization.value %% moduleName.value % mimaCompareVersion)
}
diff --git
a/slick/src/main/scala/org/apache/pekko/projection/slick/SlickProjection.scala
b/slick/src/main/scala/org/apache/pekko/projection/slick/SlickProjection.scala
index 1ae50d9..fb3a972 100644
---
a/slick/src/main/scala/org/apache/pekko/projection/slick/SlickProjection.scala
+++
b/slick/src/main/scala/org/apache/pekko/projection/slick/SlickProjection.scala
@@ -92,7 +92,7 @@ object SlickProjection {
case vsp: VerifiableSourceProvider[Offset, Envelope] =>
processedDBIO.flatMap { action =>
vsp.verifyOffset(offset) match {
- case VerificationSuccess =>
slick.dbio.DBIO.successful(action)
+ case VerificationSuccess =>
slick.dbio.DBIO.successful(action)
case VerificationFailure(reason) =>
logger.warning(
"The offset failed source provider verification after
the envelope was processed. " +
@@ -251,7 +251,7 @@ object SlickProjection {
case vsp: VerifiableSourceProvider[Offset, Envelope] =>
processedDBIO.flatMap { action =>
vsp.verifyOffset(lastOffset) match {
- case VerificationSuccess =>
slick.dbio.DBIO.successful(action)
+ case VerificationSuccess =>
slick.dbio.DBIO.successful(action)
case VerificationFailure(reason) =>
logger.warning(
"The offset failed source provider verification after
the envelope was processed. " +
diff --git
a/slick/src/main/scala/org/apache/pekko/projection/slick/internal/SlickOffsetStore.scala
b/slick/src/main/scala/org/apache/pekko/projection/slick/internal/SlickOffsetStore.scala
index 269f746..5159081 100644
---
a/slick/src/main/scala/org/apache/pekko/projection/slick/internal/SlickOffsetStore.scala
+++
b/slick/src/main/scala/org/apache/pekko/projection/slick/internal/SlickOffsetStore.scala
@@ -92,7 +92,7 @@ import slick.jdbc.JdbcProfile
val results = databaseConfig.db.run(action)
results.map {
- case Nil => None
+ case Nil => None
case reps if reps.forall(_.mergeable) =>
Some(fromStorageRepresentation[MergeableOffset[_],
Offset](MultipleOffsets(reps.toList)).asInstanceOf[Offset])
case reps =>
@@ -107,7 +107,7 @@ import slick.jdbc.JdbcProfile
def saveOffset[Offset](projectionId: ProjectionId, offset: Offset):
slick.dbio.DBIO[_] = {
val millisSinceEpoch = clock.instant().toEpochMilli
toStorageRepresentation(projectionId, offset) match {
- case offset: SingleOffset => newRow(offset, millisSinceEpoch)
+ case offset: SingleOffset => newRow(offset, millisSinceEpoch)
case MultipleOffsets(reps) =>
val actions = reps.map(rep => newRow(rep, millisSinceEpoch))
DBIO.sequence(actions)
diff --git
a/slick/src/test/scala/org/apache/pekko/projection/slick/SlickProjectionSpec.scala
b/slick/src/test/scala/org/apache/pekko/projection/slick/SlickProjectionSpec.scala
index 2c84bf7..d90174d 100644
---
a/slick/src/test/scala/org/apache/pekko/projection/slick/SlickProjectionSpec.scala
+++
b/slick/src/test/scala/org/apache/pekko/projection/slick/SlickProjectionSpec.scala
@@ -598,7 +598,7 @@ class SlickProjectionSpec
case e: Envelope =>
val o = e.offset
acc.get(e.offset) match {
- case Some(state) if state.envReceived => fail(s"Envelope
was already received for offset $o")
+ case Some(state) if state.envReceived =>
fail(s"Envelope was already received for offset $o")
case Some(state) if state.verifyCount >= 2 =>
fail(s"Envelope has already been verified at least twice
for offset $o")
case None => fail(s"Envelope has not been verified
yet for offset $o")
@@ -1355,7 +1355,7 @@ class SlickProjectionSpec
}
def handler: LifecycleHandler = _handler match {
case Some(h) => h
- case None =>
+ case None =>
handlerProbe.awaitAssert {
_handler.get
}
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]