sven-lange-last commented on a change in pull request #2878: Streamingly read 
user-logs.
URL: 
https://github.com/apache/incubator-openwhisk/pull/2878#discussion_r152029555
 
 

 ##########
 File path: 
core/invoker/src/main/scala/whisk/core/containerpool/docker/DockerContainer.scala
 ##########
 @@ -224,31 +230,73 @@ class DockerContainer(protected val id: ContainerId,
    *
    * @return a vector of Strings with log lines in our own JSON format
    */
-  def logs(limit: ByteSize, waitForSentinel: Boolean)(implicit transid: 
TransactionId): Future[Vector[String]] = {
-
-    def readLogs(retries: Int): Future[Vector[String]] = {
-      docker
-        .rawContainerLogs(id, logFileOffset)
-        .flatMap { rawLogBytes =>
-          val rawLog =
-            new String(rawLogBytes.array, rawLogBytes.arrayOffset, 
rawLogBytes.position, StandardCharsets.UTF_8)
-          val (isComplete, isTruncated, formattedLogs) = 
processJsonDriverLogContents(rawLog, waitForSentinel, limit)
-
-          if (retries > 0 && !isComplete && !isTruncated) {
-            logging.info(this, s"log cursor advanced but missing sentinel, 
trying $retries more times")
-            akka.pattern.after(filePollInterval, 
as.scheduler)(readLogs(retries - 1))
-          } else {
-            logFileOffset += rawLogBytes.position - rawLogBytes.arrayOffset
-            Future.successful(formattedLogs)
-          }
-        }
-        .andThen {
-          case Failure(e) =>
-            logging.error(this, s"Failed to obtain logs of ${id.asString}: 
${e.getClass} - ${e.getMessage}")
-        }
+  def logs(limit: ByteSize, waitForSentinel: Boolean)(implicit transid: 
TransactionId): Source[ByteString, Any] = {
+    val source = docker
+      .rawContainerLogs(id, logFileOffset.get(), if (waitForSentinel) 
Some(filePollInterval) else None)
+      .via(Framing.delimiter(delimiter, Int.MaxValue))
+      .limitWeighted(limit.toBytes) { obj =>
+        val size = obj.size + 1
+        logFileOffset.addAndGet(size)
+        size
+      }
+      .recover {
+        case _: StreamLimitReachedException =>
+          ByteString(LogLine(Instant.now.toString, "stderr", 
Messages.truncateLogs(limit)).toJson.compactPrint)
+      }
 
 Review comment:
   * What happens if we hit the `StreamLimitReachedException`? Will the stream 
continue to provide `ByteString` instances from the `Framing.delimiter` stage 
or will processing stop?
   * The overarching question is whether additional log lines will be fully 
processed and lead to additional `Log truncated` messages once the log limit is 
exceeded.
   * If processing stops once the log limit is exceeded, the `logFileOffset` is 
no more advanced behind the sentinels. If the container is re-used as warm 
container for a subsequent activation, activation log processing will start 
from a wrong offset.

----------------------------------------------------------------
This is an automated message from the Apache Git Service.
To respond to the message, please log on GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
[email protected]


With regards,
Apache Git Services

Reply via email to