Lee-W commented on code in PR #49470:
URL: https://github.com/apache/airflow/pull/49470#discussion_r2135284510


##########
airflow-core/src/airflow/utils/log/log_stream_counter.py:
##########
@@ -0,0 +1,151 @@
+#
+# Licensed to the Apache Software Foundation (ASF) under one
+# or more contributor license agreements.  See the NOTICE file
+# distributed with this work for additional information
+# regarding copyright ownership.  The ASF licenses this file
+# to you under the Apache License, Version 2.0 (the
+# "License"); you may not use this file except in compliance
+# with the License.  You may obtain a copy of the License at
+#
+#   http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing,
+# software distributed under the License is distributed on an
+# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
+# KIND, either express or implied.  See the License for the
+# specific language governing permissions and limitations
+# under the License.
+
+from __future__ import annotations
+
+import os
+import tempfile
+from typing import TYPE_CHECKING
+
+if TYPE_CHECKING:
+    from airflow.typing_compat import Self
+    from airflow.utils.log.file_task_handler import (
+        LogHandlerOutputStream,
+        StructuredLogMessage,
+        StructuredLogStream,
+    )
+
+
+class LogStreamCounter:
+    """
+    Memory-efficient log stream counter that tracks the total number of lines 
while preserving the original stream.
+
+    This class captures logs from a stream and stores them in a buffer, 
flushing them to disk when the buffer
+    exceeds a specified threshold. This approach optimizes memory usage while 
handling large log streams.
+
+    Usage:
+
+    .. code-block:: python
+
+        with LogStreamCounter(stream, threshold) as log_counter:
+            # Get total number of lines captured
+            total_lines = log_counter.get_total_lines()
+
+            # Retrieve the original stream of logs
+            for log in log_counter.get_stream():
+                print(log)
+    """
+
+    def __init__(
+        self,
+        stream: LogHandlerOutputStream,
+        threshold: int,
+    ):
+        """
+        Initialize the LogStreamCounter.
+
+        Args:
+            stream: The input log stream to capture and count.
+            threshold: Maximum number of lines to keep in memory before 
flushing to disk.
+        """
+        self._stream = stream
+        self._threshold = threshold
+        self._buffer: list[StructuredLogMessage] = []
+        self._disk_lines: int = 0
+        self._tmpfile = None
+
+    def _flush_buffer_to_disk(self):

Review Comment:
   ```suggestion
       def _flush_buffer_to_disk(self) -> None:
   ```



##########
airflow-core/src/airflow/utils/log/log_stream_counter.py:
##########
@@ -0,0 +1,151 @@
+#
+# Licensed to the Apache Software Foundation (ASF) under one
+# or more contributor license agreements.  See the NOTICE file
+# distributed with this work for additional information
+# regarding copyright ownership.  The ASF licenses this file
+# to you under the Apache License, Version 2.0 (the
+# "License"); you may not use this file except in compliance
+# with the License.  You may obtain a copy of the License at
+#
+#   http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing,
+# software distributed under the License is distributed on an
+# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
+# KIND, either express or implied.  See the License for the
+# specific language governing permissions and limitations
+# under the License.
+
+from __future__ import annotations
+
+import os
+import tempfile
+from typing import TYPE_CHECKING
+
+if TYPE_CHECKING:
+    from airflow.typing_compat import Self
+    from airflow.utils.log.file_task_handler import (
+        LogHandlerOutputStream,
+        StructuredLogMessage,
+        StructuredLogStream,
+    )
+
+
+class LogStreamCounter:
+    """
+    Memory-efficient log stream counter that tracks the total number of lines 
while preserving the original stream.
+
+    This class captures logs from a stream and stores them in a buffer, 
flushing them to disk when the buffer
+    exceeds a specified threshold. This approach optimizes memory usage while 
handling large log streams.
+
+    Usage:
+
+    .. code-block:: python
+
+        with LogStreamCounter(stream, threshold) as log_counter:
+            # Get total number of lines captured
+            total_lines = log_counter.get_total_lines()
+
+            # Retrieve the original stream of logs
+            for log in log_counter.get_stream():
+                print(log)
+    """
+
+    def __init__(
+        self,
+        stream: LogHandlerOutputStream,
+        threshold: int,
+    ):
+        """
+        Initialize the LogStreamCounter.
+
+        Args:
+            stream: The input log stream to capture and count.
+            threshold: Maximum number of lines to keep in memory before 
flushing to disk.
+        """
+        self._stream = stream
+        self._threshold = threshold
+        self._buffer: list[StructuredLogMessage] = []
+        self._disk_lines: int = 0
+        self._tmpfile = None
+
+    def _flush_buffer_to_disk(self):
+        """Flush the buffer contents to a temporary file on disk."""
+        if self._tmpfile is None:
+            self._tmpfile = tempfile.NamedTemporaryFile(delete=False, 
mode="w+", encoding="utf-8")
+
+        self._disk_lines += len(self._buffer)
+        self._tmpfile.writelines(f"{log.model_dump_json()}\n" for log in 
self._buffer)
+        self._tmpfile.flush()
+        self._buffer.clear()
+
+    def _capture(self):

Review Comment:
   ```suggestion
       def _capture(self) -> None:
   ```



##########
airflow-core/src/airflow/utils/log/log_stream_counter.py:
##########
@@ -0,0 +1,151 @@
+#
+# Licensed to the Apache Software Foundation (ASF) under one
+# or more contributor license agreements.  See the NOTICE file
+# distributed with this work for additional information
+# regarding copyright ownership.  The ASF licenses this file
+# to you under the Apache License, Version 2.0 (the
+# "License"); you may not use this file except in compliance
+# with the License.  You may obtain a copy of the License at
+#
+#   http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing,
+# software distributed under the License is distributed on an
+# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
+# KIND, either express or implied.  See the License for the
+# specific language governing permissions and limitations
+# under the License.
+
+from __future__ import annotations
+
+import os
+import tempfile
+from typing import TYPE_CHECKING
+
+if TYPE_CHECKING:
+    from airflow.typing_compat import Self
+    from airflow.utils.log.file_task_handler import (
+        LogHandlerOutputStream,
+        StructuredLogMessage,
+        StructuredLogStream,
+    )
+
+
+class LogStreamCounter:
+    """
+    Memory-efficient log stream counter that tracks the total number of lines 
while preserving the original stream.
+
+    This class captures logs from a stream and stores them in a buffer, 
flushing them to disk when the buffer
+    exceeds a specified threshold. This approach optimizes memory usage while 
handling large log streams.
+
+    Usage:
+
+    .. code-block:: python
+
+        with LogStreamCounter(stream, threshold) as log_counter:
+            # Get total number of lines captured
+            total_lines = log_counter.get_total_lines()
+
+            # Retrieve the original stream of logs
+            for log in log_counter.get_stream():
+                print(log)
+    """
+
+    def __init__(
+        self,
+        stream: LogHandlerOutputStream,
+        threshold: int,
+    ):
+        """
+        Initialize the LogStreamCounter.
+
+        Args:
+            stream: The input log stream to capture and count.
+            threshold: Maximum number of lines to keep in memory before 
flushing to disk.
+        """
+        self._stream = stream
+        self._threshold = threshold
+        self._buffer: list[StructuredLogMessage] = []
+        self._disk_lines: int = 0
+        self._tmpfile = None
+
+    def _flush_buffer_to_disk(self):
+        """Flush the buffer contents to a temporary file on disk."""
+        if self._tmpfile is None:
+            self._tmpfile = tempfile.NamedTemporaryFile(delete=False, 
mode="w+", encoding="utf-8")
+
+        self._disk_lines += len(self._buffer)
+        self._tmpfile.writelines(f"{log.model_dump_json()}\n" for log in 
self._buffer)
+        self._tmpfile.flush()
+        self._buffer.clear()
+
+    def _capture(self):
+        """Capture logs from the stream into the buffer, flushing to disk when 
threshold is reached."""
+        for log in self._stream:
+            self._buffer.append(log)
+
+            if len(self._buffer) >= self._threshold:
+                self._flush_buffer_to_disk()
+
+    def _cleanup(self):
+        """Clean up the temporary file if it exists."""
+        self._buffer.clear()
+        if self._tmpfile:
+            self._tmpfile.close()
+            os.remove(self._tmpfile.name)
+            self._tmpfile = None
+
+    def get_total_lines(self) -> int:
+        """
+        Return the total number of lines captured from the stream.
+
+        Returns:
+            The sum of lines stored in the buffer and lines written to disk.
+        """
+        return self._disk_lines + len(self._buffer)
+
+    def get_stream(self) -> StructuredLogStream:
+        """
+        Return the original stream of logs and clean up resources.
+
+        Important: This method automatically cleans up resources after all 
logs have been yielded.
+        Make sure to fully consume the returned generator to ensure proper 
cleanup.
+
+        Returns:
+            A stream of the captured log messages.
+        """
+        try:
+            if not self._tmpfile:
+                # if no temporary file was created, return from the buffer
+                yield from self._buffer
+            else:
+                # avoid circular import
+                from airflow.utils.log.file_task_handler import 
StructuredLogMessage
+
+                with open(self._tmpfile.name, encoding="utf-8") as f:
+                    for line in f:
+                        yield 
StructuredLogMessage.model_validate_json(line.strip())

Review Comment:
   ```suggestion
                       yield from 
(StructuredLogMessage.model_validate_json(line.strip()) for line in f)
   ```



##########
airflow-core/src/airflow/utils/log/log_stream_counter.py:
##########
@@ -0,0 +1,151 @@
+#
+# Licensed to the Apache Software Foundation (ASF) under one
+# or more contributor license agreements.  See the NOTICE file
+# distributed with this work for additional information
+# regarding copyright ownership.  The ASF licenses this file
+# to you under the Apache License, Version 2.0 (the
+# "License"); you may not use this file except in compliance
+# with the License.  You may obtain a copy of the License at
+#
+#   http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing,
+# software distributed under the License is distributed on an
+# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
+# KIND, either express or implied.  See the License for the
+# specific language governing permissions and limitations
+# under the License.
+
+from __future__ import annotations
+
+import os
+import tempfile
+from typing import TYPE_CHECKING
+
+if TYPE_CHECKING:
+    from airflow.typing_compat import Self
+    from airflow.utils.log.file_task_handler import (
+        LogHandlerOutputStream,
+        StructuredLogMessage,
+        StructuredLogStream,
+    )
+
+
+class LogStreamCounter:
+    """
+    Memory-efficient log stream counter that tracks the total number of lines 
while preserving the original stream.
+
+    This class captures logs from a stream and stores them in a buffer, 
flushing them to disk when the buffer
+    exceeds a specified threshold. This approach optimizes memory usage while 
handling large log streams.
+
+    Usage:
+
+    .. code-block:: python
+
+        with LogStreamCounter(stream, threshold) as log_counter:
+            # Get total number of lines captured
+            total_lines = log_counter.get_total_lines()
+
+            # Retrieve the original stream of logs
+            for log in log_counter.get_stream():
+                print(log)
+    """
+
+    def __init__(
+        self,
+        stream: LogHandlerOutputStream,
+        threshold: int,
+    ):
+        """
+        Initialize the LogStreamCounter.
+
+        Args:
+            stream: The input log stream to capture and count.
+            threshold: Maximum number of lines to keep in memory before 
flushing to disk.
+        """
+        self._stream = stream
+        self._threshold = threshold
+        self._buffer: list[StructuredLogMessage] = []
+        self._disk_lines: int = 0
+        self._tmpfile = None
+
+    def _flush_buffer_to_disk(self):
+        """Flush the buffer contents to a temporary file on disk."""
+        if self._tmpfile is None:
+            self._tmpfile = tempfile.NamedTemporaryFile(delete=False, 
mode="w+", encoding="utf-8")
+
+        self._disk_lines += len(self._buffer)
+        self._tmpfile.writelines(f"{log.model_dump_json()}\n" for log in 
self._buffer)
+        self._tmpfile.flush()
+        self._buffer.clear()
+
+    def _capture(self):
+        """Capture logs from the stream into the buffer, flushing to disk when 
threshold is reached."""
+        for log in self._stream:
+            self._buffer.append(log)
+
+            if len(self._buffer) >= self._threshold:
+                self._flush_buffer_to_disk()

Review Comment:
   ```suggestion
           while True:
               self._buffer = list(islice(self._stream, self._threshold))
   
               if self._buffer:
                   self._flush_buffer_to_disk()
               else:
                   break
   ```
   
   not sure, but might be a little better 🤔 



##########
airflow-core/src/airflow/utils/log/log_stream_counter.py:
##########
@@ -0,0 +1,151 @@
+#
+# Licensed to the Apache Software Foundation (ASF) under one
+# or more contributor license agreements.  See the NOTICE file
+# distributed with this work for additional information
+# regarding copyright ownership.  The ASF licenses this file
+# to you under the Apache License, Version 2.0 (the
+# "License"); you may not use this file except in compliance
+# with the License.  You may obtain a copy of the License at
+#
+#   http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing,
+# software distributed under the License is distributed on an
+# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
+# KIND, either express or implied.  See the License for the
+# specific language governing permissions and limitations
+# under the License.
+
+from __future__ import annotations
+
+import os
+import tempfile
+from typing import TYPE_CHECKING
+
+if TYPE_CHECKING:
+    from airflow.typing_compat import Self
+    from airflow.utils.log.file_task_handler import (
+        LogHandlerOutputStream,
+        StructuredLogMessage,
+        StructuredLogStream,
+    )
+
+
+class LogStreamCounter:

Review Comment:
   I don't quite understand this naming. This class seems to do more than just 
counting. Is this convention learned from somewhere else?



##########
airflow-core/src/airflow/utils/log/log_stream_counter.py:
##########
@@ -0,0 +1,151 @@
+#
+# Licensed to the Apache Software Foundation (ASF) under one
+# or more contributor license agreements.  See the NOTICE file
+# distributed with this work for additional information
+# regarding copyright ownership.  The ASF licenses this file
+# to you under the Apache License, Version 2.0 (the
+# "License"); you may not use this file except in compliance
+# with the License.  You may obtain a copy of the License at
+#
+#   http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing,
+# software distributed under the License is distributed on an
+# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
+# KIND, either express or implied.  See the License for the
+# specific language governing permissions and limitations
+# under the License.
+
+from __future__ import annotations
+
+import os
+import tempfile
+from typing import TYPE_CHECKING
+
+if TYPE_CHECKING:
+    from airflow.typing_compat import Self
+    from airflow.utils.log.file_task_handler import (
+        LogHandlerOutputStream,
+        StructuredLogMessage,
+        StructuredLogStream,
+    )
+
+
+class LogStreamCounter:
+    """
+    Memory-efficient log stream counter that tracks the total number of lines 
while preserving the original stream.
+
+    This class captures logs from a stream and stores them in a buffer, 
flushing them to disk when the buffer
+    exceeds a specified threshold. This approach optimizes memory usage while 
handling large log streams.
+
+    Usage:
+
+    .. code-block:: python
+
+        with LogStreamCounter(stream, threshold) as log_counter:
+            # Get total number of lines captured
+            total_lines = log_counter.get_total_lines()
+
+            # Retrieve the original stream of logs
+            for log in log_counter.get_stream():
+                print(log)
+    """
+
+    def __init__(
+        self,
+        stream: LogHandlerOutputStream,
+        threshold: int,
+    ):
+        """
+        Initialize the LogStreamCounter.
+
+        Args:
+            stream: The input log stream to capture and count.
+            threshold: Maximum number of lines to keep in memory before 
flushing to disk.
+        """
+        self._stream = stream
+        self._threshold = threshold
+        self._buffer: list[StructuredLogMessage] = []
+        self._disk_lines: int = 0
+        self._tmpfile = None
+
+    def _flush_buffer_to_disk(self):
+        """Flush the buffer contents to a temporary file on disk."""
+        if self._tmpfile is None:
+            self._tmpfile = tempfile.NamedTemporaryFile(delete=False, 
mode="w+", encoding="utf-8")
+
+        self._disk_lines += len(self._buffer)
+        self._tmpfile.writelines(f"{log.model_dump_json()}\n" for log in 
self._buffer)
+        self._tmpfile.flush()
+        self._buffer.clear()
+
+    def _capture(self):
+        """Capture logs from the stream into the buffer, flushing to disk when 
threshold is reached."""
+        for log in self._stream:
+            self._buffer.append(log)
+
+            if len(self._buffer) >= self._threshold:
+                self._flush_buffer_to_disk()
+
+    def _cleanup(self):

Review Comment:
   ```suggestion
       def _cleanup(self) -> None:
   ```



##########
airflow-core/src/airflow/utils/log/log_stream_counter.py:
##########
@@ -0,0 +1,151 @@
+#
+# Licensed to the Apache Software Foundation (ASF) under one
+# or more contributor license agreements.  See the NOTICE file
+# distributed with this work for additional information
+# regarding copyright ownership.  The ASF licenses this file
+# to you under the Apache License, Version 2.0 (the
+# "License"); you may not use this file except in compliance
+# with the License.  You may obtain a copy of the License at
+#
+#   http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing,
+# software distributed under the License is distributed on an
+# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
+# KIND, either express or implied.  See the License for the
+# specific language governing permissions and limitations
+# under the License.
+
+from __future__ import annotations
+
+import os
+import tempfile
+from typing import TYPE_CHECKING
+
+if TYPE_CHECKING:
+    from airflow.typing_compat import Self
+    from airflow.utils.log.file_task_handler import (
+        LogHandlerOutputStream,
+        StructuredLogMessage,
+        StructuredLogStream,
+    )
+
+
+class LogStreamCounter:
+    """
+    Memory-efficient log stream counter that tracks the total number of lines 
while preserving the original stream.
+
+    This class captures logs from a stream and stores them in a buffer, 
flushing them to disk when the buffer
+    exceeds a specified threshold. This approach optimizes memory usage while 
handling large log streams.
+
+    Usage:
+
+    .. code-block:: python
+
+        with LogStreamCounter(stream, threshold) as log_counter:
+            # Get total number of lines captured
+            total_lines = log_counter.get_total_lines()
+
+            # Retrieve the original stream of logs
+            for log in log_counter.get_stream():
+                print(log)
+    """
+
+    def __init__(
+        self,
+        stream: LogHandlerOutputStream,
+        threshold: int,
+    ):

Review Comment:
   ```suggestion
       ) -> None:
   ```



##########
airflow-core/src/airflow/utils/log/log_reader.py:
##########
@@ -83,20 +86,15 @@ def read_log_stream(self, ti: TaskInstance, try_number: int 
| None, metadata: di
             metadata.pop(key, None)
 
         while True:
-            logs, out_metadata = self.read_log_chunks(ti, try_number, metadata)
-            # Update the metadata dict in place so caller can get new 
values/end-of-log etc.
-
-            for log in logs:
-                # It's a bit wasteful here to parse the JSON then dump it back 
again.
-                # Optimize this so in stream mode we can just pass logs right 
through, or even better add
-                # support to 307 redirect to a signed URL etc.
-                yield (log if isinstance(log, str) else log.model_dump_json()) 
+ "\n"
+            log_stream, out_metadata = self.read_log_chunks(ti, try_number, 
metadata)
+            for log in log_stream:
+                yield log.model_dump_json() + "\n"

Review Comment:
   ```suggestion
               yield from (f"{log.model_dump_json()}\n" for log in log_stream)
   ```



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: [email protected]

For queries about this service, please contact Infrastructure at:
[email protected]

Reply via email to