[ 
https://issues.apache.org/jira/browse/HADOOP-19645?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=18011079#comment-18011079
 ] 

ASF GitHub Bot commented on HADOOP-19645:
-----------------------------------------

Copilot commented on code in PR #7837:
URL: https://github.com/apache/hadoop/pull/7837#discussion_r2244303281


##########
hadoop-tools/hadoop-azure/src/main/java/org/apache/hadoop/fs/azurebfs/utils/TracingContext.java:
##########
@@ -265,6 +289,34 @@ private String addFailureReasons(final String header,
     return String.format("%s_%s", header, previousFailure);
   }
 
+  private String getRetryHeader(final String previousFailure, String 
retryPolicyAbbreviation) {
+    String retryHeader = String.format("%d", retryCount);
+    if (previousFailure == null) {
+      return retryHeader;
+    }
+    if (CONNECTION_TIMEOUT_ABBREVIATION.equals(previousFailure) && 
retryPolicyAbbreviation != null) {
+      return String.format("%s_%s_%s", retryHeader, previousFailure, 
retryPolicyAbbreviation);
+    }
+    return String.format("%s_%s", retryHeader, previousFailure);
+  }
+
+  private String getOperationSpecificHeader(FSOperationType opType) {
+    // Similar header can be added for other operations in the future.
+    switch (opType) {
+      case READ:
+        return readSpecificHeader();
+      default:
+        return EMPTY_STRING; // no operation specific header
+    }
+  }
+
+  private String readSpecificHeader() {
+    // More information on read can be added to this header in the future.
+    // As underscore separated values.
+    String readHeader = String.format("%s", readType.toString());

Review Comment:
   The String.format with "%s" is unnecessary here. Use readType.toString() 
directly for better readability and performance.
   ```suggestion
       String readHeader = readType.toString();
   ```



##########
hadoop-tools/hadoop-azure/src/main/java/org/apache/hadoop/fs/azurebfs/utils/TracingContext.java:
##########
@@ -193,31 +213,35 @@ public void setListener(Listener listener) {
   public void constructHeader(AbfsHttpOperation httpOperation, String 
previousFailure, String retryPolicyAbbreviation) {
     clientRequestId = UUID.randomUUID().toString();
     switch (format) {
-    case ALL_ID_FORMAT: // Optional IDs (e.g. streamId) may be empty
+    case ALL_ID_FORMAT:
       header =
-          clientCorrelationID + ":" + clientRequestId + ":" + fileSystemID + 
":"
-              + getPrimaryRequestIdForHeader(retryCount > 0) + ":" + streamID
-              + ":" + opType + ":" + retryCount;
-      header = addFailureReasons(header, previousFailure, 
retryPolicyAbbreviation);
-      if (!(ingressHandler.equals(EMPTY_STRING))) {
-        header += ":" + ingressHandler;
-      }
-      if (!(position.equals(EMPTY_STRING))) {
-        header += ":" + position;
-      }
-      if (operatedBlobCount != null) {
-        header += (":" + operatedBlobCount);
-      }
-      header += (":" + httpOperation.getTracingContextSuffix());
+          AbfsHttpConstants.TracingHeaderVersion.V1 + ":" +
+          clientCorrelationID + ":" +
+          clientRequestId + ":" +
+          fileSystemID + ":" +
+          getPrimaryRequestIdForHeader(retryCount > 0) + ":" +
+          streamID + ":" +
+          opType + ":" +
+          getRetryHeader(previousFailure, retryPolicyAbbreviation) + ":" +
+          ingressHandler + ":" +
+          position + ":" +
+          operatedBlobCount + ":" +
+          httpOperation.getTracingContextSuffix() + ":" +
+          getOperationSpecificHeader(opType);
+
       metricHeader += !(metricResults.trim().isEmpty()) ? metricResults  : "";
       break;
     case TWO_ID_FORMAT:
-      header = clientCorrelationID + ":" + clientRequestId;
+      header =
+          AbfsHttpConstants.TracingHeaderVersion.V1 + ":" +
+          clientCorrelationID + ":" + clientRequestId;
       metricHeader += !(metricResults.trim().isEmpty()) ? metricResults  : "";
       break;
     default:
       //case SINGLE_ID_FORMAT
-      header = clientRequestId;
+      header =
+          AbfsHttpConstants.TracingHeaderVersion.V1 + ":" +

Review Comment:
   The hardcoded V1 version is used in multiple places. Consider using 
TracingHeaderVersion.getCurrentVersion() consistently to centralize version 
management.



##########
hadoop-tools/hadoop-azure/src/main/java/org/apache/hadoop/fs/azurebfs/utils/TracingContext.java:
##########
@@ -193,31 +213,35 @@ public void setListener(Listener listener) {
   public void constructHeader(AbfsHttpOperation httpOperation, String 
previousFailure, String retryPolicyAbbreviation) {
     clientRequestId = UUID.randomUUID().toString();
     switch (format) {
-    case ALL_ID_FORMAT: // Optional IDs (e.g. streamId) may be empty
+    case ALL_ID_FORMAT:
       header =
-          clientCorrelationID + ":" + clientRequestId + ":" + fileSystemID + 
":"
-              + getPrimaryRequestIdForHeader(retryCount > 0) + ":" + streamID
-              + ":" + opType + ":" + retryCount;
-      header = addFailureReasons(header, previousFailure, 
retryPolicyAbbreviation);
-      if (!(ingressHandler.equals(EMPTY_STRING))) {
-        header += ":" + ingressHandler;
-      }
-      if (!(position.equals(EMPTY_STRING))) {
-        header += ":" + position;
-      }
-      if (operatedBlobCount != null) {
-        header += (":" + operatedBlobCount);
-      }
-      header += (":" + httpOperation.getTracingContextSuffix());
+          AbfsHttpConstants.TracingHeaderVersion.V1 + ":" +

Review Comment:
   The hardcoded V1 version is used in multiple places. Consider using 
TracingHeaderVersion.getCurrentVersion() consistently to centralize version 
management.



##########
hadoop-tools/hadoop-azure/src/main/java/org/apache/hadoop/fs/azurebfs/utils/TracingContext.java:
##########
@@ -193,31 +213,35 @@ public void setListener(Listener listener) {
   public void constructHeader(AbfsHttpOperation httpOperation, String 
previousFailure, String retryPolicyAbbreviation) {
     clientRequestId = UUID.randomUUID().toString();
     switch (format) {
-    case ALL_ID_FORMAT: // Optional IDs (e.g. streamId) may be empty
+    case ALL_ID_FORMAT:
       header =
-          clientCorrelationID + ":" + clientRequestId + ":" + fileSystemID + 
":"
-              + getPrimaryRequestIdForHeader(retryCount > 0) + ":" + streamID
-              + ":" + opType + ":" + retryCount;
-      header = addFailureReasons(header, previousFailure, 
retryPolicyAbbreviation);
-      if (!(ingressHandler.equals(EMPTY_STRING))) {
-        header += ":" + ingressHandler;
-      }
-      if (!(position.equals(EMPTY_STRING))) {
-        header += ":" + position;
-      }
-      if (operatedBlobCount != null) {
-        header += (":" + operatedBlobCount);
-      }
-      header += (":" + httpOperation.getTracingContextSuffix());
+          AbfsHttpConstants.TracingHeaderVersion.V1 + ":" +
+          clientCorrelationID + ":" +
+          clientRequestId + ":" +
+          fileSystemID + ":" +
+          getPrimaryRequestIdForHeader(retryCount > 0) + ":" +
+          streamID + ":" +
+          opType + ":" +
+          getRetryHeader(previousFailure, retryPolicyAbbreviation) + ":" +
+          ingressHandler + ":" +
+          position + ":" +
+          operatedBlobCount + ":" +
+          httpOperation.getTracingContextSuffix() + ":" +
+          getOperationSpecificHeader(opType);
+
       metricHeader += !(metricResults.trim().isEmpty()) ? metricResults  : "";
       break;
     case TWO_ID_FORMAT:
-      header = clientCorrelationID + ":" + clientRequestId;
+      header =
+          AbfsHttpConstants.TracingHeaderVersion.V1 + ":" +
+          clientCorrelationID + ":" + clientRequestId;
       metricHeader += !(metricResults.trim().isEmpty()) ? metricResults  : "";
       break;
     default:
       //case SINGLE_ID_FORMAT
-      header = clientRequestId;
+      header =
+          AbfsHttpConstants.TracingHeaderVersion.V1 + ":" +

Review Comment:
   The hardcoded V1 version is used in multiple places. Consider using 
TracingHeaderVersion.getCurrentVersion() consistently to centralize version 
management.



##########
hadoop-tools/hadoop-azure/src/test/java/org/apache/hadoop/fs/azurebfs/utils/TracingHeaderValidator.java:
##########
@@ -81,82 +85,93 @@ public TracingHeaderValidator(String clientCorrelationId, 
String fileSystemId,
   }
 
   private void validateTracingHeader(String tracingContextHeader) {
-    String[] idList = tracingContextHeader.split(":");
+    String[] idList = tracingContextHeader.split(":", -1);

Review Comment:
   [nitpick] Consider defining the split limit (-1) as a named constant to 
improve code readability and maintainability.
   ```suggestion
       String[] idList = tracingContextHeader.split(":", SPLIT_NO_LIMIT);
   ```





> ABFS: [ReadAheadV2] Improve Metrics for Read Calls to identify type of read 
> done.
> ---------------------------------------------------------------------------------
>
>                 Key: HADOOP-19645
>                 URL: https://issues.apache.org/jira/browse/HADOOP-19645
>             Project: Hadoop Common
>          Issue Type: Sub-task
>          Components: fs/azure
>    Affects Versions: 3.3.6, 3.4.1
>            Reporter: Anuj Modi
>            Assignee: Anuj Modi
>            Priority: Major
>              Labels: pull-request-available
>
> There are a number of ways in which ABFS driver can trigger a network call to 
> read data. We need a way to identify what type of read call was made from 
> client. Plan is to add an indication for this in already present 
> ClientRequestId header.
> Following are types of read we want to identify:
>  # Direct Read: Read from a given position in remote file. This will be 
> synchronous read
>  # Normal Read: Read from current seeked position where read ahead was 
> bypassed. This will be synchronous read.
>  # Prefetch Read: Read triggered from background threads filling up in memory 
> cache. This will be asynchronous read.
>  # Missed Cache Read: Read triggered after nothing was received from read 
> ahead. This will be synchronous read.
>  # Footer Read: Read triggered as part of footer read optimization. This will 
> be synchronous.
>  # Small File Read: Read triggered as a part of small file read. This will be 
> synchronous read.
> We will add another field in the Tracing Header (Client Request Id) for each 
> request. We can call this field "Operation Specific Header" very similar to 
> how we have "Retry Header" today. As part of this we will only use it for 
> read operations keeping it empty for other operations. Moving ahead f we need 
> to publish any operation specific info, same header can be used.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

---------------------------------------------------------------------
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org

Reply via email to