jonathan-albrecht-ibm opened a new pull request, #49812:
URL: https://github.com/apache/spark/pull/49812

   ### What changes were proposed in this pull request?
   Fix minor endianness issues in the following tests.
   
   ArrayBasedMapBuilderSuite: The output of the UnsafeRow.toString() is based 
on the underlying bytes and is endian dependent. Add an expected value for big 
endian platforms. Add an expected value for big endian platforms.
   
   WriteDistributionAndOrderingSuite: Casting the id of type Int to Long 
doesn't work on big endian platforms because the BucketFunction calls 
UnsafeRow.getLong() for that column. That happens to work on little endian 
since an int field is stored in the first 4 bytes of the 8 byte field so 
positive ints are layed out the same as positive longs ie. little endian order. 
On big endian, the layout of UnsafeRow int fields does not happen to match the 
layout of long fields for the same number. Change the type of the id column to 
Long so that it matches what BucketFunction expects. Change the type of the id 
column to Long so that it matches what BucketFunction expects.
   
   ### Why are the changes needed?
   Allow tests to pass on big endian platforms
   
   
   ### Does this PR introduce _any_ user-facing change?
   No
   
   
   ### How was this patch tested?
   Ran existing tests on amd64 (little endian) and s390x (big endian)
   
   
   ### Was this patch authored or co-authored using generative AI tooling?
   No
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: [email protected]

For queries about this service, please contact Infrastructure at:
[email protected]


---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to