zeroshade commented on code in PR #654:
URL: https://github.com/apache/arrow-go/pull/654#discussion_r3001529641
##########
parquet/internal/encoding/byte_stream_split_big_endian.go:
##########
@@ -19,43 +19,98 @@
package encoding
import (
- "fmt"
+ "unsafe"
+ "github.com/apache/arrow-go/v18/parquet"
"github.com/apache/arrow-go/v18/parquet/internal/debug"
)
// decodeByteStreamSplitBatchWidth4InByteOrder decodes the batch of nValues
raw bytes representing a 4-byte datatype provided
-// by 'data', into the output buffer 'out' using BYTE_STREAM_SPLIT encoding.
The values are expected to be in little-endian
+// by 'data', into the output buffer 'out' using BYTE_STREAM_SPLIT encoding.
The values are expected to be in big-endian
// byte order and are be decoded into the 'out' array in machine's native
endianness.
// 'out' must have space for at least len(data) bytes.
-func decodeByteStreamSplitBatchWidth4InByteOrder(data []byte, nValues, stride
int, out []byte) {
+func decodeByteStreamSplitBatchWidth4InByteOrderDefault(data []byte, nValues,
stride int, out []byte) {
const width = 4
- debug.Assert(len(out) >= nValues*width, fmt.Sprintf("not enough space
in output buffer for decoding, out: %d bytes, data: %d bytes", len(out),
len(data)))
- for element := 0; element < nValues; element++ {
- // Big Endian: most significant byte first
- out[width*element+0] = data[3*stride+element]
- out[width*element+1] = data[2*stride+element]
- out[width*element+2] = data[stride+element]
- out[width*element+3] = data[element]
+ debug.Assert(len(out) >= nValues*width, "not enough space in output
buffer for decoding")
+ // the beginning of the data slice can be truncated, but for valid
encoding we need at least (width-1)*stride+nValues bytes
+ debug.Assert(len(data) >= 3*stride+nValues, "not enough data for
decoding")
+ s0 := data[:nValues]
+ s1 := data[stride : stride+nValues]
+ s2 := data[2*stride : 2*stride+nValues]
+ s3 := data[3*stride : 3*stride+nValues]
+ out = out[:width*nValues]
+ out32 := unsafe.Slice((*uint32)(unsafe.Pointer(&out[0])), nValues)
+ for i := range nValues {
+ // Big-endian machine: put s0 as MSB, s3 as LSB
+ out32[i] = uint32(s3[i])<<24 | uint32(s2[i])<<16 |
uint32(s1[i])<<8 | uint32(s0[i])
}
}
// decodeByteStreamSplitBatchWidth8InByteOrder decodes the batch of nValues
raw bytes representing a 8-byte datatype provided
-// by 'data', into the output buffer 'out' using BYTE_STREAM_SPLIT encoding.
The values are expected to be in little-endian
+// by 'data', into the output buffer 'out' using BYTE_STREAM_SPLIT encoding.
The values are expected to be in big-endian
// byte order and are be decoded into the 'out' array in machine's native
endianness.
// 'out' must have space for at least len(data) bytes.
-func decodeByteStreamSplitBatchWidth8InByteOrder(data []byte, nValues, stride
int, out []byte) {
+func decodeByteStreamSplitBatchWidth8InByteOrderDefault(data []byte, nValues,
stride int, out []byte) {
const width = 8
- debug.Assert(len(out) >= nValues*width, fmt.Sprintf("not enough space
in output buffer for decoding, out: %d bytes, data: %d bytes", len(out),
len(data)))
+ debug.Assert(len(out) >= nValues*width, "not enough space in output
buffer for decoding")
+ debug.Assert(len(data) >= 7*stride+nValues, "not enough data for
decoding")
+ s0 := data[:nValues]
+ s1 := data[stride : stride+nValues]
+ s2 := data[2*stride : 2*stride+nValues]
+ s3 := data[3*stride : 3*stride+nValues]
+ s4 := data[4*stride : 4*stride+nValues]
+ s5 := data[5*stride : 5*stride+nValues]
+ s6 := data[6*stride : 6*stride+nValues]
+ s7 := data[7*stride : 7*stride+nValues]
+ out = out[:width*nValues]
+ out64 := unsafe.Slice((*uint64)(unsafe.Pointer(&out[0])), nValues)
+ for i := range nValues {
+ // Big-endian machine: put s0 as MSB, s7 as LSB
+ out64[i] = uint64(s7[i])<<56 | uint64(s6[i])<<48 |
uint64(s5[i])<<40 | uint64(s4[i])<<32 |
+ uint64(s3[i])<<24 | uint64(s2[i])<<16 |
uint64(s1[i])<<8 | uint64(s0[i])
+ }
+}
+
+// decodeByteStreamSplitBatchFLBAWidth2 decodes the batch of nValues
FixedLenByteArrays of length 2 provided by 'data',
+// into the output slice 'out' using BYTE_STREAM_SPLIT encoding.
+// 'out' must have space for at least nValues slices.
+func decodeByteStreamSplitBatchFLBAWidth2(data []byte, nValues, stride int,
out []parquet.FixedLenByteArray) {
Review Comment:
I missed this, I believe that the output array would need the bytes to get
re-ordered on big-endian yes.
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
To unsubscribe, e-mail: [email protected]
For queries about this service, please contact Infrastructure at:
[email protected]