Use semicolons and braces. Signed-off-by: Joe Perches <j...@perches.com> --- lib/zstd/compress.c | 120 ++++++++++++++++++++++++++-------------- lib/zstd/fse_compress.c | 24 +++++--- lib/zstd/huf_compress.c | 6 +- 3 files changed, 100 insertions(+), 50 deletions(-)
diff --git a/lib/zstd/compress.c b/lib/zstd/compress.c index b080264ed3ad..d8587e458665 100644 --- a/lib/zstd/compress.c +++ b/lib/zstd/compress.c @@ -629,12 +629,17 @@ ZSTD_STATIC size_t ZSTD_compressSequences_internal(ZSTD_CCtx *zc, void *dst, siz /* Sequences Header */ if ((oend - op) < 3 /*max nbSeq Size*/ + 1 /*seqHead */) return ERROR(dstSize_tooSmall); - if (nbSeq < 0x7F) + if (nbSeq < 0x7F) { *op++ = (BYTE)nbSeq; - else if (nbSeq < LONGNBSEQ) - op[0] = (BYTE)((nbSeq >> 8) + 0x80), op[1] = (BYTE)nbSeq, op += 2; - else - op[0] = 0xFF, ZSTD_writeLE16(op + 1, (U16)(nbSeq - LONGNBSEQ)), op += 3; + } else if (nbSeq < LONGNBSEQ) { + op[0] = (BYTE)((nbSeq >> 8) + 0x80); + op[1] = (BYTE)nbSeq; + op += 2; + } else { + op[0] = 0xFF; + ZSTD_writeLE16(op + 1, (U16)(nbSeq - LONGNBSEQ)); + op += 3; + } if (nbSeq == 0) return op - ostart; @@ -1025,10 +1030,13 @@ void ZSTD_compressBlock_fast_generic(ZSTD_CCtx *cctx, const void *src, size_t sr ip += (ip == lowest); { U32 const maxRep = (U32)(ip - lowest); - if (offset_2 > maxRep) - offsetSaved = offset_2, offset_2 = 0; - if (offset_1 > maxRep) - offsetSaved = offset_1, offset_1 = 0; + if (offset_2 > maxRep) { + offsetSaved = offset_2; + offset_2 = 0; + } + if (offset_1 > maxRep) { + offsetSaved = offset_1; offset_1 = 0; + } } /* Main Search Loop */ @@ -1273,10 +1281,14 @@ void ZSTD_compressBlock_doubleFast_generic(ZSTD_CCtx *cctx, const void *src, siz ip += (ip == lowest); { U32 const maxRep = (U32)(ip - lowest); - if (offset_2 > maxRep) - offsetSaved = offset_2, offset_2 = 0; - if (offset_1 > maxRep) - offsetSaved = offset_1, offset_1 = 0; + if (offset_2 > maxRep) { + offsetSaved = offset_2; + offset_2 = 0; + } + if (offset_1 > maxRep) { + offsetSaved = offset_1; + offset_1 = 0; + } } /* Main Search Loop */ @@ -1686,8 +1698,10 @@ static size_t ZSTD_insertBtAndFindBestMatch(ZSTD_CCtx *zc, const BYTE *const ip, if (matchLength > bestLength) { if (matchLength > matchEndIdx - matchIndex) matchEndIdx = matchIndex + (U32)matchLength; - if ((4 * (int)(matchLength - bestLength)) > (int)(ZSTD_highbit32(curr - matchIndex + 1) - ZSTD_highbit32((U32)offsetPtr[0] + 1))) - bestLength = matchLength, *offsetPtr = ZSTD_REP_MOVE + curr - matchIndex; + if ((4 * (int)(matchLength - bestLength)) > (int)(ZSTD_highbit32(curr - matchIndex + 1) - ZSTD_highbit32((U32)offsetPtr[0] + 1))) { + bestLength = matchLength; + *offsetPtr = ZSTD_REP_MOVE + curr - matchIndex; + } if (ip + matchLength == iend) /* equal : no way to know if inf or sup */ break; /* drop, to guarantee consistency (miss a little bit of compression) */ } @@ -1916,10 +1930,14 @@ void ZSTD_compressBlock_lazy_generic(ZSTD_CCtx *ctx, const void *src, size_t src ctx->nextToUpdate3 = ctx->nextToUpdate; { U32 const maxRep = (U32)(ip - base); - if (offset_2 > maxRep) - savedOffset = offset_2, offset_2 = 0; - if (offset_1 > maxRep) - savedOffset = offset_1, offset_1 = 0; + if (offset_2 > maxRep) { + savedOffset = offset_2; + offset_2 = 0; + } + if (offset_1 > maxRep) { + savedOffset = offset_1; + offset_1 = 0; + } } /* Match Loop */ @@ -1940,8 +1958,11 @@ void ZSTD_compressBlock_lazy_generic(ZSTD_CCtx *ctx, const void *src, size_t src { size_t offsetFound = 99999999; size_t const ml2 = searchMax(ctx, ip, iend, &offsetFound, maxSearches, mls); - if (ml2 > matchLength) - matchLength = ml2, start = ip, offset = offsetFound; + if (ml2 > matchLength) { + matchLength = ml2; + start = ip; + offset = offsetFound; + } } if (matchLength < EQUAL_READ32) { @@ -1957,8 +1978,11 @@ void ZSTD_compressBlock_lazy_generic(ZSTD_CCtx *ctx, const void *src, size_t src size_t const mlRep = ZSTD_count(ip + EQUAL_READ32, ip + EQUAL_READ32 - offset_1, iend) + EQUAL_READ32; int const gain2 = (int)(mlRep * 3); int const gain1 = (int)(matchLength * 3 - ZSTD_highbit32((U32)offset + 1) + 1); - if ((mlRep >= EQUAL_READ32) && (gain2 > gain1)) - matchLength = mlRep, offset = 0, start = ip; + if ((mlRep >= EQUAL_READ32) && (gain2 > gain1)) { + matchLength = mlRep; + offset = 0; + start = ip; + } } { size_t offset2 = 99999999; @@ -1966,7 +1990,7 @@ void ZSTD_compressBlock_lazy_generic(ZSTD_CCtx *ctx, const void *src, size_t src int const gain2 = (int)(ml2 * 4 - ZSTD_highbit32((U32)offset2 + 1)); /* raw approx */ int const gain1 = (int)(matchLength * 4 - ZSTD_highbit32((U32)offset + 1) + 4); if ((ml2 >= EQUAL_READ32) && (gain2 > gain1)) { - matchLength = ml2, offset = offset2, start = ip; + matchLength = ml2, offset = offset2; start = ip; continue; /* search a better one */ } } @@ -1978,8 +2002,11 @@ void ZSTD_compressBlock_lazy_generic(ZSTD_CCtx *ctx, const void *src, size_t src size_t const ml2 = ZSTD_count(ip + EQUAL_READ32, ip + EQUAL_READ32 - offset_1, iend) + EQUAL_READ32; int const gain2 = (int)(ml2 * 4); int const gain1 = (int)(matchLength * 4 - ZSTD_highbit32((U32)offset + 1) + 1); - if ((ml2 >= EQUAL_READ32) && (gain2 > gain1)) - matchLength = ml2, offset = 0, start = ip; + if ((ml2 >= EQUAL_READ32) && (gain2 > gain1)) { + matchLength = ml2; + offset = 0; + start = ip; + } } { size_t offset2 = 99999999; @@ -1987,7 +2014,7 @@ void ZSTD_compressBlock_lazy_generic(ZSTD_CCtx *ctx, const void *src, size_t src int const gain2 = (int)(ml2 * 4 - ZSTD_highbit32((U32)offset2 + 1)); /* raw approx */ int const gain1 = (int)(matchLength * 4 - ZSTD_highbit32((U32)offset + 1) + 7); if ((ml2 >= EQUAL_READ32) && (gain2 > gain1)) { - matchLength = ml2, offset = offset2, start = ip; + matchLength = ml2, offset = offset2; start = ip; continue; } } @@ -2110,8 +2137,11 @@ void ZSTD_compressBlock_lazy_extDict_generic(ZSTD_CCtx *ctx, const void *src, si { size_t offsetFound = 99999999; size_t const ml2 = searchMax(ctx, ip, iend, &offsetFound, maxSearches, mls); - if (ml2 > matchLength) - matchLength = ml2, start = ip, offset = offsetFound; + if (ml2 > matchLength) { + matchLength = ml2; + start = ip; + offset = offsetFound; + } } if (matchLength < EQUAL_READ32) { @@ -2138,8 +2168,11 @@ void ZSTD_compressBlock_lazy_extDict_generic(ZSTD_CCtx *ctx, const void *src, si EQUAL_READ32; int const gain2 = (int)(repLength * 3); int const gain1 = (int)(matchLength * 3 - ZSTD_highbit32((U32)offset + 1) + 1); - if ((repLength >= EQUAL_READ32) && (gain2 > gain1)) - matchLength = repLength, offset = 0, start = ip; + if ((repLength >= EQUAL_READ32) && (gain2 > gain1)) { + matchLength = repLength; + offset = 0; + start = ip; + } } } @@ -2150,7 +2183,7 @@ void ZSTD_compressBlock_lazy_extDict_generic(ZSTD_CCtx *ctx, const void *src, si int const gain2 = (int)(ml2 * 4 - ZSTD_highbit32((U32)offset2 + 1)); /* raw approx */ int const gain1 = (int)(matchLength * 4 - ZSTD_highbit32((U32)offset + 1) + 4); if ((ml2 >= EQUAL_READ32) && (gain2 > gain1)) { - matchLength = ml2, offset = offset2, start = ip; + matchLength = ml2, offset = offset2; start = ip; continue; /* search a better one */ } } @@ -2173,8 +2206,11 @@ void ZSTD_compressBlock_lazy_extDict_generic(ZSTD_CCtx *ctx, const void *src, si EQUAL_READ32; int gain2 = (int)(repLength * 4); int gain1 = (int)(matchLength * 4 - ZSTD_highbit32((U32)offset + 1) + 1); - if ((repLength >= EQUAL_READ32) && (gain2 > gain1)) - matchLength = repLength, offset = 0, start = ip; + if ((repLength >= EQUAL_READ32) && (gain2 > gain1)) { + matchLength = repLength; + offset = 0; + start = ip; + } } } @@ -2185,7 +2221,7 @@ void ZSTD_compressBlock_lazy_extDict_generic(ZSTD_CCtx *ctx, const void *src, si int const gain2 = (int)(ml2 * 4 - ZSTD_highbit32((U32)offset2 + 1)); /* raw approx */ int const gain1 = (int)(matchLength * 4 - ZSTD_highbit32((U32)offset + 1) + 7); if ((ml2 >= EQUAL_READ32) && (gain2 > gain1)) { - matchLength = ml2, offset = offset2, start = ip; + matchLength = ml2, offset = offset2; start = ip; continue; } } @@ -3159,10 +3195,12 @@ static size_t ZSTD_compressStream_generic(ZSTD_CStream *zcs, void *dst, size_t * size_t cSize; size_t const iSize = zcs->inBuffPos - zcs->inToCompress; size_t oSize = oend - op; - if (oSize >= ZSTD_compressBound(iSize)) + if (oSize >= ZSTD_compressBound(iSize)) { cDst = op; /* compress directly into output buffer (avoid flush stage) */ - else - cDst = zcs->outBuff, oSize = zcs->outBuffSize; + } else { + cDst = zcs->outBuff; + oSize = zcs->outBuffSize; + } cSize = (flush == zsf_end) ? ZSTD_compressEnd(zcs->cctx, cDst, oSize, zcs->inBuff + zcs->inToCompress, iSize) : ZSTD_compressContinue(zcs->cctx, cDst, oSize, zcs->inBuff + zcs->inToCompress, iSize); if (ZSTD_isError(cSize)) @@ -3171,8 +3209,10 @@ static size_t ZSTD_compressStream_generic(ZSTD_CStream *zcs, void *dst, size_t * zcs->frameEnded = 1; /* prepare next block */ zcs->inBuffTarget = zcs->inBuffPos + zcs->blockSize; - if (zcs->inBuffTarget > zcs->inBuffSize) - zcs->inBuffPos = 0, zcs->inBuffTarget = zcs->blockSize; /* note : inBuffSize >= blockSize */ + if (zcs->inBuffTarget > zcs->inBuffSize) { + zcs->inBuffPos = 0; + zcs->inBuffTarget = zcs->blockSize; + } /* note : inBuffSize >= blockSize */ zcs->inToCompress = zcs->inBuffPos; if (cDst == op) { op += cSize; diff --git a/lib/zstd/fse_compress.c b/lib/zstd/fse_compress.c index ef3d1741d532..99a21c35d0ac 100644 --- a/lib/zstd/fse_compress.c +++ b/lib/zstd/fse_compress.c @@ -271,8 +271,10 @@ static size_t FSE_writeNCount_generic(void *header, size_t headerBufferSize, con previous0 = (count == 1); if (remaining < 1) return ERROR(GENERIC); - while (remaining < threshold) - nbBits--, threshold >>= 1; + while (remaining < threshold) { + nbBits--; + threshold >>= 1; + } } if (bitCount > 16) { if ((!writeIsSafe) && (out > oend - 2)) @@ -569,8 +571,10 @@ static size_t FSE_normalizeM2(short *norm, U32 tableLog, const unsigned *count, find max, then give all remaining points to max */ U32 maxV = 0, maxC = 0; for (s = 0; s <= maxSymbolValue; s++) - if (count[s] > maxC) - maxV = s, maxC = count[s]; + if (count[s] > maxC) { + maxV = s; + maxC = count[s]; + } norm[maxV] += (short)ToDistribute; return 0; } @@ -578,8 +582,10 @@ static size_t FSE_normalizeM2(short *norm, U32 tableLog, const unsigned *count, if (total == 0) { /* all of the symbols were low enough for the lowOne or lowThreshold */ for (s = 0; ToDistribute > 0; s = (s + 1) % (maxSymbolValue + 1)) - if (norm[s] > 0) - ToDistribute--, norm[s]++; + if (norm[s] > 0) { + ToDistribute--; + norm[s]++; + } return 0; } @@ -644,8 +650,10 @@ size_t FSE_normalizeCount(short *normalizedCounter, unsigned tableLog, const uns U64 restToBeat = vStep * rtbTable[proba]; proba += (count[s] * step) - ((U64)proba << scale) > restToBeat; } - if (proba > largestP) - largestP = proba, largest = s; + if (proba > largestP) { + largestP = proba; + largest = s; + } normalizedCounter[s] = proba; stillToDistribute -= proba; } diff --git a/lib/zstd/huf_compress.c b/lib/zstd/huf_compress.c index 08b4ae80aed4..ca12ab400086 100644 --- a/lib/zstd/huf_compress.c +++ b/lib/zstd/huf_compress.c @@ -404,8 +404,10 @@ static void HUF_sort(nodeElt *huffNode, const U32 *count, U32 maxSymbolValue) U32 const c = count[n]; U32 const r = BIT_highbit32(c + 1) + 1; U32 pos = rank[r].curr++; - while ((pos > rank[r].base) && (c > huffNode[pos - 1].count)) - huffNode[pos] = huffNode[pos - 1], pos--; + while ((pos > rank[r].base) && (c > huffNode[pos - 1].count)) { + huffNode[pos] = huffNode[pos - 1]; + pos--; + } huffNode[pos].count = c; huffNode[pos].byte = (BYTE)n; } -- 2.26.0