mirror of
https://github.com/containers/skopeo.git
synced 2025-09-22 18:37:21 +00:00
Image encryption/decryption support in skopeo
Signed-off-by: Harshal Patil <harshal.patil@in.ibm.com> Signed-off-by: Brandon Lum <lumjjb@gmail.com>
This commit is contained in:
11
vendor/github.com/klauspost/compress/zstd/blockdec.go
generated
vendored
11
vendor/github.com/klauspost/compress/zstd/blockdec.go
generated
vendored
@@ -11,6 +11,7 @@ import (
|
||||
"sync"
|
||||
|
||||
"github.com/klauspost/compress/huff0"
|
||||
"github.com/klauspost/compress/zstd/internal/xxhash"
|
||||
)
|
||||
|
||||
type blockType uint8
|
||||
@@ -160,7 +161,8 @@ func (b *blockDec) reset(br byteBuffer, windowSize uint64) error {
|
||||
b.data, err = br.readBig(cSize, b.dataStorage)
|
||||
if err != nil {
|
||||
if debug {
|
||||
println("Reading block:", err)
|
||||
println("Reading block:", err, "(", cSize, ")", len(b.data))
|
||||
printf("%T", br)
|
||||
}
|
||||
return err
|
||||
}
|
||||
@@ -275,7 +277,7 @@ func (b *blockDec) decodeBuf(hist *history) error {
|
||||
hist.b = nil
|
||||
err := b.decodeCompressed(hist)
|
||||
if debug {
|
||||
println("Decompressed to total", len(b.dst), "bytes, error:", err)
|
||||
println("Decompressed to total", len(b.dst), "bytes, hash:", xxhash.Sum64(b.dst), "error:", err)
|
||||
}
|
||||
hist.b = b.dst
|
||||
b.dst = saved
|
||||
@@ -368,7 +370,7 @@ func (b *blockDec) decodeCompressed(hist *history) error {
|
||||
}
|
||||
}
|
||||
if debug {
|
||||
println("literals type:", litType, "litRegenSize:", litRegenSize, "litCompSize", litCompSize)
|
||||
println("literals type:", litType, "litRegenSize:", litRegenSize, "litCompSize:", litCompSize, "sizeFormat:", sizeFormat, "4X:", fourStreams)
|
||||
}
|
||||
var literals []byte
|
||||
var huff *huff0.Scratch
|
||||
@@ -426,7 +428,6 @@ func (b *blockDec) decodeCompressed(hist *history) error {
|
||||
}
|
||||
literals = in[:litCompSize]
|
||||
in = in[litCompSize:]
|
||||
|
||||
huff = huffDecoderPool.Get().(*huff0.Scratch)
|
||||
var err error
|
||||
// Ensure we have space to store it.
|
||||
@@ -637,7 +638,7 @@ func (b *blockDec) decodeCompressed(hist *history) error {
|
||||
hist.huffTree = huff
|
||||
}
|
||||
if debug {
|
||||
println("Final literals:", len(literals), "and", nSeqs, "sequences.")
|
||||
println("Final literals:", len(literals), "hash:", xxhash.Sum64(literals), "and", nSeqs, "sequences.")
|
||||
}
|
||||
|
||||
if nSeqs == 0 {
|
||||
|
56
vendor/github.com/klauspost/compress/zstd/blockenc.go
generated
vendored
56
vendor/github.com/klauspost/compress/zstd/blockenc.go
generated
vendored
@@ -51,7 +51,7 @@ func (b *blockEnc) init() {
|
||||
b.coders.llEnc = &fseEncoder{}
|
||||
b.coders.llPrev = &fseEncoder{}
|
||||
}
|
||||
b.litEnc = &huff0.Scratch{}
|
||||
b.litEnc = &huff0.Scratch{WantLogLess: 4}
|
||||
b.reset(nil)
|
||||
}
|
||||
|
||||
@@ -391,6 +391,52 @@ func (b *blockEnc) encodeLits() error {
|
||||
return nil
|
||||
}
|
||||
|
||||
// fuzzFseEncoder can be used to fuzz the FSE encoder.
|
||||
func fuzzFseEncoder(data []byte) int {
|
||||
if len(data) > maxSequences || len(data) < 2 {
|
||||
return 0
|
||||
}
|
||||
enc := fseEncoder{}
|
||||
hist := enc.Histogram()[:256]
|
||||
maxSym := uint8(0)
|
||||
for i, v := range data {
|
||||
v = v & 63
|
||||
data[i] = v
|
||||
hist[v]++
|
||||
if v > maxSym {
|
||||
maxSym = v
|
||||
}
|
||||
}
|
||||
if maxSym == 0 {
|
||||
// All 0
|
||||
return 0
|
||||
}
|
||||
maxCount := func(a []uint32) int {
|
||||
var max uint32
|
||||
for _, v := range a {
|
||||
if v > max {
|
||||
max = v
|
||||
}
|
||||
}
|
||||
return int(max)
|
||||
}
|
||||
cnt := maxCount(hist[:maxSym])
|
||||
if cnt == len(data) {
|
||||
// RLE
|
||||
return 0
|
||||
}
|
||||
enc.HistogramFinished(maxSym, cnt)
|
||||
err := enc.normalizeCount(len(data))
|
||||
if err != nil {
|
||||
return 0
|
||||
}
|
||||
_, err = enc.writeCount(nil)
|
||||
if err != nil {
|
||||
panic(err)
|
||||
}
|
||||
return 1
|
||||
}
|
||||
|
||||
// encode will encode the block and put the output in b.output.
|
||||
func (b *blockEnc) encode() error {
|
||||
if len(b.sequences) == 0 {
|
||||
@@ -415,16 +461,10 @@ func (b *blockEnc) encode() error {
|
||||
if len(b.literals) >= 1024 {
|
||||
// Use 4 Streams.
|
||||
out, reUsed, err = huff0.Compress4X(b.literals, b.litEnc)
|
||||
if len(out) > len(b.literals)-len(b.literals)>>4 {
|
||||
err = huff0.ErrIncompressible
|
||||
}
|
||||
} else if len(b.literals) > 32 {
|
||||
// Use 1 stream
|
||||
single = true
|
||||
out, reUsed, err = huff0.Compress1X(b.literals, b.litEnc)
|
||||
if len(out) > len(b.literals)-len(b.literals)>>4 {
|
||||
err = huff0.ErrIncompressible
|
||||
}
|
||||
} else {
|
||||
err = huff0.ErrIncompressible
|
||||
}
|
||||
@@ -711,7 +751,7 @@ func (b *blockEnc) encode() error {
|
||||
return nil
|
||||
}
|
||||
|
||||
var errIncompressible = errors.New("uncompressible")
|
||||
var errIncompressible = errors.New("incompressible")
|
||||
|
||||
func (b *blockEnc) genCodes() {
|
||||
if len(b.sequences) == 0 {
|
||||
|
3
vendor/github.com/klauspost/compress/zstd/bytebuf.go
generated
vendored
3
vendor/github.com/klauspost/compress/zstd/bytebuf.go
generated
vendored
@@ -101,6 +101,9 @@ func (r *readerWrapper) readBig(n int, dst []byte) ([]byte, error) {
|
||||
dst = make([]byte, n)
|
||||
}
|
||||
n2, err := io.ReadFull(r.r, dst[:n])
|
||||
if err == io.EOF && n > 0 {
|
||||
err = io.ErrUnexpectedEOF
|
||||
}
|
||||
return dst[:n2], err
|
||||
}
|
||||
|
||||
|
26
vendor/github.com/klauspost/compress/zstd/decoder.go
generated
vendored
26
vendor/github.com/klauspost/compress/zstd/decoder.go
generated
vendored
@@ -75,6 +75,7 @@ var (
|
||||
// The Reset function can be used to initiate a new stream, which is will considerably
|
||||
// reduce the allocations normally caused by NewReader.
|
||||
func NewReader(r io.Reader, opts ...DOption) (*Decoder, error) {
|
||||
initPredefined()
|
||||
var d Decoder
|
||||
d.o.setDefault()
|
||||
for _, o := range opts {
|
||||
@@ -123,7 +124,9 @@ func (d *Decoder) Read(p []byte) (int, error) {
|
||||
if d.current.err != nil {
|
||||
break
|
||||
}
|
||||
d.nextBlock()
|
||||
if !d.nextBlock(n == 0) {
|
||||
return n, nil
|
||||
}
|
||||
}
|
||||
}
|
||||
if len(d.current.b) > 0 {
|
||||
@@ -251,7 +254,7 @@ func (d *Decoder) WriteTo(w io.Writer) (int64, error) {
|
||||
if d.current.err != nil {
|
||||
break
|
||||
}
|
||||
d.nextBlock()
|
||||
d.nextBlock(true)
|
||||
}
|
||||
err := d.current.err
|
||||
if err != nil {
|
||||
@@ -328,7 +331,10 @@ func (d *Decoder) DecodeAll(input, dst []byte) ([]byte, error) {
|
||||
|
||||
// nextBlock returns the next block.
|
||||
// If an error occurs d.err will be set.
|
||||
func (d *Decoder) nextBlock() {
|
||||
// Optionally the function can block for new output.
|
||||
// If non-blocking mode is used the returned boolean will be false
|
||||
// if no data was available without blocking.
|
||||
func (d *Decoder) nextBlock(blocking bool) (ok bool) {
|
||||
if d.current.d != nil {
|
||||
if debug {
|
||||
printf("re-adding current decoder %p", d.current.d)
|
||||
@@ -338,12 +344,22 @@ func (d *Decoder) nextBlock() {
|
||||
}
|
||||
if d.current.err != nil {
|
||||
// Keep error state.
|
||||
return
|
||||
return blocking
|
||||
}
|
||||
|
||||
if blocking {
|
||||
d.current.decodeOutput = <-d.current.output
|
||||
} else {
|
||||
select {
|
||||
case d.current.decodeOutput = <-d.current.output:
|
||||
default:
|
||||
return false
|
||||
}
|
||||
}
|
||||
d.current.decodeOutput = <-d.current.output
|
||||
if debug {
|
||||
println("got", len(d.current.b), "bytes, error:", d.current.err)
|
||||
}
|
||||
return true
|
||||
}
|
||||
|
||||
// Close will release all resources.
|
||||
|
6
vendor/github.com/klauspost/compress/zstd/enc_dfast.go
generated
vendored
6
vendor/github.com/klauspost/compress/zstd/enc_dfast.go
generated
vendored
@@ -235,7 +235,7 @@ encodeLoop:
|
||||
if debug && s-t > e.maxMatchOff {
|
||||
panic("s - t >e.maxMatchOff")
|
||||
}
|
||||
if debug {
|
||||
if debugMatches {
|
||||
println("long match")
|
||||
}
|
||||
break
|
||||
@@ -259,7 +259,7 @@ encodeLoop:
|
||||
// but the likelihood of both the first 4 bytes and the hash matching should be enough.
|
||||
t = candidateL.offset - e.cur
|
||||
s += checkAt
|
||||
if debug {
|
||||
if debugMatches {
|
||||
println("long match (after short)")
|
||||
}
|
||||
break
|
||||
@@ -275,7 +275,7 @@ encodeLoop:
|
||||
if debug && t < 0 {
|
||||
panic("t<0")
|
||||
}
|
||||
if debug {
|
||||
if debugMatches {
|
||||
println("short match")
|
||||
}
|
||||
break
|
||||
|
22
vendor/github.com/klauspost/compress/zstd/encoder.go
generated
vendored
22
vendor/github.com/klauspost/compress/zstd/encoder.go
generated
vendored
@@ -59,6 +59,7 @@ type encoderState struct {
|
||||
// NewWriter will create a new Zstandard encoder.
|
||||
// If the encoder will be used for encoding blocks a nil writer can be used.
|
||||
func NewWriter(w io.Writer, opts ...EOption) (*Encoder, error) {
|
||||
initPredefined()
|
||||
var e Encoder
|
||||
e.o.setDefault()
|
||||
for _, o := range opts {
|
||||
@@ -393,12 +394,31 @@ func (e *Encoder) Close() error {
|
||||
|
||||
// EncodeAll will encode all input in src and append it to dst.
|
||||
// This function can be called concurrently, but each call will only run on a single goroutine.
|
||||
// If empty input is given, nothing is returned.
|
||||
// If empty input is given, nothing is returned, unless WithZeroFrames is specified.
|
||||
// Encoded blocks can be concatenated and the result will be the combined input stream.
|
||||
// Data compressed with EncodeAll can be decoded with the Decoder,
|
||||
// using either a stream or DecodeAll.
|
||||
func (e *Encoder) EncodeAll(src, dst []byte) []byte {
|
||||
if len(src) == 0 {
|
||||
if e.o.fullZero {
|
||||
// Add frame header.
|
||||
fh := frameHeader{
|
||||
ContentSize: 0,
|
||||
WindowSize: MinWindowSize,
|
||||
SingleSegment: true,
|
||||
// Adding a checksum would be a waste of space.
|
||||
Checksum: false,
|
||||
DictID: 0,
|
||||
}
|
||||
dst, _ = fh.appendTo(dst)
|
||||
|
||||
// Write raw block as last one only.
|
||||
var blk blockHeader
|
||||
blk.setSize(0)
|
||||
blk.setType(blockTypeRaw)
|
||||
blk.setLast(true)
|
||||
dst = blk.appendTo(dst)
|
||||
}
|
||||
return dst
|
||||
}
|
||||
e.init.Do(func() {
|
||||
|
36
vendor/github.com/klauspost/compress/zstd/encoder_options.go
generated
vendored
36
vendor/github.com/klauspost/compress/zstd/encoder_options.go
generated
vendored
@@ -1,6 +1,7 @@
|
||||
package zstd
|
||||
|
||||
import (
|
||||
"errors"
|
||||
"fmt"
|
||||
"runtime"
|
||||
"strings"
|
||||
@@ -18,6 +19,7 @@ type encoderOptions struct {
|
||||
blockSize int
|
||||
windowSize int
|
||||
level EncoderLevel
|
||||
fullZero bool
|
||||
}
|
||||
|
||||
func (o *encoderOptions) setDefault() {
|
||||
@@ -63,6 +65,30 @@ func WithEncoderConcurrency(n int) EOption {
|
||||
}
|
||||
}
|
||||
|
||||
// WithWindowSize will set the maximum allowed back-reference distance.
|
||||
// The value must be a power of two between WindowSizeMin and WindowSizeMax.
|
||||
// A larger value will enable better compression but allocate more memory and,
|
||||
// for above-default values, take considerably longer.
|
||||
// The default value is determined by the compression level.
|
||||
func WithWindowSize(n int) EOption {
|
||||
return func(o *encoderOptions) error {
|
||||
switch {
|
||||
case n < MinWindowSize:
|
||||
return fmt.Errorf("window size must be at least %d", MinWindowSize)
|
||||
case n > MaxWindowSize:
|
||||
return fmt.Errorf("window size must be at most %d", MaxWindowSize)
|
||||
case (n & (n - 1)) != 0:
|
||||
return errors.New("window size must be a power of 2")
|
||||
}
|
||||
|
||||
o.windowSize = n
|
||||
if o.blockSize > o.windowSize {
|
||||
o.blockSize = o.windowSize
|
||||
}
|
||||
return nil
|
||||
}
|
||||
}
|
||||
|
||||
// WithEncoderPadding will add padding to all output so the size will be a multiple of n.
|
||||
// This can be used to obfuscate the exact output size or make blocks of a certain size.
|
||||
// The contents will be a skippable frame, so it will be invisible by the decoder.
|
||||
@@ -166,6 +192,16 @@ func WithEncoderLevel(l EncoderLevel) EOption {
|
||||
}
|
||||
}
|
||||
|
||||
// WithZeroFrames will encode 0 length input as full frames.
|
||||
// This can be needed for compatibility with zstandard usage,
|
||||
// but is not needed for this package.
|
||||
func WithZeroFrames(b bool) EOption {
|
||||
return func(o *encoderOptions) error {
|
||||
o.fullZero = b
|
||||
return nil
|
||||
}
|
||||
}
|
||||
|
||||
// WithSingleSegment will set the "single segment" flag when EncodeAll is used.
|
||||
// If this flag is set, data must be regenerated within a single continuous memory segment.
|
||||
// In this case, Window_Descriptor byte is skipped, but Frame_Content_Size is necessarily present.
|
||||
|
21
vendor/github.com/klauspost/compress/zstd/framedec.go
generated
vendored
21
vendor/github.com/klauspost/compress/zstd/framedec.go
generated
vendored
@@ -49,7 +49,8 @@ type frameDec struct {
|
||||
|
||||
const (
|
||||
// The minimum Window_Size is 1 KB.
|
||||
minWindowSize = 1 << 10
|
||||
MinWindowSize = 1 << 10
|
||||
MaxWindowSize = 1 << 30
|
||||
)
|
||||
|
||||
var (
|
||||
@@ -60,7 +61,7 @@ var (
|
||||
func newFrameDec(o decoderOptions) *frameDec {
|
||||
d := frameDec{
|
||||
o: o,
|
||||
maxWindowSize: 1 << 30,
|
||||
maxWindowSize: MaxWindowSize,
|
||||
}
|
||||
if d.maxWindowSize > o.maxDecodedSize {
|
||||
d.maxWindowSize = o.maxDecodedSize
|
||||
@@ -193,14 +194,14 @@ func (d *frameDec) reset(br byteBuffer) error {
|
||||
// When FCS_Field_Size is 2, the offset of 256 is added.
|
||||
d.FrameContentSize = uint64(b[0]) | (uint64(b[1]) << 8) + 256
|
||||
case 4:
|
||||
d.FrameContentSize = uint64(b[0]) | (uint64(b[1]) << 8) | (uint64(b[2]) << 16) | (uint64(b[3] << 24))
|
||||
d.FrameContentSize = uint64(b[0]) | (uint64(b[1]) << 8) | (uint64(b[2]) << 16) | (uint64(b[3]) << 24)
|
||||
case 8:
|
||||
d1 := uint32(b[0]) | (uint32(b[1]) << 8) | (uint32(b[2]) << 16) | (uint32(b[3]) << 24)
|
||||
d2 := uint32(b[4]) | (uint32(b[5]) << 8) | (uint32(b[6]) << 16) | (uint32(b[7]) << 24)
|
||||
d.FrameContentSize = uint64(d1) | (uint64(d2) << 32)
|
||||
}
|
||||
if debug {
|
||||
println("field size bits:", v, "fcsSize:", fcsSize, "FrameContentSize:", d.FrameContentSize, hex.EncodeToString(b[:fcsSize]))
|
||||
println("field size bits:", v, "fcsSize:", fcsSize, "FrameContentSize:", d.FrameContentSize, hex.EncodeToString(b[:fcsSize]), "singleseg:", d.SingleSegment, "window:", d.WindowSize)
|
||||
}
|
||||
}
|
||||
// Move this to shared.
|
||||
@@ -215,8 +216,8 @@ func (d *frameDec) reset(br byteBuffer) error {
|
||||
if d.WindowSize == 0 && d.SingleSegment {
|
||||
// We may not need window in this case.
|
||||
d.WindowSize = d.FrameContentSize
|
||||
if d.WindowSize < minWindowSize {
|
||||
d.WindowSize = minWindowSize
|
||||
if d.WindowSize < MinWindowSize {
|
||||
d.WindowSize = MinWindowSize
|
||||
}
|
||||
}
|
||||
|
||||
@@ -225,7 +226,7 @@ func (d *frameDec) reset(br byteBuffer) error {
|
||||
return ErrWindowSizeExceeded
|
||||
}
|
||||
// The minimum Window_Size is 1 KB.
|
||||
if d.WindowSize < minWindowSize {
|
||||
if d.WindowSize < MinWindowSize {
|
||||
println("got window size: ", d.WindowSize)
|
||||
return ErrWindowSizeTooSmall
|
||||
}
|
||||
@@ -309,7 +310,9 @@ func (d *frameDec) checkCRC() error {
|
||||
}
|
||||
return ErrCRCMismatch
|
||||
}
|
||||
println("CRC ok")
|
||||
if debug {
|
||||
println("CRC ok", tmp[:])
|
||||
}
|
||||
return nil
|
||||
}
|
||||
|
||||
@@ -411,6 +414,7 @@ func (d *frameDec) startDecoder(output chan decodeOutput) {
|
||||
}
|
||||
written += int64(len(r.b))
|
||||
if d.SingleSegment && uint64(written) > d.FrameContentSize {
|
||||
println("runDecoder: single segment and", uint64(written), ">", d.FrameContentSize)
|
||||
r.err = ErrFrameSizeExceeded
|
||||
output <- r
|
||||
return
|
||||
@@ -461,6 +465,7 @@ func (d *frameDec) runDecoder(dst []byte, dec *blockDec) ([]byte, error) {
|
||||
break
|
||||
}
|
||||
if d.SingleSegment && uint64(len(d.history.b)) > d.o.maxDecodedSize {
|
||||
println("runDecoder: single segment and", uint64(len(d.history.b)), ">", d.o.maxDecodedSize)
|
||||
err = ErrFrameSizeExceeded
|
||||
break
|
||||
}
|
||||
|
5
vendor/github.com/klauspost/compress/zstd/frameenc.go
generated
vendored
5
vendor/github.com/klauspost/compress/zstd/frameenc.go
generated
vendored
@@ -5,7 +5,6 @@
|
||||
package zstd
|
||||
|
||||
import (
|
||||
"errors"
|
||||
"fmt"
|
||||
"io"
|
||||
"math"
|
||||
@@ -49,9 +48,7 @@ func (f frameHeader) appendTo(dst []byte) ([]byte, error) {
|
||||
windowLog := (bits.Len32(f.WindowSize-1) - winLogMin) << 3
|
||||
dst = append(dst, uint8(windowLog))
|
||||
}
|
||||
if f.SingleSegment && f.ContentSize == 0 {
|
||||
return nil, errors.New("single segment, but no size set")
|
||||
}
|
||||
|
||||
switch fcs {
|
||||
case 0:
|
||||
if f.SingleSegment {
|
||||
|
43
vendor/github.com/klauspost/compress/zstd/fse_encoder.go
generated
vendored
43
vendor/github.com/klauspost/compress/zstd/fse_encoder.go
generated
vendored
@@ -502,21 +502,6 @@ func (s *fseEncoder) validateNorm() (err error) {
|
||||
// writeCount will write the normalized histogram count to header.
|
||||
// This is read back by readNCount.
|
||||
func (s *fseEncoder) writeCount(out []byte) ([]byte, error) {
|
||||
var (
|
||||
tableLog = s.actualTableLog
|
||||
tableSize = 1 << tableLog
|
||||
previous0 bool
|
||||
charnum uint16
|
||||
|
||||
maxHeaderSize = ((int(s.symbolLen) * int(tableLog)) >> 3) + 3
|
||||
|
||||
// Write Table Size
|
||||
bitStream = uint32(tableLog - minEncTablelog)
|
||||
bitCount = uint(4)
|
||||
remaining = int16(tableSize + 1) /* +1 for extra accuracy */
|
||||
threshold = int16(tableSize)
|
||||
nbBits = uint(tableLog + 1)
|
||||
)
|
||||
if s.useRLE {
|
||||
return append(out, s.rleVal), nil
|
||||
}
|
||||
@@ -524,7 +509,28 @@ func (s *fseEncoder) writeCount(out []byte) ([]byte, error) {
|
||||
// Never write predefined.
|
||||
return out, nil
|
||||
}
|
||||
outP := len(out)
|
||||
|
||||
var (
|
||||
tableLog = s.actualTableLog
|
||||
tableSize = 1 << tableLog
|
||||
previous0 bool
|
||||
charnum uint16
|
||||
|
||||
// maximum header size plus 2 extra bytes for final output if bitCount == 0.
|
||||
maxHeaderSize = ((int(s.symbolLen) * int(tableLog)) >> 3) + 3 + 2
|
||||
|
||||
// Write Table Size
|
||||
bitStream = uint32(tableLog - minEncTablelog)
|
||||
bitCount = uint(4)
|
||||
remaining = int16(tableSize + 1) /* +1 for extra accuracy */
|
||||
threshold = int16(tableSize)
|
||||
nbBits = uint(tableLog + 1)
|
||||
outP = len(out)
|
||||
)
|
||||
if cap(out) < outP+maxHeaderSize {
|
||||
out = append(out, make([]byte, maxHeaderSize*3)...)
|
||||
out = out[:len(out)-maxHeaderSize*3]
|
||||
}
|
||||
out = out[:outP+maxHeaderSize]
|
||||
|
||||
// stops at 1
|
||||
@@ -594,11 +600,14 @@ func (s *fseEncoder) writeCount(out []byte) ([]byte, error) {
|
||||
}
|
||||
}
|
||||
|
||||
if outP+2 > len(out) {
|
||||
return nil, fmt.Errorf("internal error: %d > %d, maxheader: %d, sl: %d, tl: %d, normcount: %v", outP+2, len(out), maxHeaderSize, s.symbolLen, int(tableLog), s.norm[:s.symbolLen])
|
||||
}
|
||||
out[outP] = byte(bitStream)
|
||||
out[outP+1] = byte(bitStream >> 8)
|
||||
outP += int((bitCount + 7) / 8)
|
||||
|
||||
if uint16(charnum) > s.symbolLen {
|
||||
if charnum > s.symbolLen {
|
||||
return nil, errors.New("internal error: charnum > s.symbolLen")
|
||||
}
|
||||
return out[:outP], nil
|
||||
|
153
vendor/github.com/klauspost/compress/zstd/fse_predefined.go
generated
vendored
153
vendor/github.com/klauspost/compress/zstd/fse_predefined.go
generated
vendored
@@ -7,6 +7,7 @@ package zstd
|
||||
import (
|
||||
"fmt"
|
||||
"math"
|
||||
"sync"
|
||||
)
|
||||
|
||||
var (
|
||||
@@ -69,85 +70,89 @@ func fillBase(dst []baseOffset, base uint32, bits ...uint8) {
|
||||
}
|
||||
}
|
||||
|
||||
func init() {
|
||||
// Literals length codes
|
||||
tmp := make([]baseOffset, 36)
|
||||
for i := range tmp[:16] {
|
||||
tmp[i] = baseOffset{
|
||||
baseLine: uint32(i),
|
||||
addBits: 0,
|
||||
}
|
||||
}
|
||||
fillBase(tmp[16:], 16, 1, 1, 1, 1, 2, 2, 3, 3, 4, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16)
|
||||
symbolTableX[tableLiteralLengths] = tmp
|
||||
var predef sync.Once
|
||||
|
||||
// Match length codes
|
||||
tmp = make([]baseOffset, 53)
|
||||
for i := range tmp[:32] {
|
||||
tmp[i] = baseOffset{
|
||||
// The transformation adds the 3 length.
|
||||
baseLine: uint32(i) + 3,
|
||||
addBits: 0,
|
||||
func initPredefined() {
|
||||
predef.Do(func() {
|
||||
// Literals length codes
|
||||
tmp := make([]baseOffset, 36)
|
||||
for i := range tmp[:16] {
|
||||
tmp[i] = baseOffset{
|
||||
baseLine: uint32(i),
|
||||
addBits: 0,
|
||||
}
|
||||
}
|
||||
}
|
||||
fillBase(tmp[32:], 35, 1, 1, 1, 1, 2, 2, 3, 3, 4, 4, 5, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16)
|
||||
symbolTableX[tableMatchLengths] = tmp
|
||||
fillBase(tmp[16:], 16, 1, 1, 1, 1, 2, 2, 3, 3, 4, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16)
|
||||
symbolTableX[tableLiteralLengths] = tmp
|
||||
|
||||
// Offset codes
|
||||
tmp = make([]baseOffset, maxOffsetBits+1)
|
||||
tmp[1] = baseOffset{
|
||||
baseLine: 1,
|
||||
addBits: 1,
|
||||
}
|
||||
fillBase(tmp[2:], 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30)
|
||||
symbolTableX[tableOffsets] = tmp
|
||||
// Match length codes
|
||||
tmp = make([]baseOffset, 53)
|
||||
for i := range tmp[:32] {
|
||||
tmp[i] = baseOffset{
|
||||
// The transformation adds the 3 length.
|
||||
baseLine: uint32(i) + 3,
|
||||
addBits: 0,
|
||||
}
|
||||
}
|
||||
fillBase(tmp[32:], 35, 1, 1, 1, 1, 2, 2, 3, 3, 4, 4, 5, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16)
|
||||
symbolTableX[tableMatchLengths] = tmp
|
||||
|
||||
// Fill predefined tables and transform them.
|
||||
// https://github.com/facebook/zstd/blob/dev/doc/zstd_compression_format.md#default-distributions
|
||||
for i := range fsePredef[:] {
|
||||
f := &fsePredef[i]
|
||||
switch tableIndex(i) {
|
||||
case tableLiteralLengths:
|
||||
// https://github.com/facebook/zstd/blob/ededcfca57366461021c922720878c81a5854a0a/lib/decompress/zstd_decompress_block.c#L243
|
||||
f.actualTableLog = 6
|
||||
copy(f.norm[:], []int16{4, 3, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 1, 1, 1,
|
||||
2, 2, 2, 2, 2, 2, 2, 2, 2, 3, 2, 1, 1, 1, 1, 1,
|
||||
-1, -1, -1, -1})
|
||||
f.symbolLen = 36
|
||||
case tableOffsets:
|
||||
// https://github.com/facebook/zstd/blob/ededcfca57366461021c922720878c81a5854a0a/lib/decompress/zstd_decompress_block.c#L281
|
||||
f.actualTableLog = 5
|
||||
copy(f.norm[:], []int16{
|
||||
1, 1, 1, 1, 1, 1, 2, 2, 2, 1, 1, 1, 1, 1, 1, 1,
|
||||
1, 1, 1, 1, 1, 1, 1, 1, -1, -1, -1, -1, -1})
|
||||
f.symbolLen = 29
|
||||
case tableMatchLengths:
|
||||
//https://github.com/facebook/zstd/blob/ededcfca57366461021c922720878c81a5854a0a/lib/decompress/zstd_decompress_block.c#L304
|
||||
f.actualTableLog = 6
|
||||
copy(f.norm[:], []int16{
|
||||
1, 4, 3, 2, 2, 2, 2, 2, 2, 1, 1, 1, 1, 1, 1, 1,
|
||||
1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1,
|
||||
1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, -1, -1,
|
||||
-1, -1, -1, -1, -1})
|
||||
f.symbolLen = 53
|
||||
// Offset codes
|
||||
tmp = make([]baseOffset, maxOffsetBits+1)
|
||||
tmp[1] = baseOffset{
|
||||
baseLine: 1,
|
||||
addBits: 1,
|
||||
}
|
||||
if err := f.buildDtable(); err != nil {
|
||||
panic(fmt.Errorf("building table %v: %v", tableIndex(i), err))
|
||||
}
|
||||
if err := f.transform(symbolTableX[i]); err != nil {
|
||||
panic(fmt.Errorf("building table %v: %v", tableIndex(i), err))
|
||||
}
|
||||
f.preDefined = true
|
||||
fillBase(tmp[2:], 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30)
|
||||
symbolTableX[tableOffsets] = tmp
|
||||
|
||||
// Create encoder as well
|
||||
enc := &fsePredefEnc[i]
|
||||
copy(enc.norm[:], f.norm[:])
|
||||
enc.symbolLen = f.symbolLen
|
||||
enc.actualTableLog = f.actualTableLog
|
||||
if err := enc.buildCTable(); err != nil {
|
||||
panic(fmt.Errorf("building encoding table %v: %v", tableIndex(i), err))
|
||||
// Fill predefined tables and transform them.
|
||||
// https://github.com/facebook/zstd/blob/dev/doc/zstd_compression_format.md#default-distributions
|
||||
for i := range fsePredef[:] {
|
||||
f := &fsePredef[i]
|
||||
switch tableIndex(i) {
|
||||
case tableLiteralLengths:
|
||||
// https://github.com/facebook/zstd/blob/ededcfca57366461021c922720878c81a5854a0a/lib/decompress/zstd_decompress_block.c#L243
|
||||
f.actualTableLog = 6
|
||||
copy(f.norm[:], []int16{4, 3, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 1, 1, 1,
|
||||
2, 2, 2, 2, 2, 2, 2, 2, 2, 3, 2, 1, 1, 1, 1, 1,
|
||||
-1, -1, -1, -1})
|
||||
f.symbolLen = 36
|
||||
case tableOffsets:
|
||||
// https://github.com/facebook/zstd/blob/ededcfca57366461021c922720878c81a5854a0a/lib/decompress/zstd_decompress_block.c#L281
|
||||
f.actualTableLog = 5
|
||||
copy(f.norm[:], []int16{
|
||||
1, 1, 1, 1, 1, 1, 2, 2, 2, 1, 1, 1, 1, 1, 1, 1,
|
||||
1, 1, 1, 1, 1, 1, 1, 1, -1, -1, -1, -1, -1})
|
||||
f.symbolLen = 29
|
||||
case tableMatchLengths:
|
||||
//https://github.com/facebook/zstd/blob/ededcfca57366461021c922720878c81a5854a0a/lib/decompress/zstd_decompress_block.c#L304
|
||||
f.actualTableLog = 6
|
||||
copy(f.norm[:], []int16{
|
||||
1, 4, 3, 2, 2, 2, 2, 2, 2, 1, 1, 1, 1, 1, 1, 1,
|
||||
1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1,
|
||||
1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, -1, -1,
|
||||
-1, -1, -1, -1, -1})
|
||||
f.symbolLen = 53
|
||||
}
|
||||
if err := f.buildDtable(); err != nil {
|
||||
panic(fmt.Errorf("building table %v: %v", tableIndex(i), err))
|
||||
}
|
||||
if err := f.transform(symbolTableX[i]); err != nil {
|
||||
panic(fmt.Errorf("building table %v: %v", tableIndex(i), err))
|
||||
}
|
||||
f.preDefined = true
|
||||
|
||||
// Create encoder as well
|
||||
enc := &fsePredefEnc[i]
|
||||
copy(enc.norm[:], f.norm[:])
|
||||
enc.symbolLen = f.symbolLen
|
||||
enc.actualTableLog = f.actualTableLog
|
||||
if err := enc.buildCTable(); err != nil {
|
||||
panic(fmt.Errorf("building encoding table %v: %v", tableIndex(i), err))
|
||||
}
|
||||
enc.setBits(bitTables[i])
|
||||
enc.preDefined = true
|
||||
}
|
||||
enc.setBits(bitTables[i])
|
||||
enc.preDefined = true
|
||||
}
|
||||
})
|
||||
}
|
||||
|
1
vendor/github.com/klauspost/compress/zstd/snappy.go
generated
vendored
1
vendor/github.com/klauspost/compress/zstd/snappy.go
generated
vendored
@@ -80,6 +80,7 @@ type SnappyConverter struct {
|
||||
// If any error is detected on the Snappy stream it is returned.
|
||||
// The number of bytes written is returned.
|
||||
func (r *SnappyConverter) Convert(in io.Reader, w io.Writer) (int64, error) {
|
||||
initPredefined()
|
||||
r.err = nil
|
||||
r.r = in
|
||||
if r.block == nil {
|
||||
|
1
vendor/github.com/klauspost/compress/zstd/zstd.go
generated
vendored
1
vendor/github.com/klauspost/compress/zstd/zstd.go
generated
vendored
@@ -11,6 +11,7 @@ import (
|
||||
|
||||
const debug = false
|
||||
const debugSequences = false
|
||||
const debugMatches = false
|
||||
|
||||
// force encoder to use predefined tables.
|
||||
const forcePreDef = false
|
||||
|
Reference in New Issue
Block a user