mirror of
				https://github.com/go-gitea/gitea
				synced 2025-10-27 17:38:25 +00:00 
			
		
		
		
	
		
			
				
	
	
		
			253 lines
		
	
	
		
			5.9 KiB
		
	
	
	
		
			Go
		
	
	
	
	
	
			
		
		
	
	
			253 lines
		
	
	
		
			5.9 KiB
		
	
	
	
		
			Go
		
	
	
	
	
	
| package flate
 | |
| 
 | |
| import (
 | |
| 	"io"
 | |
| 	"math"
 | |
| )
 | |
| 
 | |
| const (
 | |
| 	maxStatelessBlock = math.MaxInt16
 | |
| 
 | |
| 	slTableBits  = 13
 | |
| 	slTableSize  = 1 << slTableBits
 | |
| 	slTableShift = 32 - slTableBits
 | |
| )
 | |
| 
 | |
| type statelessWriter struct {
 | |
| 	dst    io.Writer
 | |
| 	closed bool
 | |
| }
 | |
| 
 | |
| func (s *statelessWriter) Close() error {
 | |
| 	if s.closed {
 | |
| 		return nil
 | |
| 	}
 | |
| 	s.closed = true
 | |
| 	// Emit EOF block
 | |
| 	return StatelessDeflate(s.dst, nil, true)
 | |
| }
 | |
| 
 | |
| func (s *statelessWriter) Write(p []byte) (n int, err error) {
 | |
| 	err = StatelessDeflate(s.dst, p, false)
 | |
| 	if err != nil {
 | |
| 		return 0, err
 | |
| 	}
 | |
| 	return len(p), nil
 | |
| }
 | |
| 
 | |
| func (s *statelessWriter) Reset(w io.Writer) {
 | |
| 	s.dst = w
 | |
| 	s.closed = false
 | |
| }
 | |
| 
 | |
| // NewStatelessWriter will do compression but without maintaining any state
 | |
| // between Write calls.
 | |
| // There will be no memory kept between Write calls,
 | |
| // but compression and speed will be suboptimal.
 | |
| // Because of this, the size of actual Write calls will affect output size.
 | |
| func NewStatelessWriter(dst io.Writer) io.WriteCloser {
 | |
| 	return &statelessWriter{dst: dst}
 | |
| }
 | |
| 
 | |
| // StatelessDeflate allows to compress directly to a Writer without retaining state.
 | |
| // When returning everything will be flushed.
 | |
| func StatelessDeflate(out io.Writer, in []byte, eof bool) error {
 | |
| 	var dst tokens
 | |
| 	bw := newHuffmanBitWriter(out)
 | |
| 	if eof && len(in) == 0 {
 | |
| 		// Just write an EOF block.
 | |
| 		// Could be faster...
 | |
| 		bw.writeStoredHeader(0, true)
 | |
| 		bw.flush()
 | |
| 		return bw.err
 | |
| 	}
 | |
| 
 | |
| 	for len(in) > 0 {
 | |
| 		todo := in
 | |
| 		if len(todo) > maxStatelessBlock {
 | |
| 			todo = todo[:maxStatelessBlock]
 | |
| 		}
 | |
| 		in = in[len(todo):]
 | |
| 		// Compress
 | |
| 		statelessEnc(&dst, todo)
 | |
| 		isEof := eof && len(in) == 0
 | |
| 
 | |
| 		if dst.n == 0 {
 | |
| 			bw.writeStoredHeader(len(todo), isEof)
 | |
| 			if bw.err != nil {
 | |
| 				return bw.err
 | |
| 			}
 | |
| 			bw.writeBytes(todo)
 | |
| 		} else if int(dst.n) > len(todo)-len(todo)>>4 {
 | |
| 			// If we removed less than 1/16th, huffman compress the block.
 | |
| 			bw.writeBlockHuff(isEof, todo, false)
 | |
| 		} else {
 | |
| 			bw.writeBlockDynamic(&dst, isEof, todo, false)
 | |
| 		}
 | |
| 		if bw.err != nil {
 | |
| 			return bw.err
 | |
| 		}
 | |
| 		dst.Reset()
 | |
| 	}
 | |
| 	if !eof {
 | |
| 		// Align.
 | |
| 		bw.writeStoredHeader(0, false)
 | |
| 	}
 | |
| 	bw.flush()
 | |
| 	return bw.err
 | |
| }
 | |
| 
 | |
| func hashSL(u uint32) uint32 {
 | |
| 	return (u * 0x1e35a7bd) >> slTableShift
 | |
| }
 | |
| 
 | |
| func load3216(b []byte, i int16) uint32 {
 | |
| 	// Help the compiler eliminate bounds checks on the read so it can be done in a single read.
 | |
| 	b = b[i:]
 | |
| 	b = b[:4]
 | |
| 	return uint32(b[0]) | uint32(b[1])<<8 | uint32(b[2])<<16 | uint32(b[3])<<24
 | |
| }
 | |
| 
 | |
| func load6416(b []byte, i int16) uint64 {
 | |
| 	// Help the compiler eliminate bounds checks on the read so it can be done in a single read.
 | |
| 	b = b[i:]
 | |
| 	b = b[:8]
 | |
| 	return uint64(b[0]) | uint64(b[1])<<8 | uint64(b[2])<<16 | uint64(b[3])<<24 |
 | |
| 		uint64(b[4])<<32 | uint64(b[5])<<40 | uint64(b[6])<<48 | uint64(b[7])<<56
 | |
| }
 | |
| 
 | |
| func statelessEnc(dst *tokens, src []byte) {
 | |
| 	const (
 | |
| 		inputMargin            = 12 - 1
 | |
| 		minNonLiteralBlockSize = 1 + 1 + inputMargin
 | |
| 	)
 | |
| 
 | |
| 	type tableEntry struct {
 | |
| 		offset int16
 | |
| 	}
 | |
| 
 | |
| 	var table [slTableSize]tableEntry
 | |
| 
 | |
| 	// This check isn't in the Snappy implementation, but there, the caller
 | |
| 	// instead of the callee handles this case.
 | |
| 	if len(src) < minNonLiteralBlockSize {
 | |
| 		// We do not fill the token table.
 | |
| 		// This will be picked up by caller.
 | |
| 		dst.n = uint16(len(src))
 | |
| 		return
 | |
| 	}
 | |
| 
 | |
| 	s := int16(1)
 | |
| 	nextEmit := int16(0)
 | |
| 	// sLimit is when to stop looking for offset/length copies. The inputMargin
 | |
| 	// lets us use a fast path for emitLiteral in the main loop, while we are
 | |
| 	// looking for copies.
 | |
| 	sLimit := int16(len(src) - inputMargin)
 | |
| 
 | |
| 	// nextEmit is where in src the next emitLiteral should start from.
 | |
| 	cv := load3216(src, s)
 | |
| 
 | |
| 	for {
 | |
| 		const skipLog = 5
 | |
| 		const doEvery = 2
 | |
| 
 | |
| 		nextS := s
 | |
| 		var candidate tableEntry
 | |
| 		for {
 | |
| 			nextHash := hashSL(cv)
 | |
| 			candidate = table[nextHash]
 | |
| 			nextS = s + doEvery + (s-nextEmit)>>skipLog
 | |
| 			if nextS > sLimit || nextS <= 0 {
 | |
| 				goto emitRemainder
 | |
| 			}
 | |
| 
 | |
| 			now := load6416(src, nextS)
 | |
| 			table[nextHash] = tableEntry{offset: s}
 | |
| 			nextHash = hashSL(uint32(now))
 | |
| 
 | |
| 			if cv == load3216(src, candidate.offset) {
 | |
| 				table[nextHash] = tableEntry{offset: nextS}
 | |
| 				break
 | |
| 			}
 | |
| 
 | |
| 			// Do one right away...
 | |
| 			cv = uint32(now)
 | |
| 			s = nextS
 | |
| 			nextS++
 | |
| 			candidate = table[nextHash]
 | |
| 			now >>= 8
 | |
| 			table[nextHash] = tableEntry{offset: s}
 | |
| 
 | |
| 			if cv == load3216(src, candidate.offset) {
 | |
| 				table[nextHash] = tableEntry{offset: nextS}
 | |
| 				break
 | |
| 			}
 | |
| 			cv = uint32(now)
 | |
| 			s = nextS
 | |
| 		}
 | |
| 
 | |
| 		// A 4-byte match has been found. We'll later see if more than 4 bytes
 | |
| 		// match. But, prior to the match, src[nextEmit:s] are unmatched. Emit
 | |
| 		// them as literal bytes.
 | |
| 		for {
 | |
| 			// Invariant: we have a 4-byte match at s, and no need to emit any
 | |
| 			// literal bytes prior to s.
 | |
| 
 | |
| 			// Extend the 4-byte match as long as possible.
 | |
| 			t := candidate.offset
 | |
| 			l := int16(matchLen(src[s+4:], src[t+4:]) + 4)
 | |
| 
 | |
| 			// Extend backwards
 | |
| 			for t > 0 && s > nextEmit && src[t-1] == src[s-1] {
 | |
| 				s--
 | |
| 				t--
 | |
| 				l++
 | |
| 			}
 | |
| 			if nextEmit < s {
 | |
| 				emitLiteral(dst, src[nextEmit:s])
 | |
| 			}
 | |
| 
 | |
| 			// Save the match found
 | |
| 			dst.AddMatchLong(int32(l), uint32(s-t-baseMatchOffset))
 | |
| 			s += l
 | |
| 			nextEmit = s
 | |
| 			if nextS >= s {
 | |
| 				s = nextS + 1
 | |
| 			}
 | |
| 			if s >= sLimit {
 | |
| 				goto emitRemainder
 | |
| 			}
 | |
| 
 | |
| 			// We could immediately start working at s now, but to improve
 | |
| 			// compression we first update the hash table at s-2 and at s. If
 | |
| 			// another emitCopy is not our next move, also calculate nextHash
 | |
| 			// at s+1. At least on GOARCH=amd64, these three hash calculations
 | |
| 			// are faster as one load64 call (with some shifts) instead of
 | |
| 			// three load32 calls.
 | |
| 			x := load6416(src, s-2)
 | |
| 			o := s - 2
 | |
| 			prevHash := hashSL(uint32(x))
 | |
| 			table[prevHash] = tableEntry{offset: o}
 | |
| 			x >>= 16
 | |
| 			currHash := hashSL(uint32(x))
 | |
| 			candidate = table[currHash]
 | |
| 			table[currHash] = tableEntry{offset: o + 2}
 | |
| 
 | |
| 			if uint32(x) != load3216(src, candidate.offset) {
 | |
| 				cv = uint32(x >> 8)
 | |
| 				s++
 | |
| 				break
 | |
| 			}
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| emitRemainder:
 | |
| 	if int(nextEmit) < len(src) {
 | |
| 		// If nothing was added, don't encode literals.
 | |
| 		if dst.n == 0 {
 | |
| 			return
 | |
| 		}
 | |
| 		emitLiteral(dst, src[nextEmit:])
 | |
| 	}
 | |
| }
 |