mirror of
https://github.com/go-gitea/gitea
synced 2024-11-15 22:54:24 +00:00
285 lines
9.8 KiB
Go
285 lines
9.8 KiB
Go
// Copyright (c) 2015 Couchbase, Inc.
|
|
// Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file
|
|
// except in compliance with the License. You may obtain a copy of the License at
|
|
// http://www.apache.org/licenses/LICENSE-2.0
|
|
// Unless required by applicable law or agreed to in writing, software distributed under the
|
|
// License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND,
|
|
// either express or implied. See the License for the specific language governing permissions
|
|
// and limitations under the License.
|
|
|
|
package segment
|
|
|
|
import (
|
|
"errors"
|
|
"io"
|
|
)
|
|
|
|
// Autogenerate the following:
|
|
// 1. Ragel rules from subset of Unicode script properties
|
|
// 2. Ragel rules from Unicode word segmentation properties
|
|
// 3. Ragel machine for word segmentation
|
|
// 4. Test tables from Unicode
|
|
//
|
|
// Requires:
|
|
// 1. Ruby (to generate ragel rules from unicode spec)
|
|
// 2. Ragel (only v6.9 tested)
|
|
// 3. sed (to rewrite build tags)
|
|
//
|
|
//go:generate ragel/unicode2ragel.rb -u http://www.unicode.org/Public/8.0.0/ucd/Scripts.txt -m SCRIPTS -p Hangul,Han,Hiragana -o ragel/uscript.rl
|
|
//go:generate ragel/unicode2ragel.rb -u http://www.unicode.org/Public/8.0.0/ucd/auxiliary/WordBreakProperty.txt -m WB -p Double_Quote,Single_Quote,Hebrew_Letter,CR,LF,Newline,Extend,Format,Katakana,ALetter,MidLetter,MidNum,MidNumLet,Numeric,ExtendNumLet,Regional_Indicator -o ragel/uwb.rl
|
|
//go:generate ragel -T1 -Z segment_words.rl -o segment_words.go
|
|
//go:generate sed -i "" -e "s/BUILDTAGS/!prod/" segment_words.go
|
|
//go:generate sed -i "" -e "s/RAGELFLAGS/-T1/" segment_words.go
|
|
//go:generate ragel -G2 -Z segment_words.rl -o segment_words_prod.go
|
|
//go:generate sed -i "" -e "s/BUILDTAGS/prod/" segment_words_prod.go
|
|
//go:generate sed -i "" -e "s/RAGELFLAGS/-G2/" segment_words_prod.go
|
|
//go:generate go run maketesttables.go -output tables_test.go
|
|
|
|
// NewWordSegmenter returns a new Segmenter to read from r.
|
|
func NewWordSegmenter(r io.Reader) *Segmenter {
|
|
return NewSegmenter(r)
|
|
}
|
|
|
|
// NewWordSegmenterDirect returns a new Segmenter to work directly with buf.
|
|
func NewWordSegmenterDirect(buf []byte) *Segmenter {
|
|
return NewSegmenterDirect(buf)
|
|
}
|
|
|
|
func SplitWords(data []byte, atEOF bool) (int, []byte, error) {
|
|
advance, token, _, err := SegmentWords(data, atEOF)
|
|
return advance, token, err
|
|
}
|
|
|
|
func SegmentWords(data []byte, atEOF bool) (int, []byte, int, error) {
|
|
vals := make([][]byte, 0, 1)
|
|
types := make([]int, 0, 1)
|
|
tokens, types, advance, err := segmentWords(data, 1, atEOF, vals, types)
|
|
if len(tokens) > 0 {
|
|
return advance, tokens[0], types[0], err
|
|
}
|
|
return advance, nil, 0, err
|
|
}
|
|
|
|
func SegmentWordsDirect(data []byte, val [][]byte, types []int) ([][]byte, []int, int, error) {
|
|
return segmentWords(data, -1, true, val, types)
|
|
}
|
|
|
|
// *** Core Segmenter
|
|
|
|
const maxConsecutiveEmptyReads = 100
|
|
|
|
// NewSegmenter returns a new Segmenter to read from r.
|
|
// Defaults to segment using SegmentWords
|
|
func NewSegmenter(r io.Reader) *Segmenter {
|
|
return &Segmenter{
|
|
r: r,
|
|
segment: SegmentWords,
|
|
maxTokenSize: MaxScanTokenSize,
|
|
buf: make([]byte, 4096), // Plausible starting size; needn't be large.
|
|
}
|
|
}
|
|
|
|
// NewSegmenterDirect returns a new Segmenter to work directly with buf.
|
|
// Defaults to segment using SegmentWords
|
|
func NewSegmenterDirect(buf []byte) *Segmenter {
|
|
return &Segmenter{
|
|
segment: SegmentWords,
|
|
maxTokenSize: MaxScanTokenSize,
|
|
buf: buf,
|
|
start: 0,
|
|
end: len(buf),
|
|
err: io.EOF,
|
|
}
|
|
}
|
|
|
|
// Segmenter provides a convenient interface for reading data such as
|
|
// a file of newline-delimited lines of text. Successive calls to
|
|
// the Segment method will step through the 'tokens' of a file, skipping
|
|
// the bytes between the tokens. The specification of a token is
|
|
// defined by a split function of type SplitFunc; the default split
|
|
// function breaks the input into lines with line termination stripped. Split
|
|
// functions are defined in this package for scanning a file into
|
|
// lines, bytes, UTF-8-encoded runes, and space-delimited words. The
|
|
// client may instead provide a custom split function.
|
|
//
|
|
// Segmenting stops unrecoverably at EOF, the first I/O error, or a token too
|
|
// large to fit in the buffer. When a scan stops, the reader may have
|
|
// advanced arbitrarily far past the last token. Programs that need more
|
|
// control over error handling or large tokens, or must run sequential scans
|
|
// on a reader, should use bufio.Reader instead.
|
|
//
|
|
type Segmenter struct {
|
|
r io.Reader // The reader provided by the client.
|
|
segment SegmentFunc // The function to split the tokens.
|
|
maxTokenSize int // Maximum size of a token; modified by tests.
|
|
token []byte // Last token returned by split.
|
|
buf []byte // Buffer used as argument to split.
|
|
start int // First non-processed byte in buf.
|
|
end int // End of data in buf.
|
|
typ int // The token type
|
|
err error // Sticky error.
|
|
}
|
|
|
|
// SegmentFunc is the signature of the segmenting function used to tokenize the
|
|
// input. The arguments are an initial substring of the remaining unprocessed
|
|
// data and a flag, atEOF, that reports whether the Reader has no more data
|
|
// to give. The return values are the number of bytes to advance the input
|
|
// and the next token to return to the user, plus an error, if any. If the
|
|
// data does not yet hold a complete token, for instance if it has no newline
|
|
// while scanning lines, SegmentFunc can return (0, nil, nil) to signal the
|
|
// Segmenter to read more data into the slice and try again with a longer slice
|
|
// starting at the same point in the input.
|
|
//
|
|
// If the returned error is non-nil, segmenting stops and the error
|
|
// is returned to the client.
|
|
//
|
|
// The function is never called with an empty data slice unless atEOF
|
|
// is true. If atEOF is true, however, data may be non-empty and,
|
|
// as always, holds unprocessed text.
|
|
type SegmentFunc func(data []byte, atEOF bool) (advance int, token []byte, segmentType int, err error)
|
|
|
|
// Errors returned by Segmenter.
|
|
var (
|
|
ErrTooLong = errors.New("bufio.Segmenter: token too long")
|
|
ErrNegativeAdvance = errors.New("bufio.Segmenter: SplitFunc returns negative advance count")
|
|
ErrAdvanceTooFar = errors.New("bufio.Segmenter: SplitFunc returns advance count beyond input")
|
|
)
|
|
|
|
const (
|
|
// Maximum size used to buffer a token. The actual maximum token size
|
|
// may be smaller as the buffer may need to include, for instance, a newline.
|
|
MaxScanTokenSize = 64 * 1024
|
|
)
|
|
|
|
// Err returns the first non-EOF error that was encountered by the Segmenter.
|
|
func (s *Segmenter) Err() error {
|
|
if s.err == io.EOF {
|
|
return nil
|
|
}
|
|
return s.err
|
|
}
|
|
|
|
func (s *Segmenter) Type() int {
|
|
return s.typ
|
|
}
|
|
|
|
// Bytes returns the most recent token generated by a call to Segment.
|
|
// The underlying array may point to data that will be overwritten
|
|
// by a subsequent call to Segment. It does no allocation.
|
|
func (s *Segmenter) Bytes() []byte {
|
|
return s.token
|
|
}
|
|
|
|
// Text returns the most recent token generated by a call to Segment
|
|
// as a newly allocated string holding its bytes.
|
|
func (s *Segmenter) Text() string {
|
|
return string(s.token)
|
|
}
|
|
|
|
// Segment advances the Segmenter to the next token, which will then be
|
|
// available through the Bytes or Text method. It returns false when the
|
|
// scan stops, either by reaching the end of the input or an error.
|
|
// After Segment returns false, the Err method will return any error that
|
|
// occurred during scanning, except that if it was io.EOF, Err
|
|
// will return nil.
|
|
func (s *Segmenter) Segment() bool {
|
|
// Loop until we have a token.
|
|
for {
|
|
// See if we can get a token with what we already have.
|
|
if s.end > s.start {
|
|
advance, token, typ, err := s.segment(s.buf[s.start:s.end], s.err != nil)
|
|
if err != nil {
|
|
s.setErr(err)
|
|
return false
|
|
}
|
|
s.typ = typ
|
|
if !s.advance(advance) {
|
|
return false
|
|
}
|
|
s.token = token
|
|
if token != nil {
|
|
return true
|
|
}
|
|
}
|
|
// We cannot generate a token with what we are holding.
|
|
// If we've already hit EOF or an I/O error, we are done.
|
|
if s.err != nil {
|
|
// Shut it down.
|
|
s.start = 0
|
|
s.end = 0
|
|
return false
|
|
}
|
|
// Must read more data.
|
|
// First, shift data to beginning of buffer if there's lots of empty space
|
|
// or space is needed.
|
|
if s.start > 0 && (s.end == len(s.buf) || s.start > len(s.buf)/2) {
|
|
copy(s.buf, s.buf[s.start:s.end])
|
|
s.end -= s.start
|
|
s.start = 0
|
|
}
|
|
// Is the buffer full? If so, resize.
|
|
if s.end == len(s.buf) {
|
|
if len(s.buf) >= s.maxTokenSize {
|
|
s.setErr(ErrTooLong)
|
|
return false
|
|
}
|
|
newSize := len(s.buf) * 2
|
|
if newSize > s.maxTokenSize {
|
|
newSize = s.maxTokenSize
|
|
}
|
|
newBuf := make([]byte, newSize)
|
|
copy(newBuf, s.buf[s.start:s.end])
|
|
s.buf = newBuf
|
|
s.end -= s.start
|
|
s.start = 0
|
|
continue
|
|
}
|
|
// Finally we can read some input. Make sure we don't get stuck with
|
|
// a misbehaving Reader. Officially we don't need to do this, but let's
|
|
// be extra careful: Segmenter is for safe, simple jobs.
|
|
for loop := 0; ; {
|
|
n, err := s.r.Read(s.buf[s.end:len(s.buf)])
|
|
s.end += n
|
|
if err != nil {
|
|
s.setErr(err)
|
|
break
|
|
}
|
|
if n > 0 {
|
|
break
|
|
}
|
|
loop++
|
|
if loop > maxConsecutiveEmptyReads {
|
|
s.setErr(io.ErrNoProgress)
|
|
break
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
// advance consumes n bytes of the buffer. It reports whether the advance was legal.
|
|
func (s *Segmenter) advance(n int) bool {
|
|
if n < 0 {
|
|
s.setErr(ErrNegativeAdvance)
|
|
return false
|
|
}
|
|
if n > s.end-s.start {
|
|
s.setErr(ErrAdvanceTooFar)
|
|
return false
|
|
}
|
|
s.start += n
|
|
return true
|
|
}
|
|
|
|
// setErr records the first error encountered.
|
|
func (s *Segmenter) setErr(err error) {
|
|
if s.err == nil || s.err == io.EOF {
|
|
s.err = err
|
|
}
|
|
}
|
|
|
|
// SetSegmenter sets the segment function for the Segmenter. If called, it must be
|
|
// called before Segment.
|
|
func (s *Segmenter) SetSegmenter(segmenter SegmentFunc) {
|
|
s.segment = segmenter
|
|
}
|