techknowlogick 3f5cdfe359 use go 1.13 (#8088) | 5 lat temu | |
---|---|---|
.. | ||
.gitignore | 5 lat temu | |
.travis.yml | 5 lat temu | |
LICENSE | 7 lat temu | |
README.md | 5 lat temu | |
doc.go | 7 lat temu | |
segment.go | 7 lat temu | |
segment_fuzz.go | 7 lat temu | |
segment_words.go | 7 lat temu | |
segment_words.rl | 5 lat temu | |
segment_words_prod.go | 7 lat temu |
A Go library for performing Unicode Text Segmentation as described in Unicode Standard Annex #29
Apache License Version 2.0
The functionality is exposed in two ways:
You can use a bufio.Scanner with the SplitWords implementation of SplitFunc. The SplitWords function will identify the appropriate word boundaries in the input text and the Scanner will return tokens at the appropriate place.
scanner := bufio.NewScanner(...)
scanner.Split(segment.SplitWords)
for scanner.Scan() {
tokenBytes := scanner.Bytes()
}
if err := scanner.Err(); err != nil {
t.Fatal(err)
}
Sometimes you would also like information returned about the type of token. To do this we have introduce a new type named Segmenter. It works just like Scanner but additionally a token type is returned.
segmenter := segment.NewWordSegmenter(...)
for segmenter.Segment() {
tokenBytes := segmenter.Bytes())
tokenType := segmenter.Type()
}
if err := segmenter.Err(); err != nil {
t.Fatal(err)
}
By default segment does NOT use the fastest runtime implementation. The reason is that it adds approximately 5s to compilation time and may require more than 1GB of ram on the machine performing compilation.
However, you can choose to build with the fastest runtime implementation by passing the build tag as follows:
-tags 'prod'
Several components in this package are generated.
All of these can be generated by running:
go generate
There is support for fuzzing the segment library with go-fuzz.
Install go-fuzz if you haven’t already:
go get github.com/dvyukov/go-fuzz/go-fuzz
go get github.com/dvyukov/go-fuzz/go-fuzz-build
Build the package with go-fuzz:
go-fuzz-build github.com/blevesearch/segment
Convert the Unicode provided test cases into the initial corpus for go-fuzz:
go test -v -run=TestGenerateWordSegmentFuzz -tags gofuzz_generate
Run go-fuzz:
go-fuzz -bin=segment-fuzz.zip -workdir=workdir