0
0
bleve/analysis/tokenizers/regexp_tokenizer/regexp_tokenizer.go

78 lines
2.1 KiB
Go
Raw Normal View History

2014-04-17 22:55:53 +02:00
// Copyright (c) 2014 Couchbase, Inc.
// Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file
// except in compliance with the License. You may obtain a copy of the License at
// http://www.apache.org/licenses/LICENSE-2.0
// Unless required by applicable law or agreed to in writing, software distributed under the
// License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND,
// either express or implied. See the License for the specific language governing permissions
// and limitations under the License.
2014-04-17 22:55:53 +02:00
package regexp_tokenizer
import (
"fmt"
2014-04-17 22:55:53 +02:00
"regexp"
"strconv"
2014-04-17 22:55:53 +02:00
"github.com/blevesearch/bleve/analysis"
"github.com/blevesearch/bleve/registry"
2014-04-17 22:55:53 +02:00
)
const Name = "regexp"
var IdeographRegexp = regexp.MustCompile(`\p{Han}|\p{Hangul}|\p{Hiragana}|\p{Katakana}`)
2014-04-17 22:55:53 +02:00
type RegexpTokenizer struct {
r *regexp.Regexp
}
func NewRegexpTokenizer(r *regexp.Regexp) *RegexpTokenizer {
return &RegexpTokenizer{
r: r,
}
}
func (rt *RegexpTokenizer) Tokenize(input []byte) analysis.TokenStream {
matches := rt.r.FindAllIndex(input, -1)
rv := make(analysis.TokenStream, len(matches))
for i, match := range matches {
matchBytes := input[match[0]:match[1]]
2014-04-17 22:55:53 +02:00
token := analysis.Token{
Term: matchBytes,
2014-04-17 22:55:53 +02:00
Start: match[0],
End: match[1],
Position: i + 1,
Type: detectTokenType(matchBytes),
2014-04-17 22:55:53 +02:00
}
rv[i] = &token
}
return rv
}
func RegexpTokenizerConstructor(config map[string]interface{}, cache *registry.Cache) (analysis.Tokenizer, error) {
rval, ok := config["regexp"].(string)
if !ok {
return nil, fmt.Errorf("must specify regexp")
}
r, err := regexp.Compile(rval)
if err != nil {
return nil, fmt.Errorf("unable to build regexp tokenizer: %v", err)
}
return NewRegexpTokenizer(r), nil
}
func init() {
registry.RegisterTokenizer(Name, RegexpTokenizerConstructor)
}
func detectTokenType(termBytes []byte) analysis.TokenType {
if IdeographRegexp.Match(termBytes) {
return analysis.Ideographic
}
_, err := strconv.ParseFloat(string(termBytes), 64)
if err == nil {
return analysis.Numeric
}
return analysis.AlphaNumeric
}