2014-08-14 22:42:41 +02:00
|
|
|
|
// Copyright (c) 2014 Couchbase, Inc.
|
|
|
|
|
// Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file
|
|
|
|
|
// except in compliance with the License. You may obtain a copy of the License at
|
|
|
|
|
// http://www.apache.org/licenses/LICENSE-2.0
|
|
|
|
|
// Unless required by applicable law or agreed to in writing, software distributed under the
|
|
|
|
|
// License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND,
|
|
|
|
|
// either express or implied. See the License for the specific language governing permissions
|
|
|
|
|
// and limitations under the License.
|
2014-08-25 18:06:10 +02:00
|
|
|
|
|
|
|
|
|
// +build libstemmer full
|
2014-08-25 18:22:01 +02:00
|
|
|
|
// +build icu full
|
2014-08-25 18:06:10 +02:00
|
|
|
|
|
2014-08-14 22:42:41 +02:00
|
|
|
|
package tr
|
|
|
|
|
|
|
|
|
|
import (
|
|
|
|
|
"reflect"
|
|
|
|
|
"testing"
|
|
|
|
|
|
2014-08-28 21:38:57 +02:00
|
|
|
|
"github.com/blevesearch/bleve/analysis"
|
|
|
|
|
"github.com/blevesearch/bleve/registry"
|
2014-08-14 22:42:41 +02:00
|
|
|
|
)
|
|
|
|
|
|
|
|
|
|
func TestTurkishAnalyzer(t *testing.T) {
|
|
|
|
|
tests := []struct {
|
|
|
|
|
input []byte
|
|
|
|
|
output analysis.TokenStream
|
|
|
|
|
}{
|
|
|
|
|
// stemming
|
|
|
|
|
{
|
|
|
|
|
input: []byte("ağacı"),
|
|
|
|
|
output: analysis.TokenStream{
|
|
|
|
|
&analysis.Token{
|
|
|
|
|
Term: []byte("ağaç"),
|
|
|
|
|
},
|
|
|
|
|
},
|
|
|
|
|
},
|
|
|
|
|
{
|
|
|
|
|
input: []byte("ağaç"),
|
|
|
|
|
output: analysis.TokenStream{
|
|
|
|
|
&analysis.Token{
|
|
|
|
|
Term: []byte("ağaç"),
|
|
|
|
|
},
|
|
|
|
|
},
|
|
|
|
|
},
|
|
|
|
|
// stop word
|
|
|
|
|
{
|
|
|
|
|
input: []byte("dolayı"),
|
|
|
|
|
output: analysis.TokenStream{},
|
|
|
|
|
},
|
|
|
|
|
// apostrophes
|
|
|
|
|
{
|
|
|
|
|
input: []byte("Kıbrıs'ta"),
|
|
|
|
|
output: analysis.TokenStream{
|
|
|
|
|
&analysis.Token{
|
|
|
|
|
Term: []byte("kıbrıs"),
|
|
|
|
|
},
|
|
|
|
|
},
|
|
|
|
|
},
|
|
|
|
|
{
|
|
|
|
|
input: []byte("Van Gölü'ne"),
|
|
|
|
|
output: analysis.TokenStream{
|
|
|
|
|
&analysis.Token{
|
|
|
|
|
Term: []byte("van"),
|
|
|
|
|
},
|
|
|
|
|
&analysis.Token{
|
|
|
|
|
Term: []byte("göl"),
|
|
|
|
|
},
|
|
|
|
|
},
|
|
|
|
|
},
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
cache := registry.NewCache()
|
|
|
|
|
analyzer, err := cache.AnalyzerNamed(AnalyzerName)
|
|
|
|
|
if err != nil {
|
|
|
|
|
t.Fatal(err)
|
|
|
|
|
}
|
|
|
|
|
for _, test := range tests {
|
|
|
|
|
actual := analyzer.Analyze(test.input)
|
|
|
|
|
if len(actual) != len(test.output) {
|
|
|
|
|
t.Fatalf("expected length: %d, got %d", len(test.output), len(actual))
|
|
|
|
|
}
|
|
|
|
|
for i, tok := range actual {
|
|
|
|
|
if !reflect.DeepEqual(tok.Term, test.output[i].Term) {
|
|
|
|
|
t.Errorf("expected term %s (% x) got %s (% x)", test.output[i].Term, test.output[i].Term, tok.Term, tok.Term)
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
}
|