0
0
bleve/analysis/token_filters/elision_filter/elision_filter_test.go
Marty Schoch 2c0bf23fac added elision filter
defined article word maps for french, italian, irish and catalan
defined elision filters for these same languages
updated analyers for french and italian to use this new filter
irish and catalan still depend on other missing pieces
closes #25
2014-08-03 19:17:35 -04:00

117 lines
2.6 KiB
Go

// Copyright (c) 2014 Couchbase, Inc.
// Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file
// except in compliance with the License. You may obtain a copy of the License at
// http://www.apache.org/licenses/LICENSE-2.0
// Unless required by applicable law or agreed to in writing, software distributed under the
// License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND,
// either express or implied. See the License for the specific language governing permissions
// and limitations under the License.
package elision_filter
import (
"reflect"
"testing"
"github.com/couchbaselabs/bleve/analysis"
)
func TestElisionFilter(t *testing.T) {
frenchArticlesMap := analysis.NewWordMap()
err := frenchArticlesMap.LoadBytes(FrenchArticles)
if err != nil {
t.Fatal(err)
}
italianArticlesMap := analysis.NewWordMap()
err = italianArticlesMap.LoadBytes(ItalianArticles)
if err != nil {
t.Fatal(err)
}
catalanArticlesMap := analysis.NewWordMap()
err = catalanArticlesMap.LoadBytes(CatalanArticles)
if err != nil {
t.Fatal(err)
}
irishArticlesMap := analysis.NewWordMap()
err = irishArticlesMap.LoadBytes(IrishArticles)
if err != nil {
t.Fatal(err)
}
tests := []struct {
articleMap analysis.WordMap
input analysis.TokenStream
output analysis.TokenStream
}{
{
articleMap: frenchArticlesMap,
input: analysis.TokenStream{
&analysis.Token{
Term: []byte("l'avion"),
},
},
output: analysis.TokenStream{
&analysis.Token{
Term: []byte("avion"),
},
},
},
{
articleMap: italianArticlesMap,
input: analysis.TokenStream{
&analysis.Token{
Term: []byte("dell'Italia"),
},
},
output: analysis.TokenStream{
&analysis.Token{
Term: []byte("Italia"),
},
},
},
{
articleMap: catalanArticlesMap,
input: analysis.TokenStream{
&analysis.Token{
Term: []byte("l'Institut"),
},
&analysis.Token{
Term: []byte("d'Estudis"),
},
},
output: analysis.TokenStream{
&analysis.Token{
Term: []byte("Institut"),
},
&analysis.Token{
Term: []byte("Estudis"),
},
},
},
{
articleMap: irishArticlesMap,
input: analysis.TokenStream{
&analysis.Token{
Term: []byte("b'fhearr"),
},
},
output: analysis.TokenStream{
&analysis.Token{
Term: []byte("fhearr"),
},
},
},
}
for _, test := range tests {
elisionFilter := NewElisionFilter(test.articleMap)
actual := elisionFilter.Filter(test.input)
if !reflect.DeepEqual(actual, test.output) {
t.Errorf("expected %s, got %s", test.output[0].Term, actual[0].Term)
}
}
}