|
@@ -0,0 +1,107 @@
|
|
|
+/*
|
|
|
+ * Licensed to Elasticsearch under one or more contributor
|
|
|
+ * license agreements. See the NOTICE file distributed with
|
|
|
+ * this work for additional information regarding copyright
|
|
|
+ * ownership. Elasticsearch licenses this file to you under
|
|
|
+ * the Apache License, Version 2.0 (the "License"); you may
|
|
|
+ * not use this file except in compliance with the License.
|
|
|
+ * You may obtain a copy of the License at
|
|
|
+ *
|
|
|
+ * http://www.apache.org/licenses/LICENSE-2.0
|
|
|
+ *
|
|
|
+ * Unless required by applicable law or agreed to in writing,
|
|
|
+ * software distributed under the License is distributed on an
|
|
|
+ * "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
|
|
|
+ * KIND, either express or implied. See the License for the
|
|
|
+ * specific language governing permissions and limitations
|
|
|
+ * under the License.
|
|
|
+ */
|
|
|
+
|
|
|
+package org.elasticsearch.index.analysis;
|
|
|
+
|
|
|
+import org.apache.lucene.analysis.icu.segmentation.ICUTokenizer;
|
|
|
+import org.elasticsearch.Version;
|
|
|
+import org.elasticsearch.cluster.metadata.IndexMetaData;
|
|
|
+import org.elasticsearch.common.settings.Settings;
|
|
|
+import org.elasticsearch.env.Environment;
|
|
|
+import org.elasticsearch.index.Index;
|
|
|
+import org.elasticsearch.plugin.analysis.icu.AnalysisICUPlugin;
|
|
|
+import org.elasticsearch.test.ESTestCase;
|
|
|
+
|
|
|
+import java.io.IOException;
|
|
|
+import java.io.InputStream;
|
|
|
+import java.io.Reader;
|
|
|
+import java.io.StringReader;
|
|
|
+import java.nio.file.Files;
|
|
|
+import java.nio.file.Path;
|
|
|
+
|
|
|
+import static org.apache.lucene.analysis.BaseTokenStreamTestCase.assertTokenStreamContents;
|
|
|
+
|
|
|
+/**
|
|
|
+ */
|
|
|
+public class IcuTokenizerFactoryTests extends ESTestCase {
|
|
|
+
|
|
|
+ public void testSimpleIcuTokenizer() throws IOException {
|
|
|
+ AnalysisService analysisService = createAnalysisService();
|
|
|
+
|
|
|
+ TokenizerFactory tokenizerFactory = analysisService.tokenizer("icu_tokenizer");
|
|
|
+ ICUTokenizer tokenizer = (ICUTokenizer) tokenizerFactory.create();
|
|
|
+
|
|
|
+ Reader reader = new StringReader("向日葵, one-two");
|
|
|
+ tokenizer.setReader(reader);
|
|
|
+ assertTokenStreamContents(tokenizer, new String[]{"向日葵", "one", "two"});
|
|
|
+ }
|
|
|
+
|
|
|
+ public void testIcuCustomizeRuleFile() throws IOException {
|
|
|
+ AnalysisService analysisService = createAnalysisService();
|
|
|
+
|
|
|
+ // test the tokenizer with single rule file
|
|
|
+ TokenizerFactory tokenizerFactory = analysisService.tokenizer("user_rule_tokenizer");
|
|
|
+ ICUTokenizer tokenizer = (ICUTokenizer) tokenizerFactory.create();
|
|
|
+ Reader reader = new StringReader
|
|
|
+ ("One-two punch. Brang-, not brung-it. This one--not that one--is the right one, -ish.");
|
|
|
+
|
|
|
+ tokenizer.setReader(reader);
|
|
|
+ assertTokenStreamContents(tokenizer,
|
|
|
+ new String[]{"One-two", "punch", "Brang", "not", "brung-it",
|
|
|
+ "This", "one", "not", "that", "one", "is", "the", "right", "one", "ish"});
|
|
|
+ }
|
|
|
+
|
|
|
+ public void testMultipleIcuCustomizeRuleFiles() throws IOException {
|
|
|
+ AnalysisService analysisService = createAnalysisService();
|
|
|
+
|
|
|
+ // test the tokenizer with two rule files
|
|
|
+ TokenizerFactory tokenizerFactory = analysisService.tokenizer("multi_rule_tokenizer");
|
|
|
+ ICUTokenizer tokenizer = (ICUTokenizer) tokenizerFactory.create();
|
|
|
+ StringReader reader = new StringReader
|
|
|
+ ("Some English. Немного русский. ข้อความภาษาไทยเล็ก ๆ น้อย ๆ More English.");
|
|
|
+
|
|
|
+ tokenizer.setReader(reader);
|
|
|
+ assertTokenStreamContents(tokenizer, new String[]{"Some", "English",
|
|
|
+ "Немного русский. ",
|
|
|
+ "ข้อความภาษาไทยเล็ก ๆ น้อย ๆ ",
|
|
|
+ "More", "English"});
|
|
|
+ }
|
|
|
+
|
|
|
+
|
|
|
+ private static AnalysisService createAnalysisService() throws IOException {
|
|
|
+ InputStream keywords = IcuTokenizerFactoryTests.class.getResourceAsStream("KeywordTokenizer.rbbi");
|
|
|
+ InputStream latin = IcuTokenizerFactoryTests.class.getResourceAsStream("Latin-dont-break-on-hyphens.rbbi");
|
|
|
+
|
|
|
+ Path home = createTempDir();
|
|
|
+ Path config = home.resolve("config");
|
|
|
+ Files.createDirectory(config);
|
|
|
+ Files.copy(keywords, config.resolve("KeywordTokenizer.rbbi"));
|
|
|
+ Files.copy(latin, config.resolve("Latin-dont-break-on-hyphens.rbbi"));
|
|
|
+
|
|
|
+ String json = "/org/elasticsearch/index/analysis/icu_analysis.json";
|
|
|
+
|
|
|
+ Settings settings = Settings.builder()
|
|
|
+ .loadFromStream(json, IcuTokenizerFactoryTests.class.getResourceAsStream(json))
|
|
|
+ .put(IndexMetaData.SETTING_VERSION_CREATED, Version.CURRENT)
|
|
|
+ .build();
|
|
|
+ Settings nodeSettings = Settings.builder().put(Environment.PATH_HOME_SETTING.getKey(), home).build();
|
|
|
+
|
|
|
+ return createAnalysisService(new Index("test", "_na_"), nodeSettings, settings, new AnalysisICUPlugin()::onModule);
|
|
|
+ }
|
|
|
+}
|