当前位置: 首页>>代码示例>>Java>>正文


Java AnalysisService.tokenFilter方法代码示例

本文整理汇总了Java中org.elasticsearch.index.analysis.AnalysisService.tokenFilter方法的典型用法代码示例。如果您正苦于以下问题:Java AnalysisService.tokenFilter方法的具体用法?Java AnalysisService.tokenFilter怎么用?Java AnalysisService.tokenFilter使用的例子?那么恭喜您, 这里精选的方法代码示例或许可以为您提供帮助。您也可以进一步了解该方法所在org.elasticsearch.index.analysis.AnalysisService的用法示例。


在下文中一共展示了AnalysisService.tokenFilter方法的15个代码示例,这些例子默认根据受欢迎程度排序。您可以为喜欢或者感觉有用的代码点赞,您的评价将有助于系统推荐出更棒的Java代码示例。

示例1: testTokenizerFilter

import org.elasticsearch.index.analysis.AnalysisService; //导入方法依赖的package包/类
@Test
public void testTokenizerFilter() throws IOException {

    String source = "eins zwei drei";

    String[] expected = {
            "d800000000001214",
            "4a00000000000000",
            "c0000000000a100"
    };
    AnalysisService analysisService = MapperTestUtils.analysisService("eudex_analysis.json");
    TokenFilterFactory tokenFilterFactory = analysisService.tokenFilter("my_phonetic");
    Tokenizer tokenizer = analysisService.tokenizer("my_phonetic").create();
    tokenizer.setReader(new StringReader(source));
    assertSimpleTSOutput(tokenFilterFactory.create(tokenizer), expected);
}
 
开发者ID:jprante,项目名称:elasticsearch-analysis-phonetic-eudex,代码行数:17,代码来源:EudexAnalyzerTests.java

示例2: testOpenNLPAnalysis

import org.elasticsearch.index.analysis.AnalysisService; //导入方法依赖的package包/类
@Test
public void testOpenNLPAnalysis() throws Exception {
    AnalysisService analysisService = MapperTestUtils.analysisService(getResource());
    TokenizerFactory tokenizerFactory = analysisService.tokenizer("opennlp");
    MatcherAssert.assertThat(tokenizerFactory, instanceOf(OpenNLPTokenizerFactory.class));
    this.tokenizerFactory = (OpenNLPTokenizerFactory) tokenizerFactory;
    TokenFilterFactory filterFactory = analysisService.tokenFilter("opennlp");
    MatcherAssert.assertThat(filterFactory, instanceOf(OpenNLPTokenFilterFactory.class));
    this.filterFactory = (OpenNLPTokenFilterFactory) filterFactory;

    TokenFilterFactory filterpayloadsfilterFactory = analysisService.tokenFilter("filterpayloads");
    MatcherAssert.assertThat(filterpayloadsfilterFactory, instanceOf(FilterPayloadsTokenFilterFactory.class));
    TokenFilterFactory strippayloadsfilterFactory = analysisService.tokenFilter("strippayloads");
    MatcherAssert.assertThat(strippayloadsfilterFactory, instanceOf(StripPayloadsTokenFilterFactory.class));

    analyze();
}
 
开发者ID:jprante,项目名称:elasticsearch-analysis-opennlp,代码行数:18,代码来源:BaseOpenNLPAnalysis.java

示例3: testTwo

import org.elasticsearch.index.analysis.AnalysisService; //导入方法依赖的package包/类
@Test
public void testTwo() throws IOException {

    String source = "Das sind Autos, die Nudeln transportieren.";

    String[] expected = {
            "Das",
            "Das",
            "sind",
            "sind",
            "Autos",
            "Auto",
            "die",
            "der",
            "Nudeln",
            "Nudel",
            "transportieren",
            "transportieren"
    };
    AnalysisService analysisService = MapperTestUtils.analysisService();
    TokenFilterFactory tokenFilter = analysisService.tokenFilter("baseform");
    Tokenizer tokenizer = analysisService.tokenizer("standard").create();
    tokenizer.setReader(new StringReader(source));
    assertSimpleTSOutput(tokenFilter.create(tokenizer), expected);
}
 
开发者ID:jprante,项目名称:elasticsearch-analysis-baseform,代码行数:26,代码来源:BaseformTokenFilterTests.java

示例4: testThree

import org.elasticsearch.index.analysis.AnalysisService; //导入方法依赖的package包/类
@Test
public void testThree() throws IOException {

    String source = "wurde zum tollen gemacht";

    String[] expected = {
            "wurde",
            "werden",
            "zum",
            "zum",
            "tollen",
            "tollen",
            "gemacht",
            "machen"
    };
    AnalysisService analysisService = MapperTestUtils.analysisService();
    TokenFilterFactory tokenFilter = analysisService.tokenFilter("baseform");
    Tokenizer tokenizer = analysisService.tokenizer("standard").create();
    tokenizer.setReader(new StringReader(source));
    assertSimpleTSOutput(tokenFilter.create(tokenizer), expected);
}
 
开发者ID:jprante,项目名称:elasticsearch-analysis-baseform,代码行数:22,代码来源:BaseformTokenFilterTests.java

示例5: test

import org.elasticsearch.index.analysis.AnalysisService; //导入方法依赖的package包/类
@Test
public void test() throws IOException {

    String source = "Ein schöner Tag in Köln im Café an der Straßenecke";

    String[] expected = {
        "Ein",
        "schoner",
        "Tag",
        "in",
        "Koln",
        "im",
        "Café",
        "an",
        "der",
        "Strassenecke"
    };

    AnalysisService analysisService = createAnalysisService();
    TokenFilterFactory tokenFilter = analysisService.tokenFilter("umlaut");
    Tokenizer tokenizer = analysisService.tokenizer("standard").create(new StringReader(source));
    assertSimpleTSOutput(tokenFilter.create(tokenizer), expected);
}
 
开发者ID:jprante,项目名称:elasticsearch-analysis-german,代码行数:24,代码来源:GermanNormalizationTests.java

示例6: testTwo

import org.elasticsearch.index.analysis.AnalysisService; //导入方法依赖的package包/类
@Test
public void testTwo() throws IOException {
    AnalysisService analysisService = createAnalysisService();
    TokenFilterFactory tokenFilter = analysisService.tokenFilter("baseform");

    String source = "Das sind Autos, die Nudeln transportieren.";

    String[] expected = {
            "Das",
            "Das",
            "sind",
            "sind",
            "Autos",
            "Auto",
            "die",
            "der",
            "Nudeln",
            "Nudel",
            "transportieren",
            "transportieren"
    };

    Tokenizer tokenizer = new StandardTokenizer(Version.LATEST, new StringReader(source));
    assertSimpleTSOutput(tokenFilter.create(tokenizer), expected);
}
 
开发者ID:jprante,项目名称:elasticsearch-analysis-german,代码行数:26,代码来源:BaseformTokenFilterTests.java

示例7: testOne

import org.elasticsearch.index.analysis.AnalysisService; //导入方法依赖的package包/类
@Test
public void testOne() throws IOException {

    String source = "Das ist ein Bindestrich-Wort.";

    String[] expected = {
            "Das",
            "ist",
            "ein",
            "Bindestrich-Wort",
            "BindestrichWort",
            "Wort",
            "Bindestrich"
    };
    AnalysisService analysisService = createAnalysisService();
    Tokenizer tokenizer = analysisService.tokenizer("my_hyphen_tokenizer").create(new StringReader(source));
    TokenFilterFactory tokenFilter = analysisService.tokenFilter("hyphen");
    TokenStream tokenStream = tokenFilter.create(tokenizer);
    assertSimpleTSOutput(tokenStream, expected);
}
 
开发者ID:jprante,项目名称:elasticsearch-analysis-german,代码行数:21,代码来源:HyphenTokenizerTests.java

示例8: testTwo

import org.elasticsearch.index.analysis.AnalysisService; //导入方法依赖的package包/类
@Test
public void testTwo() throws IOException {

    String source = "Das E-Book muss dringend zum Buchbinder.";

    String[] expected = {
            "Das",
            "E-Book",
            "EBook",
            "Book",
            "muss",
            "dringend",
            "zum",
            "Buchbinder"
    };
    AnalysisService analysisService = createAnalysisService();
    Tokenizer tokenizer = analysisService.tokenizer("my_icu_tokenizer").create(new StringReader(source));
    TokenFilterFactory tokenFilter = analysisService.tokenFilter("hyphen");
    assertSimpleTSOutput(tokenFilter.create(tokenizer), expected);
}
 
开发者ID:jprante,项目名称:elasticsearch-analysis-german,代码行数:21,代码来源:HyphenTokenizerTests.java

示例9: testFour

import org.elasticsearch.index.analysis.AnalysisService; //导入方法依赖的package包/类
@Test
public void testFour() throws IOException {

    String source = "So wird's was: das Elasticsearch-Buch erscheint beim O'Reilly-Verlag.";

    String[] expected = {
            "So",
            "wird's",
            "was",
            "das",
            "Elasticsearch-Buch",
            "ElasticsearchBuch",
            "Buch",
            "Elasticsearch",
            "erscheint",
            "beim",
            "O'Reilly-Verlag"
    };
    AnalysisService analysisService = createAnalysisService();
    Tokenizer tokenizer = analysisService.tokenizer("my_hyphen_tokenizer").create(new StringReader(source));
    TokenFilterFactory tokenFilter = analysisService.tokenFilter("hyphen");
    assertSimpleTSOutput(tokenFilter.create(tokenizer), expected);
}
 
开发者ID:jprante,项目名称:elasticsearch-analysis-german,代码行数:24,代码来源:HyphenTokenizerTests.java

示例10: testSix

import org.elasticsearch.index.analysis.AnalysisService; //导入方法依赖的package包/类
@Test
public void testSix() throws IOException {

    String source = "E-Book";

    String[] expected = {
            "E-Book",
            "EBook",
            "Book"
    };

    AnalysisService analysisService = createAnalysisService();
    Tokenizer tokenizer = analysisService.tokenizer("my_hyphen_tokenizer").create(new StringReader(source));
    TokenFilterFactory tokenFilter = analysisService.tokenFilter("hyphen");
    assertSimpleTSOutput(tokenFilter.create(tokenizer), expected);
}
 
开发者ID:jprante,项目名称:elasticsearch-analysis-german,代码行数:17,代码来源:HyphenTokenizerTests.java

示例11: testSeven

import org.elasticsearch.index.analysis.AnalysisService; //导入方法依赖的package包/类
@Test
public void testSeven() throws IOException {
    String source = "Procter & Gamble ist nicht schwarz - weiss";

    String[] expected = {
            "Procter",
            "Gamble",
            "ist",
            "nicht",
            "schwarz",
            "weiss"
    };

    AnalysisService analysisService = createAnalysisService();
    Tokenizer tokenizer = analysisService.tokenizer("my_hyphen_tokenizer").create(new StringReader(source));
    TokenFilterFactory tokenFilter = analysisService.tokenFilter("hyphen");
    assertSimpleTSOutput(tokenFilter.create(tokenizer), expected);
}
 
开发者ID:jprante,项目名称:elasticsearch-analysis-german,代码行数:19,代码来源:HyphenTokenizerTests.java

示例12: testDefaultsRSLPAnalysis

import org.elasticsearch.index.analysis.AnalysisService; //导入方法依赖的package包/类
@Test
public void testDefaultsRSLPAnalysis() throws IOException {
    AnalysisService analysisService = createAnalysisService(new Index("test", "_na_"), Settings.EMPTY, new AnalysisMetaphonePlugin());

    TokenFilterFactory tokenizerFactory = analysisService.tokenFilter("br_metaphone");

    MatcherAssert.assertThat(tokenizerFactory, instanceOf(MetaphoneTokenFilterFactory.class));
}
 
开发者ID:anaelcarvalho,项目名称:elasticsearch-analysis-metaphone_ptBR,代码行数:9,代码来源:MetaphoneAnalysisTests.java

示例13: testMetaphoneWords

import org.elasticsearch.index.analysis.AnalysisService; //导入方法依赖的package包/类
@Test
public void testMetaphoneWords() throws Exception {
    Index index = new Index("test", "_na_");
    Settings settings = Settings.builder()
            .put(IndexMetaData.SETTING_VERSION_CREATED, Version.CURRENT)
            .put("index.analysis.filter.myStemmer.type", "br_metaphone")
            .build();

    AnalysisService analysisService = createAnalysisService(index, settings, new AnalysisMetaphonePlugin());

    TokenFilterFactory filterFactory = analysisService.tokenFilter("br_metaphone");

    Tokenizer tokenizer = new KeywordTokenizer();
    
    Map<String,String> words = buildWordList();
    
    Set<String> inputWords = words.keySet();
    for(String word : inputWords) {
        tokenizer.setReader(new StringReader(word));
        TokenStream ts = filterFactory.create(tokenizer);

        CharTermAttribute term1 = ts.addAttribute(CharTermAttribute.class);
        ts.reset();
        assertThat(ts.incrementToken(), equalTo(true));
        assertThat(term1.toString(), equalTo(words.get(word)));
        ts.close();
    }
}
 
开发者ID:anaelcarvalho,项目名称:elasticsearch-analysis-metaphone_ptBR,代码行数:29,代码来源:MetaphoneTokenFilterTests.java

示例14: testDefaultsIcuAnalysis

import org.elasticsearch.index.analysis.AnalysisService; //导入方法依赖的package包/类
@Test
public void testDefaultsIcuAnalysis() {
    Settings settings = Settings.settingsBuilder()
            .put(IndexMetaData.SETTING_VERSION_CREATED, Version.CURRENT)
            .put("path.home", System.getProperty("path.home"))
            .build();

    Index index = new Index("test");

    Injector parentInjector = new ModulesBuilder()
            .add(new SettingsModule(settings),
                    new EnvironmentModule(new Environment(settings)))
            .createInjector();
    Injector injector = new ModulesBuilder().add(
            new IndexSettingsModule(index, settings),
            new IndexNameModule(index),
            new AnalysisModule(settings, parentInjector.getInstance(IndicesAnalysisService.class))
                    .addProcessor(new IcuAnalysisBinderProcessor()))
            .createChildInjector(parentInjector);

    AnalysisService analysisService = injector.getInstance(AnalysisService.class);

    TokenizerFactory tokenizerFactory = analysisService.tokenizer("icu_tokenizer");
    assertThat(tokenizerFactory, instanceOf(IcuTokenizerFactory.class));

    TokenFilterFactory filterFactory = analysisService.tokenFilter("icu_normalizer");
    assertThat(filterFactory, instanceOf(IcuNormalizerTokenFilterFactory.class));

    filterFactory = analysisService.tokenFilter("icu_folding");
    assertThat(filterFactory, instanceOf(IcuFoldingTokenFilterFactory.class));

    filterFactory = analysisService.tokenFilter("icu_transform");
    assertThat(filterFactory, instanceOf(IcuTransformTokenFilterFactory.class));

    Analyzer analyzer = analysisService.analyzer("icu_collation");
    assertThat(analyzer, instanceOf(NamedAnalyzer.class));

}
 
开发者ID:jprante,项目名称:elasticsearch-icu,代码行数:39,代码来源:IcuAnalysisTests.java

示例15: testOne

import org.elasticsearch.index.analysis.AnalysisService; //导入方法依赖的package包/类
@Test
public void testOne() throws IOException {

    String source = "Die Jahresfeier der Rechtsanwaltskanzleien auf dem Donaudampfschiff hat viel Ökosteuer gekostet";

    String[] expected = {
            "Die",
            "Die",
            "Jahresfeier",
            "Jahresfeier",
            "der",
            "der",
            "Rechtsanwaltskanzleien",
            "Rechtsanwaltskanzlei",
            "auf",
            "auf",
            "dem",
            "der",
            "Donaudampfschiff",
            "Donaudampfschiff",
            "hat",
            "haben",
            "viel",
            "viel",
            "Ökosteuer",
            "Ökosteuer",
            "gekostet",
            "kosten"
    };
    AnalysisService analysisService = MapperTestUtils.analysisService();
    TokenFilterFactory tokenFilter = analysisService.tokenFilter("baseform");
    Tokenizer tokenizer = analysisService.tokenizer("standard").create();
    tokenizer.setReader(new StringReader(source));
    assertSimpleTSOutput(tokenFilter.create(tokenizer), expected);
}
 
开发者ID:jprante,项目名称:elasticsearch-analysis-baseform,代码行数:36,代码来源:BaseformTokenFilterTests.java


注:本文中的org.elasticsearch.index.analysis.AnalysisService.tokenFilter方法示例由纯净天空整理自Github/MSDocs等开源代码及文档管理平台,相关代码片段筛选自各路编程大神贡献的开源项目,源码版权归原作者所有,传播和使用请参考对应项目的License;未经允许,请勿转载。