1 | |
|
2 | |
|
3 | |
|
4 | |
|
5 | |
|
6 | |
|
7 | |
|
8 | |
|
9 | |
|
10 | |
|
11 | |
|
12 | |
|
13 | |
|
14 | |
|
15 | |
|
16 | |
|
17 | |
|
18 | |
|
19 | |
|
20 | |
package org.crosswire.jsword.index.lucene.analysis; |
21 | |
|
22 | |
import java.io.IOException; |
23 | |
import java.io.Reader; |
24 | |
|
25 | |
import org.apache.lucene.analysis.LowerCaseFilter; |
26 | |
import org.apache.lucene.analysis.StopFilter; |
27 | |
import org.apache.lucene.analysis.TokenStream; |
28 | |
import org.apache.lucene.analysis.ar.ArabicAnalyzer; |
29 | |
import org.apache.lucene.analysis.ar.ArabicLetterTokenizer; |
30 | |
import org.apache.lucene.analysis.ar.ArabicNormalizationFilter; |
31 | |
import org.apache.lucene.analysis.ar.ArabicStemFilter; |
32 | |
import org.apache.lucene.util.Version; |
33 | |
|
34 | |
|
35 | |
|
36 | |
|
37 | |
|
38 | |
|
39 | |
|
40 | |
|
41 | |
|
42 | |
|
43 | |
public class ArabicLuceneAnalyzer extends AbstractBookAnalyzer { |
44 | 0 | public ArabicLuceneAnalyzer() { |
45 | 0 | stopSet = ArabicAnalyzer.getDefaultStopSet(); |
46 | 0 | } |
47 | |
|
48 | |
|
49 | |
|
50 | |
|
51 | |
@Override |
52 | |
public final TokenStream tokenStream(String fieldName, Reader reader) { |
53 | 0 | TokenStream result = new ArabicLetterTokenizer(reader); |
54 | 0 | result = new LowerCaseFilter(result); |
55 | 0 | result = new ArabicNormalizationFilter(result); |
56 | 0 | if (doStopWords && stopSet != null) { |
57 | 0 | result = new StopFilter(false, result, stopSet); |
58 | |
} |
59 | |
|
60 | 0 | if (doStemming) { |
61 | 0 | result = new ArabicStemFilter(result); |
62 | |
} |
63 | |
|
64 | 0 | return result; |
65 | |
} |
66 | |
|
67 | |
|
68 | |
|
69 | |
|
70 | |
@Override |
71 | |
public TokenStream reusableTokenStream(String fieldName, Reader reader) throws IOException { |
72 | 0 | SavedStreams streams = (SavedStreams) getPreviousTokenStream(); |
73 | 0 | if (streams == null) { |
74 | 0 | streams = new SavedStreams(new ArabicLetterTokenizer(reader)); |
75 | 0 | streams.setResult(new LowerCaseFilter(streams.getResult())); |
76 | 0 | streams.setResult(new ArabicNormalizationFilter(streams.getResult())); |
77 | |
|
78 | 0 | if (doStopWords && stopSet != null) { |
79 | 0 | streams.setResult(new StopFilter(StopFilter.getEnablePositionIncrementsVersionDefault(matchVersion), streams.getResult(), stopSet)); |
80 | |
} |
81 | |
|
82 | 0 | if (doStemming) { |
83 | 0 | streams.setResult(new ArabicStemFilter(streams.getResult())); |
84 | |
} |
85 | |
|
86 | 0 | setPreviousTokenStream(streams); |
87 | |
} else { |
88 | 0 | streams.getSource().reset(reader); |
89 | |
} |
90 | 0 | return streams.getResult(); |
91 | |
} |
92 | |
|
93 | 0 | private final Version matchVersion = Version.LUCENE_29; |
94 | |
} |