/lucene-3.6.0/lucene/contrib/analyzers/common/src/java/org/apache/lucene/analysis/sinks/ |
H A D | TokenTypeSinkFilter.java | 36 public boolean accept(AttributeSource source) { argument 38 typeAtt = source.addAttribute(TypeAttribute.class);
|
H A D | DateRecognizerSinkFilter.java | 52 public boolean accept(AttributeSource source) { argument 54 termAtt = source.addAttribute(CharTermAttribute.class);
|
H A D | TokenRangeSinkFilter.java | 41 public boolean accept(AttributeSource source) { argument
|
/lucene-3.6.0/lucene/contrib/icu/src/java/org/apache/lucene/collation/ |
H A D | ICUCollationKeyAnalyzer.java | 89 Tokenizer source; field in class:ICUCollationKeyAnalyzer.SavedStreams 100 streams.source = new KeywordTokenizer(reader); 101 streams.result = new ICUCollationKeyFilter(streams.source, collator); 104 streams.source.reset(reader);
|
/lucene-3.6.0/lucene/contrib/queryparser/src/java/org/apache/lucene/queryParser/standard/config/ |
H A D | NumberDateFormat.java | 60 public Number parse(String source, ParsePosition parsePosition) { argument 61 final Date date = dateFormat.parse(source, parsePosition);
|
/lucene-3.6.0/lucene/contrib/spellchecker/src/java/org/apache/lucene/search/spell/ |
H A D | NGramDistance.java | 53 public float getDistance(String source, String target) { argument 54 final int sl = source.length(); 69 if (source.charAt(i) == target.charAt(i)) { 87 sa[i] = source.charAt(i-n+1); 94 int i; // iterates through source
|
/lucene-3.6.0/lucene/contrib/spellchecker/src/java/org/apache/lucene/search/suggest/ |
H A D | BufferingTermFreqIteratorWrapper.java | 37 public BufferingTermFreqIteratorWrapper(TermFreqIterator source) throws IOException { argument 38 this.comp = source.getComparator(); 41 while((spare = source.next()) != null) { 46 freqs[freqIndex++] = source.weight();
|
H A D | UnsortedTermFreqIteratorWrapper.java | 36 public UnsortedTermFreqIteratorWrapper(TermFreqIterator source) throws IOException { argument 37 super(source);
|
/lucene-3.6.0/lucene/core/src/java/org/apache/lucene/analysis/ |
H A D | LetterTokenizer.java | 64 * @param source 65 * the attribute source to use for this {@link Tokenizer} 69 public LetterTokenizer(Version matchVersion, AttributeSource source, Reader in) { argument 70 super(matchVersion, source, in); 106 public LetterTokenizer(AttributeSource source, Reader in) { argument 107 super(Version.LUCENE_30, source, in);
|
H A D | LowerCaseTokenizer.java | 66 * @param source 67 * the attribute source to use for this {@link Tokenizer} 71 public LowerCaseTokenizer(Version matchVersion, AttributeSource source, Reader in) { argument 72 super(matchVersion, source, in); 108 public LowerCaseTokenizer(AttributeSource source, Reader in) { argument 109 super(Version.LUCENE_30, source, in);
|
H A D | Tokenizer.java | 34 /** The text source for this Tokenizer. */ 66 protected Tokenizer(AttributeSource source) { argument 67 super(source); 71 protected Tokenizer(AttributeSource source, Reader input) { argument 72 super(source);
|
H A D | WhitespaceTokenizer.java | 56 * @param source 57 * the attribute source to use for this {@link Tokenizer} 61 public WhitespaceTokenizer(Version matchVersion, AttributeSource source, Reader in) { argument 62 super(matchVersion, source, in); 99 public WhitespaceTokenizer(AttributeSource source, Reader in) { argument 100 super(source, in);
|
H A D | KeywordTokenizer.java | 48 public KeywordTokenizer(AttributeSource source, Reader input, int bufferSize) { argument 49 super(source, input);
|
H A D | ReusableAnalyzerBase.java | 64 * @param reader the reader the streams source reads from 87 * @param reader the reader the streams source reads from 104 * access to the source ({@link Tokenizer}) and the outer end (sink), an 111 protected final Tokenizer source; field in class:ReusableAnalyzerBase.TokenStreamComponents 117 * @param source 122 public TokenStreamComponents(final Tokenizer source, argument 124 this.source = source; 131 * @param source 134 public TokenStreamComponents(final Tokenizer source) { argument [all...] |
/lucene-3.6.0/lucene/core/src/java/org/apache/lucene/collation/ |
H A D | CollationKeyAnalyzer.java | 94 Tokenizer source; field in class:CollationKeyAnalyzer.SavedStreams 105 streams.source = new KeywordTokenizer(reader); 106 streams.result = new CollationKeyFilter(streams.source, collator); 109 streams.source.reset(reader);
|
/lucene-3.6.0/lucene/contrib/analyzers/common/src/java/org/apache/lucene/analysis/ar/ |
H A D | ArabicLetterTokenizer.java | 65 * @param source 66 * the attribute source to use for this Tokenizer 70 public ArabicLetterTokenizer(Version matchVersion, AttributeSource source, Reader in) { argument 71 super(matchVersion, source, in); 107 public ArabicLetterTokenizer(AttributeSource source, Reader in) { argument 108 super(source, in);
|
/lucene-3.6.0/lucene/contrib/analyzers/common/src/java/org/apache/lucene/analysis/in/ |
H A D | IndicTokenizer.java | 38 public IndicTokenizer(Version matchVersion, AttributeSource source, Reader input) { argument 39 super(matchVersion, source, input);
|
/lucene-3.6.0/lucene/contrib/analyzers/common/src/java/org/apache/lucene/analysis/ru/ |
H A D | RussianLetterTokenizer.java | 65 * @param source 66 * the attribute source to use for this {@link Tokenizer} 70 public RussianLetterTokenizer(Version matchVersion, AttributeSource source, Reader in) { argument 71 super(matchVersion, source, in); 107 public RussianLetterTokenizer(AttributeSource source, Reader in) { argument 108 super(source, in);
|
/lucene-3.6.0/lucene/contrib/benchmark/src/java/org/apache/lucene/benchmark/byTask/tasks/ |
H A D | ConsumeContentSourceTask.java | 29 * <li>content.source - the content source to use. (mandatory) 34 private ContentSource source; field in class:ConsumeContentSourceTask 40 String sourceClass = config.get("content.source", null); 42 throw new IllegalArgumentException("content.source must be defined"); 45 source = Class.forName(sourceClass).asSubclass(ContentSource.class).newInstance(); 46 source.setConfig(config); 47 source.resetInputs(); 55 return "read " + recsCount + " documents from the content source"; 60 source [all...] |
/lucene-3.6.0/solr/core/src/java/org/apache/solr/schema/ |
H A D | CopyField.java | 27 private final SchemaField source; field in class:CopyField 32 public CopyField(final SchemaField source, final SchemaField destination) { argument 33 this(source, destination, UNLIMITED); 37 * @param source The SchemaField of the source field. 39 * @param maxChars Maximum number of chars in source field to copy to destination field. 42 public CopyField(final SchemaField source, final SchemaField destination, argument 44 if (source == null || destination == null) { 52 this.source = source; [all...] |
/lucene-3.6.0/solr/core/src/java/org/apache/solr/search/function/ |
H A D | SimpleFloatFunction.java | 28 public SimpleFloatFunction(ValueSource source) { argument 29 super(source); 36 final DocValues vals = source.getValues(context, reader);
|
H A D | SingleFunction.java | 28 protected final ValueSource source; field in class:SingleFunction 30 public SingleFunction(ValueSource source) { argument 31 this.source = source; 38 return name() + '(' + source.description() + ')'; 43 return source.hashCode() + name().hashCode(); 51 && this.source.equals(other.source); 56 source.createWeight(context, searcher);
|
H A D | LinearFloatFunction.java | 35 protected final ValueSource source; field in class:LinearFloatFunction 39 public LinearFloatFunction(ValueSource source, float slope, float intercept) { argument 40 this.source = source; 47 return slope + "*float(" + source.description() + ")+" + intercept; 52 final DocValues vals = source.getValues(context, reader); 83 source.createWeight(context, searcher); 92 return h + source.hashCode(); 101 && this.source.equals(other.source); [all...] |
/lucene-3.6.0/solr/core/src/java/org/apache/solr/analysis/ |
H A D | SlowSynonymFilter.java | 40 * Generated synonyms will start at the same position as the first matched source token. 250 private void copy(AttributeSource target, AttributeSource source) { argument 251 if (target != source) 252 source.copyTo(target);
|
/lucene-3.6.0/lucene/contrib/analyzers/common/src/java/org/apache/lucene/analysis/snowball/ |
H A D | SnowballAnalyzer.java | 99 Tokenizer source; field in class:SnowballAnalyzer.SavedStreams 112 streams.source = new StandardTokenizer(matchVersion, reader); 113 streams.result = new StandardFilter(matchVersion, streams.source); 125 streams.source.reset(reader);
|