Searched defs:source (Results 1 - 25 of 58) sorted by relevance

123

/lucene-3.6.0/lucene/contrib/analyzers/common/src/java/org/apache/lucene/analysis/sinks/
H A DTokenTypeSinkFilter.java36 public boolean accept(AttributeSource source) { argument
38 typeAtt = source.addAttribute(TypeAttribute.class);
H A DDateRecognizerSinkFilter.java52 public boolean accept(AttributeSource source) { argument
54 termAtt = source.addAttribute(CharTermAttribute.class);
H A DTokenRangeSinkFilter.java41 public boolean accept(AttributeSource source) { argument
/lucene-3.6.0/lucene/contrib/icu/src/java/org/apache/lucene/collation/
H A DICUCollationKeyAnalyzer.java89 Tokenizer source; field in class:ICUCollationKeyAnalyzer.SavedStreams
100 streams.source = new KeywordTokenizer(reader);
101 streams.result = new ICUCollationKeyFilter(streams.source, collator);
104 streams.source.reset(reader);
/lucene-3.6.0/lucene/contrib/queryparser/src/java/org/apache/lucene/queryParser/standard/config/
H A DNumberDateFormat.java60 public Number parse(String source, ParsePosition parsePosition) { argument
61 final Date date = dateFormat.parse(source, parsePosition);
/lucene-3.6.0/lucene/contrib/spellchecker/src/java/org/apache/lucene/search/spell/
H A DNGramDistance.java53 public float getDistance(String source, String target) { argument
54 final int sl = source.length();
69 if (source.charAt(i) == target.charAt(i)) {
87 sa[i] = source.charAt(i-n+1);
94 int i; // iterates through source
/lucene-3.6.0/lucene/contrib/spellchecker/src/java/org/apache/lucene/search/suggest/
H A DBufferingTermFreqIteratorWrapper.java37 public BufferingTermFreqIteratorWrapper(TermFreqIterator source) throws IOException { argument
38 this.comp = source.getComparator();
41 while((spare = source.next()) != null) {
46 freqs[freqIndex++] = source.weight();
H A DUnsortedTermFreqIteratorWrapper.java36 public UnsortedTermFreqIteratorWrapper(TermFreqIterator source) throws IOException { argument
37 super(source);
/lucene-3.6.0/lucene/core/src/java/org/apache/lucene/analysis/
H A DLetterTokenizer.java64 * @param source
65 * the attribute source to use for this {@link Tokenizer}
69 public LetterTokenizer(Version matchVersion, AttributeSource source, Reader in) { argument
70 super(matchVersion, source, in);
106 public LetterTokenizer(AttributeSource source, Reader in) { argument
107 super(Version.LUCENE_30, source, in);
H A DLowerCaseTokenizer.java66 * @param source
67 * the attribute source to use for this {@link Tokenizer}
71 public LowerCaseTokenizer(Version matchVersion, AttributeSource source, Reader in) { argument
72 super(matchVersion, source, in);
108 public LowerCaseTokenizer(AttributeSource source, Reader in) { argument
109 super(Version.LUCENE_30, source, in);
H A DTokenizer.java34 /** The text source for this Tokenizer. */
66 protected Tokenizer(AttributeSource source) { argument
67 super(source);
71 protected Tokenizer(AttributeSource source, Reader input) { argument
72 super(source);
H A DWhitespaceTokenizer.java56 * @param source
57 * the attribute source to use for this {@link Tokenizer}
61 public WhitespaceTokenizer(Version matchVersion, AttributeSource source, Reader in) { argument
62 super(matchVersion, source, in);
99 public WhitespaceTokenizer(AttributeSource source, Reader in) { argument
100 super(source, in);
H A DKeywordTokenizer.java48 public KeywordTokenizer(AttributeSource source, Reader input, int bufferSize) { argument
49 super(source, input);
H A DReusableAnalyzerBase.java64 * @param reader the reader the streams source reads from
87 * @param reader the reader the streams source reads from
104 * access to the source ({@link Tokenizer}) and the outer end (sink), an
111 protected final Tokenizer source; field in class:ReusableAnalyzerBase.TokenStreamComponents
117 * @param source
122 public TokenStreamComponents(final Tokenizer source, argument
124 this.source = source;
131 * @param source
134 public TokenStreamComponents(final Tokenizer source) { argument
[all...]
/lucene-3.6.0/lucene/core/src/java/org/apache/lucene/collation/
H A DCollationKeyAnalyzer.java94 Tokenizer source; field in class:CollationKeyAnalyzer.SavedStreams
105 streams.source = new KeywordTokenizer(reader);
106 streams.result = new CollationKeyFilter(streams.source, collator);
109 streams.source.reset(reader);
/lucene-3.6.0/lucene/contrib/analyzers/common/src/java/org/apache/lucene/analysis/ar/
H A DArabicLetterTokenizer.java65 * @param source
66 * the attribute source to use for this Tokenizer
70 public ArabicLetterTokenizer(Version matchVersion, AttributeSource source, Reader in) { argument
71 super(matchVersion, source, in);
107 public ArabicLetterTokenizer(AttributeSource source, Reader in) { argument
108 super(source, in);
/lucene-3.6.0/lucene/contrib/analyzers/common/src/java/org/apache/lucene/analysis/in/
H A DIndicTokenizer.java38 public IndicTokenizer(Version matchVersion, AttributeSource source, Reader input) { argument
39 super(matchVersion, source, input);
/lucene-3.6.0/lucene/contrib/analyzers/common/src/java/org/apache/lucene/analysis/ru/
H A DRussianLetterTokenizer.java65 * @param source
66 * the attribute source to use for this {@link Tokenizer}
70 public RussianLetterTokenizer(Version matchVersion, AttributeSource source, Reader in) { argument
71 super(matchVersion, source, in);
107 public RussianLetterTokenizer(AttributeSource source, Reader in) { argument
108 super(source, in);
/lucene-3.6.0/lucene/contrib/benchmark/src/java/org/apache/lucene/benchmark/byTask/tasks/
H A DConsumeContentSourceTask.java29 * <li>content.source - the content source to use. (mandatory)
34 private ContentSource source; field in class:ConsumeContentSourceTask
40 String sourceClass = config.get("content.source", null);
42 throw new IllegalArgumentException("content.source must be defined");
45 source = Class.forName(sourceClass).asSubclass(ContentSource.class).newInstance();
46 source.setConfig(config);
47 source.resetInputs();
55 return "read " + recsCount + " documents from the content source";
60 source
[all...]
/lucene-3.6.0/solr/core/src/java/org/apache/solr/schema/
H A DCopyField.java27 private final SchemaField source; field in class:CopyField
32 public CopyField(final SchemaField source, final SchemaField destination) { argument
33 this(source, destination, UNLIMITED);
37 * @param source The SchemaField of the source field.
39 * @param maxChars Maximum number of chars in source field to copy to destination field.
42 public CopyField(final SchemaField source, final SchemaField destination, argument
44 if (source == null || destination == null) {
52 this.source = source;
[all...]
/lucene-3.6.0/solr/core/src/java/org/apache/solr/search/function/
H A DSimpleFloatFunction.java28 public SimpleFloatFunction(ValueSource source) { argument
29 super(source);
36 final DocValues vals = source.getValues(context, reader);
H A DSingleFunction.java28 protected final ValueSource source; field in class:SingleFunction
30 public SingleFunction(ValueSource source) { argument
31 this.source = source;
38 return name() + '(' + source.description() + ')';
43 return source.hashCode() + name().hashCode();
51 && this.source.equals(other.source);
56 source.createWeight(context, searcher);
H A DLinearFloatFunction.java35 protected final ValueSource source; field in class:LinearFloatFunction
39 public LinearFloatFunction(ValueSource source, float slope, float intercept) { argument
40 this.source = source;
47 return slope + "*float(" + source.description() + ")+" + intercept;
52 final DocValues vals = source.getValues(context, reader);
83 source.createWeight(context, searcher);
92 return h + source.hashCode();
101 && this.source.equals(other.source);
[all...]
/lucene-3.6.0/solr/core/src/java/org/apache/solr/analysis/
H A DSlowSynonymFilter.java40 * Generated synonyms will start at the same position as the first matched source token.
250 private void copy(AttributeSource target, AttributeSource source) { argument
251 if (target != source)
252 source.copyTo(target);
/lucene-3.6.0/lucene/contrib/analyzers/common/src/java/org/apache/lucene/analysis/snowball/
H A DSnowballAnalyzer.java99 Tokenizer source; field in class:SnowballAnalyzer.SavedStreams
112 streams.source = new StandardTokenizer(matchVersion, reader);
113 streams.result = new StandardFilter(matchVersion, streams.source);
125 streams.source.reset(reader);

Completed in 75 milliseconds

123