/lucene-3.6.0/lucene/core/src/java/org/apache/lucene/analysis/ |
H A D | ASCIIFoldingFilter.java | 66 private char[] output = new char[512]; field in class:ASCIIFoldingFilter 83 termAtt.copyBuffer(output, 0, outputPos); 103 if (output.length < maxSizeNeeded) { 104 output = new char[ArrayUtil.oversize(maxSizeNeeded, RamUsageEstimator.NUM_BYTES_CHAR)]; 107 outputPos = foldToASCII(input, 0, output, 0, length); 115 * @param output The result of the folding. Should be of size >= {@code length * 4}. 116 * @param outputPos Index of output where to put the result of the folding 118 * @return length of output 121 public static final int foldToASCII(char input[], int inputPos, char output[], int outputPos, int length) argument 129 output[outputPo [all...] |
H A D | ISOLatin1AccentFilter.java | 40 private char[] output = new char[256]; field in class:ISOLatin1AccentFilter 55 termAtt.copyBuffer(output, 0, outputPos); 72 int size = output.length; 76 if (size != output.length) 77 output = new char[size]; 89 output[outputPos++] = c; 98 output[outputPos++] = 'A'; 101 output[outputPos++] = 'A'; 102 output[outputPos++] = 'E'; 105 output[outputPo [all...] |
/lucene-3.6.0/lucene/core/src/java/org/apache/lucene/util/fst/ |
H A D | PositiveIntOutputs.java | 26 * An FST {@link Outputs} implementation where each output 67 public Long subtract(Long output, Long inc) { argument 68 assert valid(output); 70 assert output >= inc; 73 return output; 74 } else if (output.equals(inc)) { 77 return output - inc; 82 public Long add(Long prefix, Long output) { argument 84 assert valid(output); 86 return output; 95 write(Long output, DataOutput out) argument 123 outputToString(Long output) argument [all...] |
H A D | ByteSequenceOutputs.java | 27 * An FST {@link Outputs} implementation where each output 76 public BytesRef subtract(BytesRef output, BytesRef inc) { argument 77 assert output != null; 81 return output; 82 } else if (inc.length == output.length) { 83 // entire output removed 86 assert inc.length < output.length: "inc.length=" + inc.length + " vs output.length=" + output.length; 88 return new BytesRef(output 93 add(BytesRef prefix, BytesRef output) argument 137 outputToString(BytesRef output) argument [all...] |
H A D | IntSequenceOutputs.java | 27 * An FST {@link Outputs} implementation where each output 76 public IntsRef subtract(IntsRef output, IntsRef inc) { argument 77 assert output != null; 81 return output; 82 } else if (inc.length == output.length) { 83 // entire output removed 86 assert inc.length < output.length: "inc.length=" + inc.length + " vs output.length=" + output.length; 88 return new IntsRef(output 93 add(IntsRef prefix, IntsRef output) argument 141 outputToString(IntsRef output) argument [all...] |
H A D | UpToTwoPositiveIntOutputs.java | 26 * An FST {@link Outputs} implementation where each output 28 * single output, Long is returned; else, TwoLongs. Order 30 * input/output added to Builder, and .second is the 31 * second. You cannot store 0 output with this (that's 32 * reserved to mean "no output")! 125 final Long output = (Long) _output; 127 assert output >= inc; 130 return output; 131 } else if (output.equals(inc)) { 134 return output 216 outputToString(Object output) argument [all...] |
H A D | Outputs.java | 39 // output instances -- this is an insane amount of garbage 47 public abstract T subtract(T output, T inc); argument 50 public abstract T add(T prefix, T output); argument 52 public abstract void write(T output, DataOutput out) throws IOException; argument 56 /** NOTE: this output is compared with == so you must 58 * it's really no output */ 61 public abstract String outputToString(T output); argument 63 // TODO: maybe make valid(T output) public...? for asserts
|
H A D | NoOutputs.java | 33 // NodeHash calls hashCode for this output; we fix this 63 public Object subtract(Object output, Object inc) { argument 64 assert output == NO_OUTPUT; 70 public Object add(Object prefix, Object output) { argument 72 assert output == NO_OUTPUT; 94 public String outputToString(Object output) { argument
|
H A D | Builder.java | 29 * output) from pre-sorted terms with outputs. The FST 38 * The parameterized type T is the output type. See the 122 * @param outputs The output type for each input sequence. Applies only if building an FST. For 124 * singleton output object. 259 final T nextFinalOutput = node.output; 308 public void add(IntsRef input, T output) throws IOException { argument 316 if (output == NO_OUTPUT) { 319 System.out.println("\nFST ADD: input=" + toString(b) + " " + b + " output=" + fst.outputs.outputToString(output)); 325 if (output 428 validOutput(T output) argument 478 public T output; field in class:Builder.Arc 506 public T output; field in class:Builder.UnCompiledNode [all...] |
H A D | PairOutputs.java | 123 public Pair<A,B> subtract(Pair<A,B> output, Pair<A,B> inc) { argument 124 assert valid(output); 126 return newPair(outputs1.subtract(output.output1, inc.output1), 127 outputs2.subtract(output.output2, inc.output2)); 131 public Pair<A,B> add(Pair<A,B> prefix, Pair<A,B> output) { argument 133 assert valid(output); 134 return newPair(outputs1.add(prefix.output1, output.output1), 135 outputs2.add(prefix.output2, output.output2)); 139 public void write(Pair<A,B> output, DataOutput writer) throws IOException { argument 140 assert valid(output); 158 outputToString(Pair<A,B> output) argument [all...] |
H A D | FSTEnum.java | 35 @SuppressWarnings({"rawtypes","unchecked"}) protected T[] output = (T[]) new Object[10]; field in class:FSTEnum 50 output[0] = NO_OUTPUT; 177 output[upto] = fst.outputs.add(output[upto-1], arc.output); 218 output[upto] = fst.outputs.add(output[upto-1], arc.output); 317 output[upto] = fst.outputs.add(output[upt [all...] |
/lucene-3.6.0/lucene/backwards/src/test-framework/java/org/apache/lucene/analysis/ |
H A D | BaseTokenStreamTestCase.java | 85 public static void assertTokenStreamContents(TokenStream ts, String[] output, int startOffsets[], int endOffsets[], String types[], int posIncrements[], Integer finalOffset) throws IOException { argument 86 assertNotNull(output); 111 for (int i = 0; i < output.length; i++) { 123 assertEquals("term "+i, output[i], termAtt.toString()); 153 public static void assertTokenStreamContents(TokenStream ts, String[] output, int startOffsets[], int endOffsets[], String types[], int posIncrements[]) throws IOException { argument 154 assertTokenStreamContents(ts, output, startOffsets, endOffsets, types, posIncrements, null); 157 public static void assertTokenStreamContents(TokenStream ts, String[] output) throws IOException { argument 158 assertTokenStreamContents(ts, output, null, null, null, null, null); 161 public static void assertTokenStreamContents(TokenStream ts, String[] output, String[] types) throws IOException { argument 162 assertTokenStreamContents(ts, output, nul 165 assertTokenStreamContents(TokenStream ts, String[] output, int[] posIncrements) argument 169 assertTokenStreamContents(TokenStream ts, String[] output, int startOffsets[], int endOffsets[]) argument 173 assertTokenStreamContents(TokenStream ts, String[] output, int startOffsets[], int endOffsets[], Integer finalOffset) argument 177 assertTokenStreamContents(TokenStream ts, String[] output, int startOffsets[], int endOffsets[], int[] posIncrements) argument 181 assertTokenStreamContents(TokenStream ts, String[] output, int startOffsets[], int endOffsets[], int[] posIncrements, Integer finalOffset) argument 185 assertAnalyzesTo(Analyzer a, String input, String[] output, int startOffsets[], int endOffsets[], String types[], int posIncrements[]) argument 189 assertAnalyzesTo(Analyzer a, String input, String[] output) argument 193 assertAnalyzesTo(Analyzer a, String input, String[] output, String[] types) argument 197 assertAnalyzesTo(Analyzer a, String input, String[] output, int[] posIncrements) argument 201 assertAnalyzesTo(Analyzer a, String input, String[] output, int startOffsets[], int endOffsets[]) argument 205 assertAnalyzesTo(Analyzer a, String input, String[] output, int startOffsets[], int endOffsets[], int[] posIncrements) argument 210 assertAnalyzesToReuse(Analyzer a, String input, String[] output, int startOffsets[], int endOffsets[], String types[], int posIncrements[]) argument 214 assertAnalyzesToReuse(Analyzer a, String input, String[] output) argument 218 assertAnalyzesToReuse(Analyzer a, String input, String[] output, String[] types) argument 222 assertAnalyzesToReuse(Analyzer a, String input, String[] output, int[] posIncrements) argument 226 assertAnalyzesToReuse(Analyzer a, String input, String[] output, int startOffsets[], int endOffsets[]) argument 230 assertAnalyzesToReuse(Analyzer a, String input, String[] output, int startOffsets[], int endOffsets[], int[] posIncrements) argument [all...] |
/lucene-3.6.0/lucene/core/src/java/org/apache/lucene/index/ |
H A D | TermInfosWriter.java | 45 private IndexOutput output; field in class:TermInfosWriter 95 IOUtils.closeWhileHandlingException(output, other); 110 output = directory.createOutput(segment + (isIndex ? ".tii" : ".tis")); 113 output.writeInt(FORMAT_CURRENT); // write format 114 output.writeLong(0); // leave space for size 115 output.writeInt(indexInterval); // write indexInterval 116 output.writeInt(skipInterval); // write skipInterval 117 output.writeInt(maxSkipLevels); // write maxSkipLevels 122 IOUtils.closeWhileHandlingException(output); 198 output [all...] |
H A D | MultiLevelSkipListWriter.java | 130 * Writes the buffered skip lists to the given output. 132 * @param output the IndexOutput the skip lists shall be written to 135 long writeSkip(IndexOutput output) throws IOException { argument 136 long skipPointer = output.getFilePointer(); 142 output.writeVLong(length); 143 skipBuffer[level].writeTo(output); 146 skipBuffer[0].writeTo(output);
|
/lucene-3.6.0/lucene/contrib/analyzers/common/src/java/org/tartarus/snowball/ |
H A D | TestApp.java | 48 System.err.println("Usage: TestApp <algorithm> <input file> [-o <output file>]"); 80 Writer output = new OutputStreamWriter(outstream); 81 output = new BufferedWriter(output); 98 output.write(stemmer.getCurrent()); 99 output.write('\n'); 106 output.flush();
|
/lucene-3.6.0/lucene/core/src/java/org/apache/lucene/util/ |
H A D | IndexableBinaryStringTools.java | 46 * respectively; similarly, on the output buffer, arrayOffset() is the first 47 * position written to, and limit() is set to one past the final output array 153 * Encodes the input byte sequence into the output char sequence. Before 154 * calling this method, ensure that the output CharBuffer has sufficient 158 * @param output Where the char sequence encoding result will go. The limit is 160 * @throws IllegalArgumentException If either the input or the output buffer 166 public static void encode(ByteBuffer input, CharBuffer output) { argument 167 if (input.hasArray() && output.hasArray()) { 170 final int outputOffset = output.arrayOffset(); 173 output 250 decode(CharBuffer input, ByteBuffer output) argument [all...] |
H A D | BitVector.java | 187 IndexOutput output = d.createOutput(name); 189 output.writeInt(-2); 190 CodecUtil.writeHeader(output, CODEC, VERSION_CURRENT); 192 writeDgaps(output); // sparse bit-set more efficiently saved as d-gaps. 194 writeBits(output); 197 output.close(); 202 private void writeBits(IndexOutput output) throws IOException { argument 203 output.writeInt(size()); // write size 204 output.writeInt(count()); // write count 205 output 209 writeDgaps(IndexOutput output) argument [all...] |
/lucene-3.6.0/lucene/contrib/spellchecker/src/java/org/apache/lucene/search/suggest/fst/ |
H A D | FSTCompletion.java | 130 * Cache the root node's output arcs starting with completions with the 260 final BytesRef output = BytesRef.deepCopyOf(key); 270 output.length = key.length - 1; 271 if (collect(res, num, rootArc.label, output, arc) && !collectAll) { 353 BytesRef output, Arc<Object> arc) throws IOException { 354 if (output.length == output.bytes.length) { 355 output.bytes = ArrayUtil.grow(output.bytes); 357 assert output 352 collect(List<Completion> res, int num, int bucket, BytesRef output, Arc<Object> arc) argument [all...] |
/lucene-3.6.0/lucene/test-framework/src/java/org/apache/lucene/analysis/ |
H A D | BaseTokenStreamTestCase.java | 101 public static void assertTokenStreamContents(TokenStream ts, String[] output, int startOffsets[], int endOffsets[], String types[], int posIncrements[], int posLengths[], Integer finalOffset) throws IOException { argument 102 assertNotNull(output); 133 for (int i = 0; i < output.length; i++) { 146 assertEquals("term "+i, output[i], termAtt.toString()); 192 public static void assertTokenStreamContents(TokenStream ts, String[] output, int startOffsets[], int endOffsets[], String types[], int posIncrements[], Integer finalOffset) throws IOException { argument 193 assertTokenStreamContents(ts, output, startOffsets, endOffsets, types, posIncrements, null, finalOffset); 196 public static void assertTokenStreamContents(TokenStream ts, String[] output, int startOffsets[], int endOffsets[], String types[], int posIncrements[]) throws IOException { argument 197 assertTokenStreamContents(ts, output, startOffsets, endOffsets, types, posIncrements, null, null); 200 public static void assertTokenStreamContents(TokenStream ts, String[] output) throws IOException { argument 201 assertTokenStreamContents(ts, output, nul 204 assertTokenStreamContents(TokenStream ts, String[] output, String[] types) argument 208 assertTokenStreamContents(TokenStream ts, String[] output, int[] posIncrements) argument 212 assertTokenStreamContents(TokenStream ts, String[] output, int startOffsets[], int endOffsets[]) argument 216 assertTokenStreamContents(TokenStream ts, String[] output, int startOffsets[], int endOffsets[], Integer finalOffset) argument 220 assertTokenStreamContents(TokenStream ts, String[] output, int startOffsets[], int endOffsets[], int[] posIncrements) argument 224 assertTokenStreamContents(TokenStream ts, String[] output, int startOffsets[], int endOffsets[], int[] posIncrements, Integer finalOffset) argument 228 assertTokenStreamContents(TokenStream ts, String[] output, int startOffsets[], int endOffsets[], int[] posIncrements, int[] posLengths, Integer finalOffset) argument 232 assertAnalyzesTo(Analyzer a, String input, String[] output, int startOffsets[], int endOffsets[], String types[], int posIncrements[]) argument 236 assertAnalyzesTo(Analyzer a, String input, String[] output, int startOffsets[], int endOffsets[], String types[], int posIncrements[], int posLengths[]) argument 240 assertAnalyzesTo(Analyzer a, String input, String[] output) argument 244 assertAnalyzesTo(Analyzer a, String input, String[] output, String[] types) argument 248 assertAnalyzesTo(Analyzer a, String input, String[] output, int[] posIncrements) argument 252 assertAnalyzesToPositions(Analyzer a, String input, String[] output, int[] posIncrements, int[] posLengths) argument 256 assertAnalyzesTo(Analyzer a, String input, String[] output, int startOffsets[], int endOffsets[]) argument 260 assertAnalyzesTo(Analyzer a, String input, String[] output, int startOffsets[], int endOffsets[], int[] posIncrements) argument 265 assertAnalyzesToReuse(Analyzer a, String input, String[] output, int startOffsets[], int endOffsets[], String types[], int posIncrements[]) argument 269 assertAnalyzesToReuse(Analyzer a, String input, String[] output) argument 273 assertAnalyzesToReuse(Analyzer a, String input, String[] output, String[] types) argument 277 assertAnalyzesToReuse(Analyzer a, String input, String[] output, int[] posIncrements) argument 281 assertAnalyzesToReuse(Analyzer a, String input, String[] output, int startOffsets[], int endOffsets[]) argument 285 assertAnalyzesToReuse(Analyzer a, String input, String[] output, int startOffsets[], int endOffsets[], int[] posIncrements) argument [all...] |
/lucene-3.6.0/lucene/core/src/test/org/apache/lucene/analysis/ |
H A D | TestKeywordMarkerFilter.java | 40 String[] output = new String[] { "the", "quick", "brown", "LuceneFox", 44 "The quIck browN LuceneFox Jumps"), MockTokenizer.WHITESPACE, false), set)), output); 49 "The quIck browN LuceneFox Jumps"), MockTokenizer.WHITESPACE, false), jdkSet)), output); 53 "The quIck browN LuceneFox Jumps"), MockTokenizer.WHITESPACE, false), set2)), output);
|
/lucene-3.6.0/lucene/core/src/test/org/apache/lucene/index/ |
H A D | TestFieldInfos.java | 49 IndexOutput output = dir.createOutput(name); 50 assertTrue(output != null); 53 fieldInfos.write(output); 54 output.close();
|
/lucene-3.6.0/solr/contrib/dataimporthandler/src/test/org/apache/solr/handler/dataimport/ |
H A D | TestDateFormatTransformer.java | 82 List output = new ArrayList(); 83 output.add(now1); 84 output.add(now2); 85 assertEquals(output, row.get("dateAdded"));
|
/lucene-3.6.0/lucene/backwards/src/test/org/apache/lucene/analysis/ |
H A D | TestKeywordMarkerFilter.java | 40 String[] output = new String[] { "the", "quick", "brown", "LuceneFox", 44 "The quIck browN LuceneFox Jumps"), MockTokenizer.WHITESPACE, false), set)), output); 49 "The quIck browN LuceneFox Jumps"), MockTokenizer.WHITESPACE, false), jdkSet)), output); 53 "The quIck browN LuceneFox Jumps"), MockTokenizer.WHITESPACE, false), set2)), output);
|
/lucene-3.6.0/lucene/backwards/src/test/org/apache/lucene/index/ |
H A D | TestFieldInfos.java | 49 IndexOutput output = dir.createOutput(name); 50 assertTrue(output != null); 53 fieldInfos.write(output); 54 output.close();
|
/lucene-3.6.0/lucene/contrib/analyzers/common/src/java/org/apache/lucene/analysis/synonym/ |
H A D | SynonymMap.java | 76 * same output) will be added only once. */ 173 private void add(CharsRef input, int numInputWords, CharsRef output, int numOutputWords, boolean includeOrig) { argument 184 if (output.length <= 0) { 185 throw new IllegalArgumentException("output.length must be > 0 (got " + output.length + ")"); 189 assert !hasHoles(output): "output has holes: " + output; 191 //System.out.println("fmap.add input=" + input + " numInputWords=" + numInputWords + " output=" + output 238 add(CharsRef input, CharsRef output, boolean includeOrig) argument [all...] |