/lucene-3.6.0/lucene/contrib/queryparser/src/java/org/apache/lucene/queryParser/core/util/ |
H A D | UnescapedCharSequence.java | 21 * CharsSequence with escaped chars information. 24 private char[] chars; field in class:UnescapedCharSequence 31 * @param chars 36 public UnescapedCharSequence(char[] chars, boolean[] wasEscaped, int offset, argument 38 this.chars = new char[length]; 40 System.arraycopy(chars, offset, this.chars, 0, length); 50 this.chars = new char[text.length()]; 53 this.chars[i] = text.charAt(i); 65 this.chars [all...] |
/lucene-3.6.0/lucene/core/src/java/org/apache/lucene/util/ |
H A D | CharsRef.java | 24 * The {@link #chars} member should never be null; use 30 public char[] chars; field in class:CharsRef 46 chars = new char[capacity]; 53 public CharsRef(char[] chars, int offset, int length) { argument 54 assert chars != null; 55 assert chars.length >= offset + length; 56 this.chars = chars; 66 this.chars = string.toCharArray(); 68 this.length = chars [all...] |
H A D | CharacterUtils.java | 56 * @param chars 59 * the offset to the char values in the chars array to be converted 68 public abstract int codePointAt(final char[] chars, final int offset); argument 80 * the offset to the char values in the chars array to be converted 99 * @param chars 102 * the offset to the char values in the chars array to be converted 113 public abstract int codePointAt(final char[] chars, final int offset, final int limit); argument 163 public int codePointAt(final char[] chars, final int offset) { argument 164 return Character.codePointAt(chars, offset); 173 public int codePointAt(final char[] chars, fina argument 235 codePointAt(final char[] chars, final int offset) argument 245 codePointAt(final char[] chars, final int offset, final int limit) argument [all...] |
H A D | UnicodeUtil.java | 157 * offset for length chars. Returns a hash of the resulting bytes. After encoding, result.offset will always be 0. */ 268 * offset for length chars. Returns the number of bytes 426 * offset for length chars. After encoding, result.offset will always be 0. 676 char[] chars = new char[count]; 686 chars[w] = (char) cp; 689 chars[w] = (char) (LEAD_SURROGATE_OFFSET_ + (cp >> LEAD_SURROGATE_SHIFT_)); 690 chars[w + 1] = (char) (TRAIL_SURROGATE_MIN_VALUE + (cp & TRAIL_SURROGATE_MASK_)); 698 System.arraycopy(chars, 0, temp, 0, w); 699 chars = temp; 703 return new String(chars, 715 UTF8toUTF16(byte[] utf8, int offset, int length, CharsRef chars) argument 749 UTF8toUTF16(BytesRef bytesRef, CharsRef chars) argument [all...] |
/lucene-3.6.0/lucene/contrib/facet/src/java/org/apache/lucene/facet/taxonomy/ |
H A D | CategoryPath.java | 40 // internally as one large character array "chars" with all the string 48 protected char[] chars; field in class:CategoryPath 83 return chars.length; 106 chars = new char[capacityChars]; 139 // Copy the new component's characters, increasing the "chars" array 141 if (ends[ncomponents] > chars.length) { 143 System.arraycopy(chars, 0, newchars, 0, chars.length); 144 chars = newchars; 147 chars[preven [all...] |
/lucene-3.6.0/lucene/contrib/facet/src/java/org/apache/lucene/facet/taxonomy/writercache/cl2o/ |
H A D | CharBlockArray.java | 44 char[] chars; field in class:CharBlockArray.Block 48 this.chars = new char[size]; 81 public CharBlockArray append(CharSequence chars) { argument 82 return append(chars, 0, chars.length()); 89 this.current.chars[this.current.length++] = c; 95 public CharBlockArray append(CharSequence chars, int start, int length) { argument 98 append(chars.charAt(i)); 103 public CharBlockArray append(char[] chars, int start, int length) { argument 115 System.arraycopy(chars, offse [all...] |
/lucene-3.6.0/lucene/contrib/analyzers/common/src/java/org/apache/lucene/analysis/synonym/ |
H A D | SynonymMap.java | 88 * SynonymMap#WORD_SEPARATOR}. reuse and its chars 92 char[] buffer = reuse.chars; 97 buffer = reuse.chars; 106 buffer = reuse.chars; 118 * reuse and its chars must not be null. */ 136 reuse.chars[end++] = SynonymMap.WORD_SEPARATOR; 139 System.arraycopy(termAtt.buffer(), 0, reuse.chars, end, length); 151 private boolean hasHoles(CharsRef chars) { argument 152 final int end = chars.offset + chars 215 countWords(CharsRef chars) argument [all...] |
/lucene-3.6.0/solr/core/src/java/org/apache/solr/highlight/ |
H A D | SimpleBoundaryScanner.java | 30 Character[] chars = new Character[str.length()]; 32 chars[i] = str.charAt(i); 34 return new org.apache.lucene.search.vectorhighlight.SimpleBoundaryScanner(maxScan, chars);
|
/lucene-3.6.0/lucene/contrib/analyzers/kuromoji/src/test/org/apache/lucene/analysis/ja/dict/ |
H A D | TestTokenInfoDictionary.java | 46 char chars[] = new char[input.length]; 47 for (int i = 0; i < chars.length; i++) { 48 chars[i] = (char)input.ints[input.offset+i]; 50 assertTrue(UnicodeUtil.validUTF16String(new String(chars))); 64 String baseForm = tid.getBaseForm(wordId, chars, 0, chars.length); 94 String pronunciation = tid.getPronunciation(wordId, chars, 0, chars.length); 98 String reading = tid.getReading(wordId, chars, 0, chars [all...] |
/lucene-3.6.0/lucene/contrib/analyzers/common/src/java/org/apache/lucene/analysis/ngram/ |
H A D | NGramTokenizer.java | 107 char[] chars = new char[1024]; 110 while (charsRead < chars.length) { 111 int inc = input.read(chars, charsRead, chars.length-charsRead); 117 inStr = new String(chars, 0, charsRead).trim(); // remove any trailing empty strings 119 if (charsRead == chars.length) { 120 // Read extra throwaway chars so that on end() we
|
H A D | EdgeNGramTokenizer.java | 187 char[] chars = new char[1024]; 190 while (charsRead < chars.length) { 191 final int inc = input.read(chars, charsRead, chars.length-charsRead); 198 inStr = new String(chars, 0, charsRead).trim(); // remove any trailing empty strings 200 if (charsRead == chars.length) { 201 // Read extra throwaway chars so that on end() we 229 // grab gramSize chars from front or back
|
/lucene-3.6.0/lucene/contrib/analyzers/common/src/java/org/apache/lucene/analysis/compound/hyphenation/ |
H A D | PatternParser.java | 145 protected String readToken(StringBuffer chars) { argument 149 for (i = 0; i < chars.length(); i++) { 150 if (Character.isWhitespace(chars.charAt(i))) { 157 // chars.delete(0,i); 158 for (int countr = i; countr < chars.length(); countr++) { 159 chars.setCharAt(countr - i, chars.charAt(countr)); 161 chars.setLength(chars.length() - i); 169 for (i = 0; i < chars [all...] |
/lucene-3.6.0/lucene/backwards/src/test/org/apache/lucene/index/ |
H A D | TestIndexInput.java | 127 theBytes[bytes.length] = (byte)utf8Str.length();//Add in the number of chars we are storing, which should fit in a byte for this test 134 int charsToRead = is.readVInt();//number of chars in the Lucene string 137 char [] chars = new char[3];//there should be 6 chars remaining 138 is.readChars(chars, 0, 3); 139 String tmpStr = new String(chars); 145 chars = new char[charsToRead]; 146 is.readChars(chars, 0, charsToRead); 147 tmpStr = new String(chars);
|
H A D | TestIndexWriterUnicode.java | 166 char[] chars = new char[2]; 175 chars[len++] = (char) ch; 177 chars[len++] = (char) (((ch-0x0010000) >> 10) + UnicodeUtil.UNI_SUR_HIGH_START); 178 chars[len++] = (char) (((ch-0x0010000) & 0x3FFL) + UnicodeUtil.UNI_SUR_LOW_START); 181 UnicodeUtil.UTF16toUTF8(chars, 0, len, utf8); 183 String s1 = new String(chars, 0, len);
|
/lucene-3.6.0/solr/core/src/java/org/apache/solr/internal/csv/ |
H A D | CharBuffer.java | 180 char[] chars = new char[length]; 181 System.arraycopy(c, 0, chars, 0, length); 182 return chars;
|
/lucene-3.6.0/solr/core/src/java/org/apache/solr/analysis/ |
H A D | JapanesePartOfSpeechStopFilterFactory.java | 54 char chars[] = (char[]) element; 55 stopTags.add(new String(chars));
|
/lucene-3.6.0/lucene/core/src/test/org/apache/lucene/analysis/ |
H A D | TestClassicAnalyzer.java | 54 char[] chars = new char[255]; 56 chars[i] = 'a'; 57 String longTerm = new String(chars, 0, 255); 252 char[] chars = new char[IndexWriter.MAX_TERM_LENGTH]; 253 Arrays.fill(chars, 'x'); 255 final String bigTerm = new String(chars);
|
/lucene-3.6.0/lucene/backwards/src/test/org/apache/lucene/analysis/ |
H A D | TestClassicAnalyzer.java | 54 char[] chars = new char[255]; 56 chars[i] = 'a'; 57 String longTerm = new String(chars, 0, 255); 252 char[] chars = new char[IndexWriter.MAX_TERM_LENGTH]; 253 Arrays.fill(chars, 'x'); 255 final String bigTerm = new String(chars);
|
/lucene-3.6.0/solr/core/src/java/org/apache/solr/search/ |
H A D | SolrQueryParser.java | 241 char[] chars = new char[len+1]; 242 chars[0] = factory.getMarkerChar(); 243 termStr.getChars(0, len, chars, 1); 244 ReversedWildcardFilter.reverse(chars, 1, len); 245 termStr = new String(chars);
|
/lucene-3.6.0/lucene/core/src/java/org/apache/lucene/store/ |
H A D | DataInput.java | 35 * in the old modified UTF8 format (length in java chars 192 final char[] chars = new char[length]; 193 readChars(chars, 0, length); 194 return new String(chars, 0, length);
|
/lucene-3.6.0/lucene/test-framework/src/java/org/apache/lucene/analysis/ |
H A D | MockTokenizer.java | 104 char chars[] = Character.toChars(normalize(cp)); 105 for (int i = 0; i < chars.length; i++) 106 termAtt.append(chars[i]);
|
/lucene-3.6.0/lucene/backwards/src/test-framework/java/org/apache/lucene/analysis/ |
H A D | MockTokenizer.java | 102 char chars[] = Character.toChars(normalize(cp)); 103 for (int i = 0; i < chars.length; i++) 104 termAtt.append(chars[i]);
|
/lucene-3.6.0/lucene/core/src/test/org/apache/lucene/index/ |
H A D | TestIndexInput.java | 220 theBytes[bytes.length] = (byte)utf8Str.length();//Add in the number of chars we are storing, which should fit in a byte for this test 227 int charsToRead = is.readVInt();//number of chars in the Lucene string 230 char [] chars = new char[3];//there should be 6 chars remaining 231 is.readChars(chars, 0, 3); 232 String tmpStr = new String(chars); 238 chars = new char[charsToRead]; 239 is.readChars(chars, 0, charsToRead); 240 tmpStr = new String(chars);
|
H A D | TestIndexWriterUnicode.java | 166 char[] chars = new char[2]; 175 chars[len++] = (char) ch; 177 chars[len++] = (char) (((ch-0x0010000) >> 10) + UnicodeUtil.UNI_SUR_HIGH_START); 178 chars[len++] = (char) (((ch-0x0010000) & 0x3FFL) + UnicodeUtil.UNI_SUR_LOW_START); 181 UnicodeUtil.UTF16toUTF8(chars, 0, len, utf8); 183 String s1 = new String(chars, 0, len);
|
/lucene-3.6.0/lucene/contrib/analyzers/kuromoji/src/java/org/apache/lucene/analysis/ja/ |
H A D | JapaneseAnalyzer.java | 78 char chars[] = (char[]) element; 79 DEFAULT_STOP_TAGS.add(new String(chars));
|