Searched refs:token (Results 1 - 7 of 7) sorted by relevance
/pkg/src/tests/cli/ |
H A D | t_change_variant.py | 165 def f_verify(self, path, token=None, negate=False): 167 the specified token. If negate is true, then make sure 169 it doesn't contain the specified token.""" 180 if negate and not token: 185 "^" + token + "$" \ 186 "|^" + token + "[/_]" \ 187 "|[/_]" + token + "$" \ 188 "|[/_]" + token + "[/_]") 199 path, file_path, token)) 202 path, file_path, token)) [all...] |
H A D | t_pkg_search.py | 383 def _search_op(self, remote, token, test_value, case_sensitive=False, 388 token = "-r " + token 390 token = "-l " + token 392 token = "-I " + token 394 token = "-a " + token 396 token [all...] |
/pkg/src/modules/ |
H A D | search_storage.py | 219 # Each line begins with a urllib quoted search token. It's followed by 224 # matched the token. It's separated from its sublist by a '#'. The 225 # next token (579) is the fmri id. The subsequent comma separated 227 # that token. 299 """Pulls the token out of a line from a main dictionary file. 309 def transform_main_dict_line(token, entries): 310 """Paired with parse_main_dict_line above. Transforms a token 314 The "token" parameter is the token whose index line is being 318 It contains information about where and how "token" wa [all...] |
H A D | query_parser.py | 67 # reported as a token. 100 # all colons are escaped, it produces a TERM token whose value 102 # there are unescaped colons, it produces an FTERM token whose 104 # type, the action key, and the token that followed the last 113 # the escaped colon in a token. The colon is matching either a 115 # token or field. The part after the colon is attempting to 128 token = fields[-1] 134 t.type = self.reserved.get(token, "TERM") 135 t.value = token 147 t.value = (pkg_name, action_type, key, token) 170 def token(self): member in class:QueryLexer [all...] |
H A D | indexer.py | 385 def _write_main_dict_line(self, file_handle, token, 388 token offsets to _data_token_offset. file_handle is the file 389 handle for the output main dictionary file. token is the token 394 the values which matched the token for the action and key it's 403 self.old_out_token >= token: 404 raise RuntimeError("In writing dict line, token:{0}, " 405 "old_out_token:{1}".format(token, 407 self.old_out_token = token 411 self._data_token_offset.write_entity(token, cur_locatio [all...] |
/pkg/src/modules/flavor/ |
H A D | elf.py | 50 token in their path that we're unable to decode.""" 52 def __init__(self, proto_path, installed_path, run_path, token): 57 self.tok = token 61 "token, {tok}, in its run path: {rp}. It is not " 62 "currently possible to automatically expand this token. " 118 paramter 'paths' with the values for that token provided in the 139 # The first dynamic token has been replaced, but
|
/pkg/src/tests/api/ |
H A D | t_api_search.py | 645 def _search_op(self, api_obj, remote, token, test_value, 648 query = [api.Query(token, case_sensitive, return_actions, 656 query = [api.Query(token, case_sensitive, return_actions, 657 num_to_return, start_point) for token in tokens] 687 def _search_op_slow(self, api_obj, remote, token, test_value, 690 query = [api.Query(token, case_sensitive, return_actions, 698 query = [api.Query(token, case_sensitive, return_actions, 699 num_to_return, start_point) for token in tokens] 818 # token isn't in the packages. 1014 # token is [all...] |
Completed in 56 milliseconds