Searched refs:tokens (Results 1 - 5 of 5) sorted by relevance

/pkg/src/util/apache2/depot/
H A Ddepot_index.py371 def repo_index(self, *tokens, **params):
415 def default(self, *tokens, **params):
429 if "_themes" in tokens:
431 theme_index = tokens.index("_themes")
433 tokens[theme_index:])
444 return dh.default(*tokens[theme_index:])
446 elif tokens[0] not in repositories:
451 repo_prefix = tokens[0]
493 def manifest(self, *tokens):
547 def __strip_pub(self, tokens, rep
[all...]
/pkg/src/modules/server/
H A Ddepot.py410 def default(self, *tokens, **params):
417 if not pub and tokens:
418 op = tokens[0]
419 elif pub and len(tokens) > 1:
420 op = tokens[1]
436 ver = int(tokens[1])
438 ver = int(tokens[2])
478 def versions_0(self, *tokens):
571 def catalog_1(self, *tokens):
576 name = tokens[
[all...]
H A Dapi.py437 def search(self, tokens, case_sensitive=False,
441 by 'return_type') matching the specified 'tokens'.
443 'tokens' is a string using pkg(7) query syntax.
447 the provided tokens.
470 if not tokens:
473 tokens = tokens.split()
547 query = qp.Query(" ".join(tokens), case_sensitive,
556 query = qp.Query(" ".join(tokens), case_sensitive,
/pkg/src/modules/
H A Dquery_parser.py55 # These are the types of tokens that the lexer can produce.
56 tokens = ("FTERM", "TERM", "LPAREN", "RPAREN", "AND", "OR", "QUOTE1", variable in class:QueryLexer
66 # This rule causes spaces to break tokens, but the space itself is not
75 # tokens.
181 shows the user exactly which tokens are produced from the input
193 """This class defines the parser which converts a stream of tokens into
197 # Use the same set of tokens as the lexer.
198 tokens = QueryLexer.tokens variable in class:QueryParser
200 # Define the precendence and associativity of certain tokens t
[all...]
/pkg/src/tests/api/
H A Dt_api_search.py653 def _search_op_multi(self, api_obj, remote, tokens, test_value,
657 num_to_return, start_point) for token in tokens]
695 def _search_op_slow_multi(self, api_obj, remote, tokens, test_value,
699 num_to_return, start_point) for token in tokens]

Completed in 26 milliseconds