/netbsd/external/bsd/nsd/dist/contrib/bind2nsd/bind2nsd/ |
H A D | Parser.py | 62 def handle_include(self, tokens): argument 79 def handle_key(self, tokens): argument 114 def handle_server(self, tokens): argument 306 self.skip_value(val, tokens) 327 def handle_zone(self, tokens): argument 414 def parse(self, tokens): argument 424 self.handle_key(tokens) 427 self.handle_include(tokens) 436 self.handle_options(tokens) 439 self.handle_server(tokens) [all …]
|
/netbsd/external/bsd/ntp/dist/libjsmn/ |
H A D | jsmn.c | 68 if (tokens == NULL) { in jsmn_parse_primitive() 102 if (tokens == NULL) { in jsmn_parse_string() 170 if (tokens == NULL) { in jsmn_parse() 187 if (tokens == NULL) in jsmn_parse() 211 token = &tokens[i]; in jsmn_parse() 224 token = &tokens[i]; in jsmn_parse() 245 if (tokens != NULL && in jsmn_parse() 252 if (tokens[i].type == JSMN_ARRAY || tokens[i].type == JSMN_OBJECT) { in jsmn_parse() 253 if (tokens[i].start != -1 && tokens[i].end == -1) { in jsmn_parse() 293 if (tokens != NULL) { in jsmn_parse() [all …]
|
H A D | jsmn_test.c | 88 jsmntok_t tokens[10]; in test_simple() local 337 jsmntok_t tokens[10]; in test_objects_arrays() local 366 jsmntok_t tokens[128]; in test_issue_22() local 381 for (i = 1; tokens[i].end < tokens[0].end; i++) { in test_issue_22() 382 if (tokens[i].type == JSMN_STRING || tokens[i].type == JSMN_PRIMITIVE) { in test_issue_22() 383 printf("%.*s\n", tokens[i].end - tokens[i].start, js + tokens[i].start); in test_issue_22() 398 jsmntok_t tokens[10]; in test_unicode_characters() local 444 jsmntok_t tokens[10]; in test_input_length() local 509 jsmntok_t tokens[10]; in test_keyvalue() local 517 check(tokens[1].size == 1 && tokens[3].size == 1); /* one value per key */ in test_keyvalue() [all …]
|
H A D | README.md | 51 It holds the following tokens: 58 string instead. In the example above jsmn will create tokens like: Object 71 Besides start/end positions, jsmn tokens for complex types (like arrays 106 **Note:** Unlike JSON data types, primitive tokens are not divided into 120 int size; // Number of child (nested) tokens 123 **Note:** string tokens point to the first character after 130 jsmntok_t tokens[10]; 135 // tokens - an array of tokens available 136 // 10 - number of tokens available 137 jsmn_parse(&parser, js, tokens, 10); [all …]
|
/netbsd/external/bsd/tradcpp/dist/ |
H A D | eval.c | 63 enum tokens { enum 94 enum tokens tok; 109 enum tokens tok; 132 enum tokens tok; 171 num = tokenarray_num(&tokens); in printtokens() 209 isuop(enum tokens tok) in isuop() 225 isbop(enum tokens tok) in isbop() 245 isop(enum tokens tok) in isop() 261 getprec(enum tokens tok) in getprec() 282 looser(enum tokens t1, enum tokens t2) in looser() [all …]
|
/netbsd/external/apache2/llvm/dist/clang/bindings/python/tests/cindex/ |
H A D | test_tokens.py | 22 tokens = list(tu.get_tokens(extent=r)) 24 self.assertEqual(len(tokens), 4) 25 self.assertEqual(tokens[1].spelling, 'i') 28 cursor = tokens[1].cursor 30 self.assertEqual(tokens[1].cursor, tokens[2].cursor) 38 tokens = list(tu.get_tokens(extent=r)) 39 self.assertEqual(len(tokens), 4) 41 loc = tokens[1].location 52 tokens = list(tu.get_tokens(extent=r)) 53 self.assertEqual(len(tokens), 4) [all …]
|
/netbsd/crypto/external/bsd/openssl.old/dist/test/recipes/ |
H A D | 02-test_ordinals.t | 30 my @tokens = split(/(?:\s+|\s*:\s*)/, $line); 32 if ($#tokens < 5 || $#tokens > 6) { 37 if ($tokens[3] eq "NOEXIST") { 42 $newqual = $tokens[4]; 44 if ($cnt > $tokens[1] 45 || ($cnt == $tokens[1] && ($qualifier ne $newqual 47 print STDERR "Invalid ordinal detected: ".$tokens[1]."\n"; 51 $cnt = $tokens[1]; 53 $lastfunc = $tokens[0];
|
/netbsd/crypto/external/bsd/openssl/dist/test/recipes/ |
H A D | 02-test_ordinals.t | 30 my @tokens = split(/(?:\s+|\s*:\s*)/, $line); 32 if ($#tokens < 5 || $#tokens > 6) { 37 if ($tokens[3] eq "NOEXIST") { 42 $newqual = $tokens[4]; 44 my $number = $tokens[1]; 50 print STDERR "Invalid ordinal detected: ".$tokens[1]."\n"; 54 $cnt = $tokens[1]; 56 $lastfunc = $tokens[0];
|
/netbsd/sys/dev/microcode/siop/ |
H A D | ncr53cxxx.c | 49 } tokens[MAXTOKENS]; variable 437 if (tokens[i].type) in main() 600 tokens[tokenix].type); in process() 610 tokens[tokenix].name); in process() 712 if (tokens[t].type == 0 && strcmpi (tokens[t].name, string) == 0) in reserved() 769 if (tokens[t].type) { in evaluate() 773 name = tokens[t].name; in evaluate() 868 setarch(tokens[i].name); in f_arch() 877 if (tokens[tokenix].type != 0 || tokens[tokenix + 1].type != ':') in f_proc() 929 if (tokens[i].type) { in f_define() [all …]
|
/netbsd/lib/libc/locale/ |
H A D | generic_lc_all.c | 59 const char *tokens[_LC_LAST], *s, *t; in _generic_LC_ALL_setlocale() local 67 tokens[1] = &head[0]; in _generic_LC_ALL_setlocale() 68 tail = strchr(tokens[1], '/'); in _generic_LC_ALL_setlocale() 71 tokens[i] = tokens[1]; in _generic_LC_ALL_setlocale() 75 tokens[i] = (const char *)tail; in _generic_LC_ALL_setlocale() 76 tail = strchr(tokens[i], '/'); in _generic_LC_ALL_setlocale() 81 tokens[_LC_LAST - 1] = (const char *)tail; in _generic_LC_ALL_setlocale() 82 tail = strchr(tokens[i], '/'); in _generic_LC_ALL_setlocale() 86 if ((*sl)(tokens[1], locale) != NULL) in _generic_LC_ALL_setlocale() 96 if ((*sl)(tokens[i], locale) != NULL) in _generic_LC_ALL_setlocale()
|
H A D | newlocale.c | 48 const char *tokens[_LC_LAST - 1]; in __weak_alias() local 64 tokens[0] = (const char *)&head[0]; in __weak_alias() 65 tail = strchr(tokens[0], '/'); in __weak_alias() 71 (*l)(tokens[0], dst); in __weak_alias() 84 tokens[i] = (const char *)tail; in __weak_alias() 85 tail = strchr(tokens[i], '/'); in __weak_alias() 92 tokens[howmany] = tail; in __weak_alias() 93 tail = strchr(tokens[howmany], '/'); in __weak_alias() 101 (*l)(tokens[i], dst); in __weak_alias()
|
/netbsd/external/mpl/bind/dist/bin/pkcs11/ |
H A D | Makefile.in | 34 pkcs11-keygen@EXEEXT@ pkcs11-tokens@EXEEXT@ 36 pkcs11-keygen.c pkcs11-tokens.c 38 pkcs11-keygen.@O@ pkcs11-tokens.@O@ 54 pkcs11-tokens@EXEEXT@: pkcs11-tokens.@O@ ${DEPLIBS} 56 -o $@ pkcs11-tokens.@O@ ${LIBS} 68 ${LIBTOOL_MODE_INSTALL} ${INSTALL_PROGRAM} pkcs11-tokens@EXEEXT@ \ 73 ${DESTDIR}${sbindir}/pkcs11-tokens@EXEEXT@
|
H A D | pkcs11-tokens.rst | 14 .. _man_pkcs11-tokens: 16 pkcs11-tokens - list PKCS#11 available tokens 22 :program:`pkcs11-tokens` [**-m** module] [**-v**] 27 ``pkcs11-tokens`` lists the PKCS#11 available tokens with defaults from
|
/netbsd/lib/libc/stdlib/ |
H A D | getsubopt.c | 65 getsubopt(char **optionp, char * const *tokens, char **valuep) 70 _DIAGASSERT(tokens != NULL); 112 for (cnt = 0; *tokens; ++tokens, ++cnt) 113 if (!strcmp(suboptarg, *tokens))
|
/netbsd/external/apache2/llvm/dist/llvm/tools/llvm-c-test/ |
H A D | calc.c | 41 static LLVMValueRef build_from_tokens(char **tokens, int ntokens, in build_from_tokens() argument 49 char tok = tokens[i][0]; in build_from_tokens() 85 long val = strtol(tokens[i], &end, 0); in build_from_tokens() 112 static void handle_line(char **tokens, int ntokens) { in handle_line() argument 113 char *name = tokens[0]; in handle_line() 130 res = build_from_tokens(tokens + 1, ntokens - 1, builder, param); in handle_line()
|
H A D | disassemble.c | 65 static void handle_line(char **tokens, int ntokens) { in handle_line() argument 68 const char *triple = tokens[0]; in handle_line() 69 const char *features = tokens[1]; in handle_line() 77 disbuf[disbuflen++] = strtol(tokens[i], NULL, 16); in handle_line()
|
/netbsd/external/apache2/llvm/dist/clang/utils/ |
H A D | token-delta.py | 117 tokens = [] 128 tokens.append(Token(*kTokenRE.match(ln).groups())) 130 return tokens 155 for i,(file,tokens) in enumerate(self.tokenLists): 158 f.write(tokens[j]) 209 for i,(file,tokens) in enumerate(self.tokenLists) 210 for j in range(len(tokens))]) 221 numTokens = sum([len(tokens) for _,tokens in tokenLists])
|
/netbsd/external/bsd/mdocml/dist/ |
H A D | compat_getsubopt.c | 46 getsubopt(char **optionp, char * const *tokens, char **valuep) in getsubopt() argument 90 for (cnt = 0; *tokens; ++tokens, ++cnt) in getsubopt() 91 if (!strcmp(suboptarg, *tokens)) in getsubopt()
|
/netbsd/external/bsd/ipf/dist/perl/ |
H A D | ipfmeta.pl | 19 my @tokens; 26 push (@tokens, split); 32 while (@tokens) { 33 my $token = shift(@tokens);
|
/netbsd/external/mit/libuv/dist/test/ |
H A D | test-strtok.c | 34 const char* tokens[] = { variable 68 size_t tokens_len = ARRAY_SIZE(tokens); in TEST_IMPL() 80 ASSERT_STRCMP(tok_r, tokens[j]); in TEST_IMPL() 85 ASSERT_STRCMP(tok_r, tokens[j]); in TEST_IMPL()
|
/netbsd/external/bsd/wpa/dist/src/utils/ |
H A D | json.c | 209 static struct json_token * json_alloc_token(unsigned int *tokens) in json_alloc_token() argument 211 (*tokens)++; in json_alloc_token() 212 if (*tokens > JSON_MAX_TOKENS) { in json_alloc_token() 227 unsigned int tokens = 0; in json_parse() local 237 token = json_alloc_token(&tokens); in json_parse() 262 token->child = json_alloc_token(&tokens); in json_parse() 302 token = json_alloc_token(&tokens); in json_parse() 351 curr_token->sibling = json_alloc_token(&tokens); in json_parse() 372 token = json_alloc_token(&tokens); in json_parse() 424 token = json_alloc_token(&tokens); in json_parse()
|
/netbsd/external/gpl3/gcc.old/dist/gcc/doc/ |
H A D | cppinternals.texi | 140 returns preprocessing tokens individually, not a line at a time. 360 in the directive's tokens if necessary. 396 calling @code{_cpp_lex_direct} to lex new tokens, or returning 514 unexpanded tokens. 523 context, cpplib reads tokens directly from the lexer. 587 argument only uses tokens from that argument; it cannot take tokens 609 any intervening padding tokens. 620 @section Marking tokens ineligible for future expansion 621 As discussed above, cpplib needs a way of marking tokens as 627 remember to free these tokens, they are allocated as temporary tokens [all …]
|
H A D | cppinternals.info | 327 directive's tokens if necessary. 364 calling '_cpp_lex_direct' to lex new tokens, or returning 481 in which case it holds that argument's unexpanded tokens. 490 reads tokens directly from the lexer. 553 tokens from that argument; it cannot take tokens from whatever follows 576 padding tokens. 587 Marking tokens ineligible for future expansion 590 As discussed above, cpplib needs a way of marking tokens as 596 to remember to free these tokens, they are allocated as temporary tokens 599 current line of tokens has been read in. [all …]
|
/netbsd/external/gpl3/gcc/dist/gcc/doc/ |
H A D | cppinternals.info | 327 directive's tokens if necessary. 364 calling '_cpp_lex_direct' to lex new tokens, or returning 481 in which case it holds that argument's unexpanded tokens. 490 reads tokens directly from the lexer. 553 tokens from that argument; it cannot take tokens from whatever follows 576 padding tokens. 587 Marking tokens ineligible for future expansion 590 As discussed above, cpplib needs a way of marking tokens as 596 to remember to free these tokens, they are allocated as temporary tokens 599 current line of tokens has been read in. [all …]
|
H A D | cppinternals.texi | 140 returns preprocessing tokens individually, not a line at a time. 360 in the directive's tokens if necessary. 396 calling @code{_cpp_lex_direct} to lex new tokens, or returning 514 unexpanded tokens. 523 context, cpplib reads tokens directly from the lexer. 587 argument only uses tokens from that argument; it cannot take tokens 609 any intervening padding tokens. 620 @section Marking tokens ineligible for future expansion 621 As discussed above, cpplib needs a way of marking tokens as 627 remember to free these tokens, they are allocated as temporary tokens [all …]
|