/dports/print/a2ps/a2ps-4.13/src/ |
H A D | sheets-map.l | 54 static char *token_val; variable 99 {key} token_val = xstrndup (yytext, yyleng - 1); return tKEY; 127 token_val = xustrdup (string); 167 token_val = xustrdup (string); 286 key = token_val; 292 _("no key defined for `%s'"), quotearg (token_val)); 293 sheets_map_add (token_val, false, insensitive_p, key); 299 _("no key defined for `%s'"), quotearg (token_val)); 300 sheets_map_add (token_val, true, insensitive_p, key);
|
/dports/databases/percona57-server/percona-server-5.7.36-39/plugin/version_token/ |
H A D | version_token.cc | 50 LEX_STRING token_val; member 210 if ((token_obj->token_val).str) in set_vtoken_string_length() 297 LEX_STRING token_name, token_val; in parse_vtokens() local 306 token_val.str= lasts_val; in parse_vtokens() 373 val_len= token_val.length; in parse_vtokens() 386 memcpy(val, token_val.str, val_len); in parse_vtokens() 388 v_token->token_val.str= val; in parse_vtokens() 390 v_token->token_val.length= val_len; in parse_vtokens() 421 if ((token_obj->token_val.length != token_val.length) || in parse_vtokens() 422 (memcmp(token_obj->token_val.str, token_val.str, token_val.length) != 0)) in parse_vtokens() [all …]
|
/dports/databases/xtrabackup/percona-xtrabackup-2.4.21/plugin/version_token/ |
H A D | version_token.cc | 50 LEX_STRING token_val; member 210 if ((token_obj->token_val).str) in set_vtoken_string_length() 297 LEX_STRING token_name, token_val; in parse_vtokens() local 306 token_val.str= lasts_val; in parse_vtokens() 373 val_len= token_val.length; in parse_vtokens() 386 memcpy(val, token_val.str, val_len); in parse_vtokens() 388 v_token->token_val.str= val; in parse_vtokens() 390 v_token->token_val.length= val_len; in parse_vtokens() 421 if ((token_obj->token_val.length != token_val.length) || in parse_vtokens() 422 (memcmp(token_obj->token_val.str, token_val.str, token_val.length) != 0)) in parse_vtokens() [all …]
|
/dports/databases/mysqlwsrep57-server/mysql-wsrep-wsrep_5.7.35-25.27/plugin/version_token/ |
H A D | version_token.cc | 50 LEX_STRING token_val; member 210 if ((token_obj->token_val).str) in set_vtoken_string_length() 297 LEX_STRING token_name, token_val; in parse_vtokens() local 306 token_val.str= lasts_val; in parse_vtokens() 373 val_len= token_val.length; in parse_vtokens() 386 memcpy(val, token_val.str, val_len); in parse_vtokens() 388 v_token->token_val.str= val; in parse_vtokens() 390 v_token->token_val.length= val_len; in parse_vtokens() 421 if ((token_obj->token_val.length != token_val.length) || in parse_vtokens() 422 (memcmp(token_obj->token_val.str, token_val.str, token_val.length) != 0)) in parse_vtokens() [all …]
|
/dports/databases/percona57-pam-for-mysql/percona-server-5.7.36-39/plugin/version_token/ |
H A D | version_token.cc | 50 LEX_STRING token_val; member 210 if ((token_obj->token_val).str) in set_vtoken_string_length() 297 LEX_STRING token_name, token_val; in parse_vtokens() local 306 token_val.str= lasts_val; in parse_vtokens() 373 val_len= token_val.length; in parse_vtokens() 386 memcpy(val, token_val.str, val_len); in parse_vtokens() 388 v_token->token_val.str= val; in parse_vtokens() 390 v_token->token_val.length= val_len; in parse_vtokens() 421 if ((token_obj->token_val.length != token_val.length) || in parse_vtokens() 422 (memcmp(token_obj->token_val.str, token_val.str, token_val.length) != 0)) in parse_vtokens() [all …]
|
/dports/databases/percona57-client/percona-server-5.7.36-39/plugin/version_token/ |
H A D | version_token.cc | 50 LEX_STRING token_val; member 210 if ((token_obj->token_val).str) in set_vtoken_string_length() 297 LEX_STRING token_name, token_val; in parse_vtokens() local 306 token_val.str= lasts_val; in parse_vtokens() 373 val_len= token_val.length; in parse_vtokens() 386 memcpy(val, token_val.str, val_len); in parse_vtokens() 388 v_token->token_val.str= val; in parse_vtokens() 390 v_token->token_val.length= val_len; in parse_vtokens() 421 if ((token_obj->token_val.length != token_val.length) || in parse_vtokens() 422 (memcmp(token_obj->token_val.str, token_val.str, token_val.length) != 0)) in parse_vtokens() [all …]
|
/dports/databases/mysql57-client/mysql-5.7.36/plugin/version_token/ |
H A D | version_token.cc | 50 LEX_STRING token_val; member 210 if ((token_obj->token_val).str) in set_vtoken_string_length() 297 LEX_STRING token_name, token_val; in parse_vtokens() local 306 token_val.str= lasts_val; in parse_vtokens() 373 val_len= token_val.length; in parse_vtokens() 386 memcpy(val, token_val.str, val_len); in parse_vtokens() 388 v_token->token_val.str= val; in parse_vtokens() 390 v_token->token_val.length= val_len; in parse_vtokens() 421 if ((token_obj->token_val.length != token_val.length) || in parse_vtokens() 422 (memcmp(token_obj->token_val.str, token_val.str, token_val.length) != 0)) in parse_vtokens() [all …]
|
/dports/security/pamtester/pamtester-0.1.2/src/ |
H A D | expr_parser.c | 75 …r(pamtester_expr_parser_t *parser, pamtester_token_t token, const char *token_val, size_t token_va… in pamtester_expr_parser_hdlr() argument 83 parser->last_error = build_error_message("unexpected token", token_val, token_val_len); in pamtester_expr_parser_hdlr() 87 val = strtol(token_val, &next_ptr, 10); in pamtester_expr_parser_hdlr() 89 if (next_ptr - token_val != token_val_len in pamtester_expr_parser_hdlr() 92 || parser->const_resolve_hdlr(parser->hdlr_param, &val, token_val, token_val_len)) { in pamtester_expr_parser_hdlr() 93 parser->last_error = build_error_message("undefined constant", token_val, token_val_len); in pamtester_expr_parser_hdlr() 111 parser->last_error = build_error_message("unexpected token", token_val, token_val_len); in pamtester_expr_parser_hdlr() 147 parser->last_error = build_error_message("unexpected token", token_val, token_val_len); in pamtester_expr_parser_hdlr() 179 parser->last_error = build_error_message("unexpected token", token_val, token_val_len); in pamtester_expr_parser_hdlr()
|
/dports/biology/lamarc/lamarc-2.1.8/boost/wave/grammars/ |
H A D | cpp_chlit_grammar.hpp | 292 typename TokenT::string_type const &token_val = token.get_value(); in evaluate() local 294 parse(token_val.begin(), token_val.end(), g[spirit_assign_actor(result)]); in evaluate() 298 token_val.c_str(), token.get_position()); in evaluate() 302 if ('L' == token_val[0]) { in evaluate() 310 token_val.c_str(), token.get_position()); in evaluate() 321 token_val.c_str(), token.get_position()); in evaluate()
|
H A D | cpp_intlit_grammar.hpp | 164 typename TokenT::string_type const &token_val = token.get_value(); in evaluate() local 166 parse(token_val.begin(), token_val.end(), g[spirit_assign_actor(result)]); in evaluate() 170 token_val.c_str(), token.get_position()); in evaluate()
|
/dports/devel/boost-docs/boost_1_72_0/libs/wave/samples/list_includes/lexertl/ |
H A D | lexertl_lexer.hpp | 610 string_type token_val; in get() local 612 wave::token_id id = lexer_.next_token(first, last, token_val); in get() 623 validate_identifier_name(token_val, in get() 634 token_val = convert_trigraphs(token_val); in get() 638 validate_literal(token_val, in get() 648 invalid_long_long_literal, token_val.c_str(), in get() 663 if (0 == token_val.compare(start, 12, "include_next", 12)) in get() 673 token_val.clear(); in get() 688 token_val = convert_trigraph(token_val); in get() 702 sprintf(buffer, "%02x'", token_val[0]); in get() [all …]
|
/dports/devel/boost-python-libs/boost_1_72_0/libs/wave/samples/list_includes/lexertl/ |
H A D | lexertl_lexer.hpp | 610 string_type token_val; in get() local 612 wave::token_id id = lexer_.next_token(first, last, token_val); in get() 623 validate_identifier_name(token_val, in get() 634 token_val = convert_trigraphs(token_val); in get() 638 validate_literal(token_val, in get() 648 invalid_long_long_literal, token_val.c_str(), in get() 663 if (0 == token_val.compare(start, 12, "include_next", 12)) in get() 673 token_val.clear(); in get() 688 token_val = convert_trigraph(token_val); in get() 702 sprintf(buffer, "%02x'", token_val[0]); in get() [all …]
|
/dports/databases/percona57-pam-for-mysql/boost_1_59_0/libs/wave/samples/list_includes/lexertl/ |
H A D | lexertl_lexer.hpp | 610 string_type token_val; in get() local 612 wave::token_id id = lexer_.next_token(first, last, token_val); in get() 623 validate_identifier_name(token_val, in get() 634 token_val = convert_trigraphs(token_val); in get() 638 validate_literal(token_val, in get() 648 invalid_long_long_literal, token_val.c_str(), in get() 663 if (0 == token_val.compare(start, 12, "include_next", 12)) in get() 673 token_val.clear(); in get() 688 token_val = convert_trigraph(token_val); in get() 702 sprintf(buffer, "%02x'", token_val[0]); in get() [all …]
|
/dports/databases/mysqlwsrep57-server/boost_1_59_0/libs/wave/samples/list_includes/lexertl/ |
H A D | lexertl_lexer.hpp | 610 string_type token_val; in get() local 612 wave::token_id id = lexer_.next_token(first, last, token_val); in get() 623 validate_identifier_name(token_val, in get() 634 token_val = convert_trigraphs(token_val); in get() 638 validate_literal(token_val, in get() 648 invalid_long_long_literal, token_val.c_str(), in get() 663 if (0 == token_val.compare(start, 12, "include_next", 12)) in get() 673 token_val.clear(); in get() 688 token_val = convert_trigraph(token_val); in get() 702 sprintf(buffer, "%02x'", token_val[0]); in get() [all …]
|
/dports/databases/percona57-server/boost_1_59_0/libs/wave/samples/list_includes/lexertl/ |
H A D | lexertl_lexer.hpp | 610 string_type token_val; in get() local 612 wave::token_id id = lexer_.next_token(first, last, token_val); in get() 623 validate_identifier_name(token_val, in get() 634 token_val = convert_trigraphs(token_val); in get() 638 validate_literal(token_val, in get() 648 invalid_long_long_literal, token_val.c_str(), in get() 663 if (0 == token_val.compare(start, 12, "include_next", 12)) in get() 673 token_val.clear(); in get() 688 token_val = convert_trigraph(token_val); in get() 702 sprintf(buffer, "%02x'", token_val[0]); in get() [all …]
|
/dports/databases/xtrabackup/boost_1_59_0/libs/wave/samples/list_includes/lexertl/ |
H A D | lexertl_lexer.hpp | 610 string_type token_val; in get() local 612 wave::token_id id = lexer_.next_token(first, last, token_val); in get() 623 validate_identifier_name(token_val, in get() 634 token_val = convert_trigraphs(token_val); in get() 638 validate_literal(token_val, in get() 648 invalid_long_long_literal, token_val.c_str(), in get() 663 if (0 == token_val.compare(start, 12, "include_next", 12)) in get() 673 token_val.clear(); in get() 688 token_val = convert_trigraph(token_val); in get() 702 sprintf(buffer, "%02x'", token_val[0]); in get() [all …]
|
/dports/databases/percona57-client/boost_1_59_0/libs/wave/samples/list_includes/lexertl/ |
H A D | lexertl_lexer.hpp | 610 string_type token_val; in get() local 612 wave::token_id id = lexer_.next_token(first, last, token_val); in get() 623 validate_identifier_name(token_val, in get() 634 token_val = convert_trigraphs(token_val); in get() 638 validate_literal(token_val, in get() 648 invalid_long_long_literal, token_val.c_str(), in get() 663 if (0 == token_val.compare(start, 12, "include_next", 12)) in get() 673 token_val.clear(); in get() 688 token_val = convert_trigraph(token_val); in get() 702 sprintf(buffer, "%02x'", token_val[0]); in get() [all …]
|
/dports/devel/boost-libs/boost_1_72_0/libs/wave/samples/list_includes/lexertl/ |
H A D | lexertl_lexer.hpp | 610 string_type token_val; in get() local 612 wave::token_id id = lexer_.next_token(first, last, token_val); in get() 623 validate_identifier_name(token_val, in get() 634 token_val = convert_trigraphs(token_val); in get() 638 validate_literal(token_val, in get() 648 invalid_long_long_literal, token_val.c_str(), in get() 663 if (0 == token_val.compare(start, 12, "include_next", 12)) in get() 673 token_val.clear(); in get() 688 token_val = convert_trigraph(token_val); in get() 702 sprintf(buffer, "%02x'", token_val[0]); in get() [all …]
|
/dports/editors/poke/poke-1.0/poke/ |
H A D | pk-hserver.c | 93 pk_val token_val = pk_make_int (token, 32); in pk_hserver_token_p() local 98 ret = pk_call (poke_compiler, cls, &token_p, token_val, PK_NULL); in pk_hserver_token_p() 108 pk_val token_val = pk_make_int (token, 32); in pk_hserver_token_kind() local 113 ret = pk_call (poke_compiler, cls, &token_kind, token_val, PK_NULL); in pk_hserver_token_kind() 123 pk_val token_val = pk_make_int (token, 32); in pk_hserver_cmd() local 128 ret = pk_call (poke_compiler, cls, &cmd, token_val, PK_NULL); in pk_hserver_cmd()
|
/dports/math/stanmath/math-4.2.0/lib/boost_1.75.0/libs/wave/samples/list_includes/lexertl/ |
H A D | lexertl_lexer.hpp | 691 string_type token_val; in get() local 693 wave::token_id id = lexer_.next_token(first, last, token_val); in get() 704 validate_identifier_name(token_val, in get() 715 token_val = convert_trigraphs(token_val); in get() 719 validate_literal(token_val, in get() 729 invalid_long_long_literal, token_val.c_str(), in get() 744 if (0 == token_val.compare(start, 12, "include_next", 12)) in get() 754 token_val.clear(); in get() 769 token_val = convert_trigraph(token_val); in get() 783 sprintf(buffer, "%02x'", token_val[0]); in get() [all …]
|
/dports/devel/hyperscan/boost_1_75_0/libs/wave/samples/list_includes/lexertl/ |
H A D | lexertl_lexer.hpp | 691 string_type token_val; in get() local 693 wave::token_id id = lexer_.next_token(first, last, token_val); in get() 704 validate_identifier_name(token_val, in get() 715 token_val = convert_trigraphs(token_val); in get() 719 validate_literal(token_val, in get() 729 invalid_long_long_literal, token_val.c_str(), in get() 744 if (0 == token_val.compare(start, 12, "include_next", 12)) in get() 754 token_val.clear(); in get() 769 token_val = convert_trigraph(token_val); in get() 783 sprintf(buffer, "%02x'", token_val[0]); in get() [all …]
|
/dports/devel/boost-docs/boost_1_72_0/boost/wave/grammars/ |
H A D | cpp_intlit_grammar.hpp | 165 typename TokenT::string_type const &token_val = token.get_value(); in evaluate() local 167 parse(token_val.begin(), token_val.end(), g[spirit_assign_actor(result)]); in evaluate() 171 token_val.c_str(), token.get_position()); in evaluate()
|
/dports/games/frogatto/frogatto-1.3.1/MacOSJet/boost/include/boost/wave/grammars/ |
H A D | cpp_intlit_grammar.hpp | 165 typename TokenT::string_type const &token_val = token.get_value(); in evaluate() local 167 parse(token_val.begin(), token_val.end(), g[spirit_assign_actor(result)]); in evaluate() 171 token_val.c_str(), token.get_position()); in evaluate()
|
/dports/databases/percona57-pam-for-mysql/boost_1_59_0/boost/wave/grammars/ |
H A D | cpp_intlit_grammar.hpp | 165 typename TokenT::string_type const &token_val = token.get_value(); in evaluate() local 167 parse(token_val.begin(), token_val.end(), g[spirit_assign_actor(result)]); in evaluate() 171 token_val.c_str(), token.get_position()); in evaluate()
|
/dports/math/stanmath/math-4.2.0/lib/boost_1.75.0/boost/wave/grammars/ |
H A D | cpp_intlit_grammar.hpp | 163 typename TokenT::string_type const &token_val = token.get_value(); in evaluate() local 166 parse(token_val.begin(), token_val.end(), g[spirit_assign_actor(result)]); in evaluate() 170 token_val.c_str(), token.get_position()); in evaluate()
|