/dports/textproc/py-jsx-lexer/jsx-lexer-0.0.7/tests/ |
H A D | tokens.py | 1 from pygments.token import Token 6 (Token.Text, ' '), 8 (Token.Text, ' '), 9 (Token.Operator, '='), 10 (Token.Text, ' '), 14 (Token.Text, ' '), 17 (Token.Text, ' '), 19 (Token.Text, ' '), 23 (Token.Text, ' '), 29 (Token.Text, ' '), [all …]
|
/dports/devel/py-moto/moto-2.2.16/tests/test_dynamodb2/ |
H A D | test_dynamodb_expression_tokenizer.py | 12 Token(Token.ATTRIBUTE, "SET"), 13 Token(Token.WHITESPACE, " "), 15 Token(Token.WHITESPACE, " "), 16 Token(Token.EQUAL_SIGN, "="), 17 Token(Token.WHITESPACE, " "), 27 Token(Token.WHITESPACE, " "), 29 Token(Token.WHITESPACE, " "), 30 Token(Token.EQUAL_SIGN, "="), 141 Token(Token.DOT, "."), 157 Token(Token.DOT, "."), [all …]
|
/dports/textproc/py-pygments/Pygments-2.7.2/tests/ |
H A D | test_python.py | 35 (Token.Text, ' '), 40 (Token.Text, '\n'), 43 (Token.Text, '\n'), 46 (Token.Text, ' '), 66 (Token.Name, 'S'), 67 (Token.Text, ' '), 69 (Token.Text, ' '), 71 (Token.Name, 'H'), 72 (Token.Text, ' '), 74 (Token.Text, ' '), [all …]
|
H A D | test_promql.py | 13 from pygments.token import Token 34 (Token.Punctuation, "{"), 36 (Token.Operator, "="), 37 (Token.Punctuation, '"'), 39 (Token.Punctuation, '"'), 52 (Token.Operator, "="), 58 (Token.Operator, "="), 75 (Token.Operator, "="), 82 (Token.Operator, "="), 99 (Token.Operator, "="), [all …]
|
H A D | test_bibtex.py | 28 (Token.String, '"'), 30 (Token.String, '"'), 32 (Token.Text, '\n'), 43 (Token.Text, ' '), 45 (Token.Text, ' '), 50 (Token.Text, '\n'), 76 (Token.Text, ' '), 78 (Token.Text, ' '), 82 (Token.Text, ' '), 84 (Token.Text, ' '), [all …]
|
H A D | test_ezhil.py | 45 (Token.Name, 'gcd'), 47 (Token.Name, 'a'), 49 (Token.Name, 'b'), 51 (Token.Text, '\n') 63 (Token.Text, ' '), 65 (Token.Text, ' '), 67 (Token.Text, ' '), 69 (Token.Text, ' '), 71 (Token.Text, ' '), 76 (Token.Text, ' '), [all …]
|
H A D | test_clexer.py | 49 (Token.Text, ' '), 53 (Token.Text, '\n'), 55 (Token.Text, '\n'), 58 (Token.Text, ' '), 62 (Token.Text, '\n'), 65 (Token.Text, '\n'), 68 (Token.Text, ' '), 71 (Token.Text, '\n'), 75 (Token.Text, '\n'), 102 (Token.Text, ' '), [all …]
|
H A D | test_matlab.py | 12 from pygments.token import Token 27 (Token.Name, 'set'), 40 (Token.Text, '\n'), 77 (Token.Text, ' '), 78 (Token.Name, 'x'), 79 (Token.Text, ' '), 81 (Token.Text, ' '), 84 (Token.Text, ' '), 85 (Token.Text, ' '), 86 (Token.Text, ' '), [all …]
|
H A D | test_gdscript.py | 25 (Token.Text, " "), 26 (Token.Name, "abc"), 27 (Token.Text, " "), 29 (Token.Text, " "), 31 (Token.Text, "\n"), 40 (Token.Text, " "), 46 (Token.Text, "\n"), 47 (Token.Text, "\t"), 63 (Token.Text, " "), 68 (Token.Text, " "), [all …]
|
H A D | test_praat.py | 12 from pygments.token import Token 24 (Token.Text, 'var'), 25 (Token.Text, ' '), 27 (Token.Text, ' '), 30 (Token.Text, '\n'), 39 (Token.Text, ' '), 41 (Token.Text, ' '), 45 (Token.Text, '\n'), 75 (Token.Text, ' '), 76 (Token.Text, 'i'), [all …]
|
/dports/devel/jsonnet/jsonnet-0.17.0/core/ |
H A D | lexer_test.cpp | 28 test_tokens.push_back(Token(Token::Kind::END_OF_FILE, "")); in testLex() 56 {Token(Token::Kind::OPERATOR, "<"), Token(Token::Kind::OPERATOR, "-")}, in TEST() 58 testLex("comma", ",", {Token(Token::Kind::COMMA, "")}, ""); in TEST() 60 testLex("dot", ".", {Token(Token::Kind::DOT, "")}, ""); in TEST() 100 {Token(Token::Kind::NUMBER, "0"), Token(Token::Kind::NUMBER, "100")}, in TEST() 104 {Token(Token::Kind::NUMBER, "10"), in TEST() 105 Token(Token::Kind::OPERATOR, "+"), in TEST() 106 Token(Token::Kind::NUMBER, "10")}, in TEST() 125 {Token(Token::Kind::STRING_DOUBLE, "hi\\\"")}, in TEST() 129 {Token(Token::Kind::STRING_DOUBLE, "hi\\\n")}, in TEST() [all …]
|
/dports/devel/py-jsonnet/jsonnet-0.17.0/core/ |
H A D | lexer_test.cpp | 28 test_tokens.push_back(Token(Token::Kind::END_OF_FILE, "")); in testLex() 56 {Token(Token::Kind::OPERATOR, "<"), Token(Token::Kind::OPERATOR, "-")}, in TEST() 58 testLex("comma", ",", {Token(Token::Kind::COMMA, "")}, ""); in TEST() 60 testLex("dot", ".", {Token(Token::Kind::DOT, "")}, ""); in TEST() 100 {Token(Token::Kind::NUMBER, "0"), Token(Token::Kind::NUMBER, "100")}, in TEST() 104 {Token(Token::Kind::NUMBER, "10"), in TEST() 105 Token(Token::Kind::OPERATOR, "+"), in TEST() 106 Token(Token::Kind::NUMBER, "10")}, in TEST() 125 {Token(Token::Kind::STRING_DOUBLE, "hi\\\"")}, in TEST() 129 {Token(Token::Kind::STRING_DOUBLE, "hi\\\n")}, in TEST() [all …]
|
/dports/devel/p5-Compiler-Lexer/Compiler-Lexer-0.23/src/compiler/util/ |
H A D | Compiler_gen_token_decl.cpp | 4 {Enum::Token::Type::Return, Enum::Token::Kind::Return, "Return", "return"}, 5 {Enum::Token::Type::Add, Enum::Token::Kind::Operator, "Add", "+"}, 6 {Enum::Token::Type::Sub, Enum::Token::Kind::Operator, "Sub", "-"}, 7 {Enum::Token::Type::Mul, Enum::Token::Kind::Operator, "Mul", "*"}, 8 {Enum::Token::Type::Div, Enum::Token::Kind::Operator, "Div", "/"}, 9 {Enum::Token::Type::Mod, Enum::Token::Kind::Operator, "Mod", "%"}, 12 {Enum::Token::Type::Less, Enum::Token::Kind::Operator, "Less", "<"}, 14 {Enum::Token::Type::Ref, Enum::Token::Kind::Operator, "Ref", "\\"}, 15 {Enum::Token::Type::Glob, Enum::Token::Kind::Operator, "Glob", "*"}, 17 {Enum::Token::Type::BitOr, Enum::Token::Kind::Operator, "BitOr", "|"}, [all …]
|
H A D | Compiler_reserved_keyword.cpp | 104 {"=", {Enum::Token::Type::Assign, Enum::Token::Kind::Assign, "Assign", "="}}, in in_word_set() 127 {"next", {Enum::Token::Type::Next, Enum::Token::Kind::Control, "Next", "next"}}, in in_word_set() 155 {"&&", {Enum::Token::Type::And, Enum::Token::Kind::Operator, "And", "&&"}}, in in_word_set() 282 {":", {Enum::Token::Type::Colon, Enum::Token::Kind::Colon, "Colon", ":"}}, in in_word_set() 475 {"!", {Enum::Token::Type::Not, Enum::Token::Kind::SingleTerm, "Not", "!"}}, in in_word_set() 522 {"do", {Enum::Token::Type::Do, Enum::Token::Kind::Do, "Do", "do"}}, in in_word_set() 611 {"%", {Enum::Token::Type::Mod, Enum::Token::Kind::Operator, "Mod", "%"}}, in in_word_set() 740 {"+", {Enum::Token::Type::Add, Enum::Token::Kind::Operator, "Add", "+"}}, in in_word_set() 859 {"/", {Enum::Token::Type::Div, Enum::Token::Kind::Operator, "Div", "/"}}, in in_word_set() 1077 {"-", {Enum::Token::Type::Sub, Enum::Token::Kind::Operator, "Sub", "-"}}, in in_word_set() [all …]
|
/dports/devel/py-prompt-toolkit1/prompt_toolkit-1.0.18/examples/ |
H A D | ansi-colors.py | 53 (Token.Black, '#ansiblack'), (Token, '\n'), 54 (Token.White, '#ansiwhite'), (Token, '\n'), 55 (Token.Red, '#ansired'), (Token, '\n'), 56 (Token.Green, '#ansigreen'), (Token, '\n'), 57 (Token.Yellow, '#ansiyellow'), (Token, '\n'), 58 (Token.Blue, '#ansiblue'), (Token, '\n'), 66 (Token.Brown, '#ansibrown'), (Token, '\n'), 69 (Token.Teal, '#ansiteal'), (Token, '\n'), 74 (Token.BgRed, '#ansired'), (Token, '\n'), 77 (Token.BgBlue, '#ansiblue'), (Token, '\n'), [all …]
|
/dports/www/chromium-legacy/chromium-88.0.4324.182/third_party/tint/src/src/reader/wgsl/ |
H A D | token.cc | 28 case Token::Type::kEOF: in TypeToName() 357 Token::Token() : type_(Type::kUninitialized) {} in Token() function in tint::reader::wgsl::Token 359 Token::Token(Type type, const Source& source, const std::string& val) in Token() function in tint::reader::wgsl::Token 362 Token::Token(const Source& source, uint32_t val) in Token() function in tint::reader::wgsl::Token 365 Token::Token(const Source& source, int32_t val) in Token() function in tint::reader::wgsl::Token 368 Token::Token(const Source& source, float val) in Token() function in tint::reader::wgsl::Token 371 Token::Token(Type type, const Source& source) : Token(type, source, "") {} in Token() function in tint::reader::wgsl::Token 373 Token::Token(Token&&) = default; 375 Token::Token(const Token&) = default; 377 Token::~Token() = default; [all …]
|
/dports/lang/mono/mono-5.10.1.57/mcs/ilasm/scanner/ |
H A D | ILTables.cs | 86 directives [".os"] = new ILToken (Token.D_OS, ".os"); 123 keywords ["at"] = new ILToken (Token.K_AT, "at"); 124 keywords ["as"] = new ILToken (Token.K_AS, "as"); 141 keywords ["in"] = new ILToken (Token.K_IN, "in"); 189 keywords ["il"] = new ILToken (Token.K_IL, "il"); 253 keywords ["cf"] = new ILToken (Token.K_CF, "cf"); 271 keywords ["to"] = new ILToken (Token.K_TO, "to"); 316 keywords ["is"] = new ILToken (Token.K_IS, "is"); 317 keywords ["on"] = new ILToken (Token.K_ON, "on"); 319 keywords ["strict"] = new ILToken (Token.K_STRICT, "strict"); [all …]
|
/dports/devel/p5-Compiler-Lexer/Compiler-Lexer-0.23/gen/ |
H A D | reserved_keywords.gperf | 13 "return", {Enum::Token::Type::Return, Enum::Token::Kind::Return, "Return", "return"} 14 "+", {Enum::Token::Type::Add, Enum::Token::Kind::Operator, "Add", "+"} 15 "-", {Enum::Token::Type::Sub, Enum::Token::Kind::Operator, "Sub", "-"} 16 "/", {Enum::Token::Type::Div, Enum::Token::Kind::Operator, "Div", "/"} 17 "%", {Enum::Token::Type::Mod, Enum::Token::Kind::Operator, "Mod", "%"} 19 ">", {Enum::Token::Type::Greater, Enum::Token::Kind::Operator, "Greater", ">"} 20 "<", {Enum::Token::Type::Less, Enum::Token::Kind::Operator, "Less", "<"} 22 "\\", {Enum::Token::Type::Ref, Enum::Token::Kind::Operator, "Ref", "\\"} 23 "~", {Enum::Token::Type::BitNot, Enum::Token::Kind::Operator, "BitNot", "~"} 24 "|", {Enum::Token::Type::BitOr, Enum::Token::Kind::Operator, "BitOr", "|"} [all …]
|
/dports/textproc/py-pygments-25/Pygments-2.5.2/tests/ |
H A D | test_bibtex.py | 15 from pygments.token import Token 28 (Token.String, u'"'), 30 (Token.String, u'"'), 32 (Token.Text, u'\n'), 43 (Token.Text, u' '), 45 (Token.Text, u' '), 50 (Token.Text, u'\n'), 76 (Token.Text, u' '), 78 (Token.Text, u' '), 82 (Token.Text, u' '), [all …]
|
H A D | test_ezhil.py | 45 (Token.Name, u'gcd'), 47 (Token.Name, u'a'), 49 (Token.Name, u'b'), 51 (Token.Text, u'\n') 63 (Token.Text, u' '), 65 (Token.Text, u' '), 67 (Token.Text, u' '), 69 (Token.Text, u' '), 71 (Token.Text, u' '), 76 (Token.Text, u' '), [all …]
|
H A D | test_clexer.py | 49 (Token.Text, u' '), 53 (Token.Text, u'\n'), 55 (Token.Text, u'\n'), 58 (Token.Text, u' '), 62 (Token.Text, u'\n'), 65 (Token.Text, u'\n'), 68 (Token.Text, u' '), 71 (Token.Text, u'\n'), 75 (Token.Text, u'\n'), 102 (Token.Text, u' '), [all …]
|
/dports/devel/cppcheck/cppcheck-2.6.3/lib/ |
H A D | astutils.h | 53 const Token* findAstNode(const Token* ast, const std::function<bool(const Token*)>& pred); 103 Token* previousBeforeAstLeftmostLeaf(Token* tok); 106 Token* nextAfterAstRightmostLeaf(Token* tok); 108 Token* astParentSkipParens(Token* tok); 111 const Token* getParentMember(const Token * tok); 113 const Token* getParentLifetime(const Token* tok); 118 Token* getCondTok(Token* tok); 119 const Token* getCondTok(const Token* tok); 121 Token* getCondTokFromEnd(Token* endBlock); 283 Token* findLambdaEndToken(Token* first); [all …]
|
/dports/devel/cppcheck-gui/cppcheck-2.6.3/lib/ |
H A D | astutils.h | 53 const Token* findAstNode(const Token* ast, const std::function<bool(const Token*)>& pred); 103 Token* previousBeforeAstLeftmostLeaf(Token* tok); 106 Token* nextAfterAstRightmostLeaf(Token* tok); 108 Token* astParentSkipParens(Token* tok); 111 const Token* getParentMember(const Token * tok); 113 const Token* getParentLifetime(const Token* tok); 118 Token* getCondTok(Token* tok); 119 const Token* getCondTok(const Token* tok); 121 Token* getCondTokFromEnd(Token* endBlock); 283 Token* findLambdaEndToken(Token* first); [all …]
|
/dports/textproc/p5-PPIx-Utils/PPIx-Utils-0.003/t/ |
H A D | split_ppi_node_by_namespace.t | 34 [ 1, 8, 8 ] PPI::Token::Whitespace ' ' 35 [ 1, 9, 9 ] PPI::Token::Word 'Foo' 36 [ 1, 12, 12 ] PPI::Token::Structure ';' 37 [ 1, 13, 13 ] PPI::Token::Whitespace '\n' 39 [ 2, 1, 1 ] PPI::Token::Symbol '$x' 40 [ 2, 3, 3 ] PPI::Token::Whitespace ' ' 41 [ 2, 4, 4 ] PPI::Token::Operator '=' 43 [ 2, 6, 6 ] PPI::Token::Number '1' 45 [ 2, 8, 8 ] PPI::Token::Whitespace '\n' 62 [ 3, 21, 21 ] PPI::Token::Whitespace '\n' [all …]
|
/dports/shells/xonsh/xonsh-0.7.9/xonsh/ |
H A D | style_tools.py | 58 Token = _TokenType() variable 59 Color = Token.Color 265 Token: "", 357 Token.Escape: "", 358 Token.Generic: "", 376 Token.Literal: "", 402 Token.Name: "", 415 Token.Name.Other: "", 425 Token.Other: "", 426 Token.Punctuation: "", [all …]
|