/dports/devel/protobuf25/protobuf-2.5.0/src/google/protobuf/io/ |
H A D | tokenizer_unittest.cc | 198 Tokenizer::TokenType type; 220 { "123.45", Tokenizer::TYPE_FLOAT }, 221 { "1.", Tokenizer::TYPE_FLOAT }, 222 { "1e3", Tokenizer::TYPE_FLOAT }, 223 { "1E3", Tokenizer::TYPE_FLOAT }, 224 { "1e-3", Tokenizer::TYPE_FLOAT }, 225 { "1e+3", Tokenizer::TYPE_FLOAT }, 226 { "1.e3", Tokenizer::TYPE_FLOAT }, 445 Tokenizer::Token token; in TEST_2D() 824 Tokenizer::ParseString("'", &output); in TEST_F() [all …]
|
/dports/databases/percona57-pam-for-mysql/percona-server-5.7.36-39/extra/protobuf/protobuf-2.6.1/src/google/protobuf/io/ |
H A D | tokenizer_unittest.cc | 198 Tokenizer::TokenType type; 220 { "123.45", Tokenizer::TYPE_FLOAT }, 221 { "1.", Tokenizer::TYPE_FLOAT }, 222 { "1e3", Tokenizer::TYPE_FLOAT }, 223 { "1E3", Tokenizer::TYPE_FLOAT }, 224 { "1e-3", Tokenizer::TYPE_FLOAT }, 225 { "1e+3", Tokenizer::TYPE_FLOAT }, 226 { "1.e3", Tokenizer::TYPE_FLOAT }, 439 Tokenizer::Token token; in TEST_2D() 818 Tokenizer::ParseString("'", &output); in TEST_F() [all …]
|
/dports/databases/mysqlwsrep57-server/mysql-wsrep-wsrep_5.7.35-25.27/extra/protobuf/protobuf-2.6.1/src/google/protobuf/io/ |
H A D | tokenizer_unittest.cc | 198 Tokenizer::TokenType type; 220 { "123.45", Tokenizer::TYPE_FLOAT }, 221 { "1.", Tokenizer::TYPE_FLOAT }, 222 { "1e3", Tokenizer::TYPE_FLOAT }, 223 { "1E3", Tokenizer::TYPE_FLOAT }, 224 { "1e-3", Tokenizer::TYPE_FLOAT }, 225 { "1e+3", Tokenizer::TYPE_FLOAT }, 226 { "1.e3", Tokenizer::TYPE_FLOAT }, 439 Tokenizer::Token token; in TEST_2D() 818 Tokenizer::ParseString("'", &output); in TEST_F() [all …]
|
/dports/databases/percona57-server/percona-server-5.7.36-39/extra/protobuf/protobuf-2.6.1/src/google/protobuf/io/ |
H A D | tokenizer_unittest.cc | 198 Tokenizer::TokenType type; 220 { "123.45", Tokenizer::TYPE_FLOAT }, 221 { "1.", Tokenizer::TYPE_FLOAT }, 222 { "1e3", Tokenizer::TYPE_FLOAT }, 223 { "1E3", Tokenizer::TYPE_FLOAT }, 224 { "1e-3", Tokenizer::TYPE_FLOAT }, 225 { "1e+3", Tokenizer::TYPE_FLOAT }, 226 { "1.e3", Tokenizer::TYPE_FLOAT }, 439 Tokenizer::Token token; in TEST_2D() 818 Tokenizer::ParseString("'", &output); in TEST_F() [all …]
|
/dports/databases/xtrabackup/percona-xtrabackup-2.4.21/extra/protobuf/protobuf-2.6.1/src/google/protobuf/io/ |
H A D | tokenizer_unittest.cc | 198 Tokenizer::TokenType type; 220 { "123.45", Tokenizer::TYPE_FLOAT }, 221 { "1.", Tokenizer::TYPE_FLOAT }, 222 { "1e3", Tokenizer::TYPE_FLOAT }, 223 { "1E3", Tokenizer::TYPE_FLOAT }, 224 { "1e-3", Tokenizer::TYPE_FLOAT }, 225 { "1e+3", Tokenizer::TYPE_FLOAT }, 226 { "1.e3", Tokenizer::TYPE_FLOAT }, 439 Tokenizer::Token token; in TEST_2D() 818 Tokenizer::ParseString("'", &output); in TEST_F() [all …]
|
/dports/databases/percona57-client/percona-server-5.7.36-39/extra/protobuf/protobuf-2.6.1/src/google/protobuf/io/ |
H A D | tokenizer_unittest.cc | 198 Tokenizer::TokenType type; 220 { "123.45", Tokenizer::TYPE_FLOAT }, 221 { "1.", Tokenizer::TYPE_FLOAT }, 222 { "1e3", Tokenizer::TYPE_FLOAT }, 223 { "1E3", Tokenizer::TYPE_FLOAT }, 224 { "1e-3", Tokenizer::TYPE_FLOAT }, 225 { "1e+3", Tokenizer::TYPE_FLOAT }, 226 { "1.e3", Tokenizer::TYPE_FLOAT }, 439 Tokenizer::Token token; in TEST_2D() 818 Tokenizer::ParseString("'", &output); in TEST_F() [all …]
|
/dports/databases/mysql57-client/mysql-5.7.36/extra/protobuf/protobuf-2.6.1/src/google/protobuf/io/ |
H A D | tokenizer_unittest.cc | 198 Tokenizer::TokenType type; 220 { "123.45", Tokenizer::TYPE_FLOAT }, 221 { "1.", Tokenizer::TYPE_FLOAT }, 222 { "1e3", Tokenizer::TYPE_FLOAT }, 223 { "1E3", Tokenizer::TYPE_FLOAT }, 224 { "1e-3", Tokenizer::TYPE_FLOAT }, 225 { "1e+3", Tokenizer::TYPE_FLOAT }, 226 { "1.e3", Tokenizer::TYPE_FLOAT }, 439 Tokenizer::Token token; in TEST_2D() 818 Tokenizer::ParseString("'", &output); in TEST_F() [all …]
|
/dports/devel/protobuf/protobuf-3.17.3/src/google/protobuf/io/ |
H A D | tokenizer_unittest.cc | 198 Tokenizer::TokenType type; 221 {"1.", Tokenizer::TYPE_FLOAT}, 222 {"1e3", Tokenizer::TYPE_FLOAT}, 223 {"1E3", Tokenizer::TYPE_FLOAT}, 224 {"1e-3", Tokenizer::TYPE_FLOAT}, 225 {"1e+3", Tokenizer::TYPE_FLOAT}, 226 {"1.e3", Tokenizer::TYPE_FLOAT}, 228 {".1", Tokenizer::TYPE_FLOAT}, 244 {"+", Tokenizer::TYPE_SYMBOL}, 245 {".", Tokenizer::TYPE_SYMBOL}, [all …]
|
/dports/devel/py-grpcio-tools/grpcio-tools-1.43.0/third_party/protobuf/src/google/protobuf/io/ |
H A D | tokenizer_unittest.cc | 198 Tokenizer::TokenType type; 221 {"1.", Tokenizer::TYPE_FLOAT}, 222 {"1e3", Tokenizer::TYPE_FLOAT}, 223 {"1E3", Tokenizer::TYPE_FLOAT}, 224 {"1e-3", Tokenizer::TYPE_FLOAT}, 225 {"1e+3", Tokenizer::TYPE_FLOAT}, 226 {"1.e3", Tokenizer::TYPE_FLOAT}, 228 {".1", Tokenizer::TYPE_FLOAT}, 244 {"+", Tokenizer::TYPE_SYMBOL}, 245 {".", Tokenizer::TYPE_SYMBOL}, [all …]
|
/dports/www/qt5-webengine/qtwebengine-everywhere-src-5.15.2/src/3rdparty/chromium/third_party/protobuf/src/google/protobuf/io/ |
H A D | tokenizer_unittest.cc | 198 Tokenizer::TokenType type; 221 {"1.", Tokenizer::TYPE_FLOAT}, 222 {"1e3", Tokenizer::TYPE_FLOAT}, 223 {"1E3", Tokenizer::TYPE_FLOAT}, 224 {"1e-3", Tokenizer::TYPE_FLOAT}, 225 {"1e+3", Tokenizer::TYPE_FLOAT}, 226 {"1.e3", Tokenizer::TYPE_FLOAT}, 228 {".1", Tokenizer::TYPE_FLOAT}, 244 {"+", Tokenizer::TYPE_SYMBOL}, 245 {".", Tokenizer::TYPE_SYMBOL}, [all …]
|
/dports/www/chromium-legacy/chromium-88.0.4324.182/third_party/protobuf/src/google/protobuf/io/ |
H A D | tokenizer_unittest.cc | 198 Tokenizer::TokenType type; 221 {"1.", Tokenizer::TYPE_FLOAT}, 222 {"1e3", Tokenizer::TYPE_FLOAT}, 223 {"1E3", Tokenizer::TYPE_FLOAT}, 224 {"1e-3", Tokenizer::TYPE_FLOAT}, 225 {"1e+3", Tokenizer::TYPE_FLOAT}, 226 {"1.e3", Tokenizer::TYPE_FLOAT}, 228 {".1", Tokenizer::TYPE_FLOAT}, 244 {"+", Tokenizer::TYPE_SYMBOL}, 245 {".", Tokenizer::TYPE_SYMBOL}, [all …]
|
/dports/dns/dnsjava/dnsjava-2.1.9/tests/org/xbill/DNS/ |
H A D | TokenizerTest.java | 46 private Tokenizer m_t; 201 m_t = new Tokenizer(")"); in test_get_invalid() 205 m_t = new Tokenizer("\\"); in test_get_invalid() 256 m_t = new Tokenizer(" "); in test_unwanted_ungotten_whitespace() 274 m_t = new Tokenizer(""); in test_empty_string() 278 m_t = new Tokenizer(" "); in test_empty_string() 466 m_t = new Tokenizer(""); in test_getName() 476 m_t = new Tokenizer("id"); in test_getEOL() 485 m_t = new Tokenizer("\n"); in test_getEOL() 494 m_t = new Tokenizer("id"); in test_getEOL() [all …]
|
/dports/www/firefox-legacy/firefox-52.8.0esr/xpcom/tests/gtest/ |
H A D | TestTokenizer.cpp | 30 Tokenizer::Token t; in TEST() 97 TEST(Tokenizer, Main) in TEST() argument 99 Tokenizer::Token t; in TEST() 109 Tokenizer::Token u; in TEST() 235 Tokenizer::Token t; in TEST() 248 Tokenizer::Token t; in TEST() 315 Tokenizer p1(a); in TEST() 322 Tokenizer p1(a); in TEST() 330 Tokenizer p1(a); in TEST() 338 Tokenizer p1(a); in TEST() [all …]
|
/dports/lang/spidermonkey60/firefox-60.9.0/xpcom/tests/gtest/ |
H A D | TestTokenizer.cpp | 30 Tokenizer::Token t; in TEST() 97 TEST(Tokenizer, Main) in TEST() argument 99 Tokenizer::Token t; in TEST() 109 Tokenizer::Token u; in TEST() 235 Tokenizer::Token t; in TEST() 248 Tokenizer::Token t; in TEST() 315 Tokenizer p1(a); in TEST() 322 Tokenizer p1(a); in TEST() 330 Tokenizer p1(a); in TEST() 338 Tokenizer p1(a); in TEST() [all …]
|
/dports/devel/android-tools-adb/platform_system_core-android-9.0.0_r3/init/parser/ |
H A D | tokenizer_test.cpp | 26 Tokenizer tokenizer(data); \ 38 TEST(Tokenizer, Empty) { in TEST() argument 43 TEST(Tokenizer, Simple) { in TEST() argument 61 TEST(Tokenizer, WhiteSpace) { in TEST() argument 68 TEST(Tokenizer, TwoTokens) { in TEST() argument 76 TEST(Tokenizer, MultiToken) { in TEST() argument 87 TEST(Tokenizer, NewLine) { in TEST() argument 94 TEST(Tokenizer, TextNewLine) { in TEST() argument 125 TEST(Tokenizer, Comment) { in TEST() argument 215 TEST(Tokenizer, SimpleQuotes) { in TEST() argument [all …]
|
/dports/devel/android-tools-fastboot/platform_system_core-platform-tools-29.0.5/init/parser/ |
H A D | tokenizer_test.cpp | 26 Tokenizer tokenizer(data); \ 38 TEST(Tokenizer, Empty) { in TEST() argument 43 TEST(Tokenizer, Simple) { in TEST() argument 61 TEST(Tokenizer, WhiteSpace) { in TEST() argument 68 TEST(Tokenizer, TwoTokens) { in TEST() argument 76 TEST(Tokenizer, MultiToken) { in TEST() argument 87 TEST(Tokenizer, NewLine) { in TEST() argument 94 TEST(Tokenizer, TextNewLine) { in TEST() argument 125 TEST(Tokenizer, Comment) { in TEST() argument 215 TEST(Tokenizer, SimpleQuotes) { in TEST() argument [all …]
|
/dports/mail/thunderbird/thunderbird-91.8.0/xpcom/tests/gtest/ |
H A D | TestTokenizer.cpp | 26 Tokenizer::Token t; in TEST() 30 Tokenizer p( in TEST() 98 TEST(Tokenizer, Main) in TEST() argument 100 Tokenizer::Token t; in TEST() 110 Tokenizer::Token u; in TEST() 429 Tokenizer p1(a); in TEST() 436 Tokenizer p1(a); in TEST() 444 Tokenizer p1(a); in TEST() 562 Tokenizer p(str); in TEST() 848 Tokenizer p( in TEST() [all …]
|
/dports/www/firefox/firefox-99.0/xpcom/tests/gtest/ |
H A D | TestTokenizer.cpp | 26 Tokenizer::Token t; in TEST() 30 Tokenizer p( in TEST() 103 TEST(Tokenizer, Main) in TEST() argument 105 Tokenizer::Token t; in TEST() 115 Tokenizer::Token u; in TEST() 436 Tokenizer p1(a); in TEST() 443 Tokenizer p1(a); in TEST() 451 Tokenizer p1(a); in TEST() 569 Tokenizer p(str); in TEST() 855 Tokenizer p( in TEST() [all …]
|
/dports/www/firefox-esr/firefox-91.8.0/xpcom/tests/gtest/ |
H A D | TestTokenizer.cpp | 26 Tokenizer::Token t; in TEST() 30 Tokenizer p( in TEST() 98 TEST(Tokenizer, Main) in TEST() argument 100 Tokenizer::Token t; in TEST() 110 Tokenizer::Token u; in TEST() 429 Tokenizer p1(a); in TEST() 436 Tokenizer p1(a); in TEST() 444 Tokenizer p1(a); in TEST() 562 Tokenizer p(str); in TEST() 848 Tokenizer p( in TEST() [all …]
|
/dports/lang/spidermonkey78/firefox-78.9.0/xpcom/tests/gtest/ |
H A D | TestTokenizer.cpp | 26 Tokenizer::Token t; in TEST() 30 Tokenizer p( in TEST() 98 TEST(Tokenizer, Main) in TEST() argument 100 Tokenizer::Token t; in TEST() 110 Tokenizer::Token u; in TEST() 430 Tokenizer p1(a); in TEST() 437 Tokenizer p1(a); in TEST() 445 Tokenizer p1(a); in TEST() 565 Tokenizer p(str); in TEST() 852 Tokenizer p( in TEST() [all …]
|
/dports/converters/wkhtmltopdf/qt-5db36ec/tools/porting/src/ |
H A D | tokenizer.cpp | 54 bool Tokenizer::s_initialized = false; 55 Tokenizer::scan_fun_ptr Tokenizer::s_scan_table[128 + 1]; 56 int Tokenizer::s_attr_table[256]; 58 Tokenizer::Tokenizer() in Tokenizer() function in Tokenizer 65 Tokenizer::~Tokenizer() in ~Tokenizer() 77 void Tokenizer::setupScanTable() in setupScanTable() 160 bool Tokenizer::nextToken(Token &tok) in nextToken() 174 void Tokenizer::scanChar(int *kind) in scanChar() 190 void Tokenizer::scanNewline(int *kind) in scanNewline() 303 void Tokenizer::scanComment(int *kind) in scanComment() [all …]
|
/dports/textproc/py-mwparserfromhell/mwparserfromhell-0.6.3/src/mwparserfromhell/parser/ctokenizer/ |
H A D | tok_support.h | 29 int Tokenizer_push(Tokenizer *, uint64_t); 30 int Tokenizer_push_textbuffer(Tokenizer *); 31 void Tokenizer_delete_top_of_stack(Tokenizer *); 32 PyObject *Tokenizer_pop(Tokenizer *); 34 void Tokenizer_memoize_bad_route(Tokenizer *); 35 void *Tokenizer_fail_route(Tokenizer *); 36 int Tokenizer_check_route(Tokenizer *, uint64_t); 37 void Tokenizer_free_bad_route_tree(Tokenizer *); 41 int Tokenizer_emit_char(Tokenizer *, Py_UCS4); 44 int Tokenizer_emit_all(Tokenizer *, PyObject *); [all …]
|
/dports/www/mediawiki137/mediawiki-1.37.1/vendor/wikimedia/remex-html/bin/ |
H A D | test.php | 11 use Wikimedia\RemexHtml\Tokenizer; alias 17 $tokenizer = new Tokenizer\Tokenizer( $handler, $text, [] ); 27 $tokenizer = new Tokenizer\Tokenizer( $handler, $text, [] ); 36 reserializeState( $text, Tokenizer\Tokenizer::STATE_SCRIPT_DATA, 'script' ); 40 reserializeState( $text, Tokenizer\Tokenizer::STATE_RCDATA, 'xmp' ); 53 $tokenizer = new Tokenizer\Tokenizer( $dispatchTracer, $text, [] ); 68 $tokenizer = new Tokenizer\Tokenizer( $dispatchTracer, $text, [] ); 107 $tokenizer = new Tokenizer\Tokenizer( $dispatcher, $text, [] ); 121 $tokenizer = new Tokenizer\Tokenizer( $dispatcher, $text, [] ); 129 $tokenizer = new Tokenizer\Tokenizer( $handler, $text, $GLOBALS['tokenizerOptions'] ); [all …]
|
/dports/textproc/luceneplusplus/LucenePlusPlus-rel_3.0.8/src/core/analysis/ |
H A D | Tokenizer.cpp | 13 Tokenizer::Tokenizer() { in Tokenizer() function in Lucene::Tokenizer 16 Tokenizer::Tokenizer(const ReaderPtr& input) { in Tokenizer() function in Lucene::Tokenizer 21 Tokenizer::Tokenizer(const AttributeFactoryPtr& factory) : TokenStream(factory) { in Tokenizer() function in Lucene::Tokenizer 24 Tokenizer::Tokenizer(const AttributeFactoryPtr& factory, const ReaderPtr& input) : TokenStream(fact… in Tokenizer() function in Lucene::Tokenizer 29 Tokenizer::Tokenizer(const AttributeSourcePtr& source) : TokenStream(source) { in Tokenizer() function in Lucene::Tokenizer 32 Tokenizer::Tokenizer(const AttributeSourcePtr& source, const ReaderPtr& input) : TokenStream(source… in Tokenizer() function in Lucene::Tokenizer 37 Tokenizer::~Tokenizer() { in ~Tokenizer() 40 void Tokenizer::close() { in close() 47 int32_t Tokenizer::correctOffset(int32_t currentOff) { in correctOffset() 51 void Tokenizer::reset(const ReaderPtr& input) { in reset()
|
/dports/science/lammps/lammps-stable_29Sep2021/unittest/utils/ |
H A D | test_tokenizer.cpp | 22 TEST(Tokenizer, empty_string) in TEST() argument 24 Tokenizer t("", " "); in TEST() 30 Tokenizer t(" ", " "); in TEST() 34 TEST(Tokenizer, single_word) in TEST() argument 36 Tokenizer t("test", " "); in TEST() 40 TEST(Tokenizer, two_words) in TEST() argument 46 TEST(Tokenizer, skip) in TEST() argument 89 Tokenizer u(t); in TEST() 112 Tokenizer v = u; in TEST() 146 TEST(Tokenizer, unix_paths) in TEST() argument [all …]
|