Home
last modified time | relevance | path

Searched refs:Tokenizer (Results 1 – 25 of 3225) sorted by relevance

12345678910>>...129

/dports/devel/protobuf25/protobuf-2.5.0/src/google/protobuf/io/
H A Dtokenizer_unittest.cc198 Tokenizer::TokenType type;
220 { "123.45", Tokenizer::TYPE_FLOAT },
221 { "1.", Tokenizer::TYPE_FLOAT },
222 { "1e3", Tokenizer::TYPE_FLOAT },
223 { "1E3", Tokenizer::TYPE_FLOAT },
224 { "1e-3", Tokenizer::TYPE_FLOAT },
225 { "1e+3", Tokenizer::TYPE_FLOAT },
226 { "1.e3", Tokenizer::TYPE_FLOAT },
445 Tokenizer::Token token; in TEST_2D()
824 Tokenizer::ParseString("'", &output); in TEST_F()
[all …]
/dports/databases/percona57-pam-for-mysql/percona-server-5.7.36-39/extra/protobuf/protobuf-2.6.1/src/google/protobuf/io/
H A Dtokenizer_unittest.cc198 Tokenizer::TokenType type;
220 { "123.45", Tokenizer::TYPE_FLOAT },
221 { "1.", Tokenizer::TYPE_FLOAT },
222 { "1e3", Tokenizer::TYPE_FLOAT },
223 { "1E3", Tokenizer::TYPE_FLOAT },
224 { "1e-3", Tokenizer::TYPE_FLOAT },
225 { "1e+3", Tokenizer::TYPE_FLOAT },
226 { "1.e3", Tokenizer::TYPE_FLOAT },
439 Tokenizer::Token token; in TEST_2D()
818 Tokenizer::ParseString("'", &output); in TEST_F()
[all …]
/dports/databases/mysqlwsrep57-server/mysql-wsrep-wsrep_5.7.35-25.27/extra/protobuf/protobuf-2.6.1/src/google/protobuf/io/
H A Dtokenizer_unittest.cc198 Tokenizer::TokenType type;
220 { "123.45", Tokenizer::TYPE_FLOAT },
221 { "1.", Tokenizer::TYPE_FLOAT },
222 { "1e3", Tokenizer::TYPE_FLOAT },
223 { "1E3", Tokenizer::TYPE_FLOAT },
224 { "1e-3", Tokenizer::TYPE_FLOAT },
225 { "1e+3", Tokenizer::TYPE_FLOAT },
226 { "1.e3", Tokenizer::TYPE_FLOAT },
439 Tokenizer::Token token; in TEST_2D()
818 Tokenizer::ParseString("'", &output); in TEST_F()
[all …]
/dports/databases/percona57-server/percona-server-5.7.36-39/extra/protobuf/protobuf-2.6.1/src/google/protobuf/io/
H A Dtokenizer_unittest.cc198 Tokenizer::TokenType type;
220 { "123.45", Tokenizer::TYPE_FLOAT },
221 { "1.", Tokenizer::TYPE_FLOAT },
222 { "1e3", Tokenizer::TYPE_FLOAT },
223 { "1E3", Tokenizer::TYPE_FLOAT },
224 { "1e-3", Tokenizer::TYPE_FLOAT },
225 { "1e+3", Tokenizer::TYPE_FLOAT },
226 { "1.e3", Tokenizer::TYPE_FLOAT },
439 Tokenizer::Token token; in TEST_2D()
818 Tokenizer::ParseString("'", &output); in TEST_F()
[all …]
/dports/databases/xtrabackup/percona-xtrabackup-2.4.21/extra/protobuf/protobuf-2.6.1/src/google/protobuf/io/
H A Dtokenizer_unittest.cc198 Tokenizer::TokenType type;
220 { "123.45", Tokenizer::TYPE_FLOAT },
221 { "1.", Tokenizer::TYPE_FLOAT },
222 { "1e3", Tokenizer::TYPE_FLOAT },
223 { "1E3", Tokenizer::TYPE_FLOAT },
224 { "1e-3", Tokenizer::TYPE_FLOAT },
225 { "1e+3", Tokenizer::TYPE_FLOAT },
226 { "1.e3", Tokenizer::TYPE_FLOAT },
439 Tokenizer::Token token; in TEST_2D()
818 Tokenizer::ParseString("'", &output); in TEST_F()
[all …]
/dports/databases/percona57-client/percona-server-5.7.36-39/extra/protobuf/protobuf-2.6.1/src/google/protobuf/io/
H A Dtokenizer_unittest.cc198 Tokenizer::TokenType type;
220 { "123.45", Tokenizer::TYPE_FLOAT },
221 { "1.", Tokenizer::TYPE_FLOAT },
222 { "1e3", Tokenizer::TYPE_FLOAT },
223 { "1E3", Tokenizer::TYPE_FLOAT },
224 { "1e-3", Tokenizer::TYPE_FLOAT },
225 { "1e+3", Tokenizer::TYPE_FLOAT },
226 { "1.e3", Tokenizer::TYPE_FLOAT },
439 Tokenizer::Token token; in TEST_2D()
818 Tokenizer::ParseString("'", &output); in TEST_F()
[all …]
/dports/databases/mysql57-client/mysql-5.7.36/extra/protobuf/protobuf-2.6.1/src/google/protobuf/io/
H A Dtokenizer_unittest.cc198 Tokenizer::TokenType type;
220 { "123.45", Tokenizer::TYPE_FLOAT },
221 { "1.", Tokenizer::TYPE_FLOAT },
222 { "1e3", Tokenizer::TYPE_FLOAT },
223 { "1E3", Tokenizer::TYPE_FLOAT },
224 { "1e-3", Tokenizer::TYPE_FLOAT },
225 { "1e+3", Tokenizer::TYPE_FLOAT },
226 { "1.e3", Tokenizer::TYPE_FLOAT },
439 Tokenizer::Token token; in TEST_2D()
818 Tokenizer::ParseString("'", &output); in TEST_F()
[all …]
/dports/devel/protobuf/protobuf-3.17.3/src/google/protobuf/io/
H A Dtokenizer_unittest.cc198 Tokenizer::TokenType type;
221 {"1.", Tokenizer::TYPE_FLOAT},
222 {"1e3", Tokenizer::TYPE_FLOAT},
223 {"1E3", Tokenizer::TYPE_FLOAT},
224 {"1e-3", Tokenizer::TYPE_FLOAT},
225 {"1e+3", Tokenizer::TYPE_FLOAT},
226 {"1.e3", Tokenizer::TYPE_FLOAT},
228 {".1", Tokenizer::TYPE_FLOAT},
244 {"+", Tokenizer::TYPE_SYMBOL},
245 {".", Tokenizer::TYPE_SYMBOL},
[all …]
/dports/devel/py-grpcio-tools/grpcio-tools-1.43.0/third_party/protobuf/src/google/protobuf/io/
H A Dtokenizer_unittest.cc198 Tokenizer::TokenType type;
221 {"1.", Tokenizer::TYPE_FLOAT},
222 {"1e3", Tokenizer::TYPE_FLOAT},
223 {"1E3", Tokenizer::TYPE_FLOAT},
224 {"1e-3", Tokenizer::TYPE_FLOAT},
225 {"1e+3", Tokenizer::TYPE_FLOAT},
226 {"1.e3", Tokenizer::TYPE_FLOAT},
228 {".1", Tokenizer::TYPE_FLOAT},
244 {"+", Tokenizer::TYPE_SYMBOL},
245 {".", Tokenizer::TYPE_SYMBOL},
[all …]
/dports/www/qt5-webengine/qtwebengine-everywhere-src-5.15.2/src/3rdparty/chromium/third_party/protobuf/src/google/protobuf/io/
H A Dtokenizer_unittest.cc198 Tokenizer::TokenType type;
221 {"1.", Tokenizer::TYPE_FLOAT},
222 {"1e3", Tokenizer::TYPE_FLOAT},
223 {"1E3", Tokenizer::TYPE_FLOAT},
224 {"1e-3", Tokenizer::TYPE_FLOAT},
225 {"1e+3", Tokenizer::TYPE_FLOAT},
226 {"1.e3", Tokenizer::TYPE_FLOAT},
228 {".1", Tokenizer::TYPE_FLOAT},
244 {"+", Tokenizer::TYPE_SYMBOL},
245 {".", Tokenizer::TYPE_SYMBOL},
[all …]
/dports/www/chromium-legacy/chromium-88.0.4324.182/third_party/protobuf/src/google/protobuf/io/
H A Dtokenizer_unittest.cc198 Tokenizer::TokenType type;
221 {"1.", Tokenizer::TYPE_FLOAT},
222 {"1e3", Tokenizer::TYPE_FLOAT},
223 {"1E3", Tokenizer::TYPE_FLOAT},
224 {"1e-3", Tokenizer::TYPE_FLOAT},
225 {"1e+3", Tokenizer::TYPE_FLOAT},
226 {"1.e3", Tokenizer::TYPE_FLOAT},
228 {".1", Tokenizer::TYPE_FLOAT},
244 {"+", Tokenizer::TYPE_SYMBOL},
245 {".", Tokenizer::TYPE_SYMBOL},
[all …]
/dports/dns/dnsjava/dnsjava-2.1.9/tests/org/xbill/DNS/
H A DTokenizerTest.java46 private Tokenizer m_t;
201 m_t = new Tokenizer(")"); in test_get_invalid()
205 m_t = new Tokenizer("\\"); in test_get_invalid()
256 m_t = new Tokenizer(" "); in test_unwanted_ungotten_whitespace()
274 m_t = new Tokenizer(""); in test_empty_string()
278 m_t = new Tokenizer(" "); in test_empty_string()
466 m_t = new Tokenizer(""); in test_getName()
476 m_t = new Tokenizer("id"); in test_getEOL()
485 m_t = new Tokenizer("\n"); in test_getEOL()
494 m_t = new Tokenizer("id"); in test_getEOL()
[all …]
/dports/www/firefox-legacy/firefox-52.8.0esr/xpcom/tests/gtest/
H A DTestTokenizer.cpp30 Tokenizer::Token t; in TEST()
97 TEST(Tokenizer, Main) in TEST() argument
99 Tokenizer::Token t; in TEST()
109 Tokenizer::Token u; in TEST()
235 Tokenizer::Token t; in TEST()
248 Tokenizer::Token t; in TEST()
315 Tokenizer p1(a); in TEST()
322 Tokenizer p1(a); in TEST()
330 Tokenizer p1(a); in TEST()
338 Tokenizer p1(a); in TEST()
[all …]
/dports/lang/spidermonkey60/firefox-60.9.0/xpcom/tests/gtest/
H A DTestTokenizer.cpp30 Tokenizer::Token t; in TEST()
97 TEST(Tokenizer, Main) in TEST() argument
99 Tokenizer::Token t; in TEST()
109 Tokenizer::Token u; in TEST()
235 Tokenizer::Token t; in TEST()
248 Tokenizer::Token t; in TEST()
315 Tokenizer p1(a); in TEST()
322 Tokenizer p1(a); in TEST()
330 Tokenizer p1(a); in TEST()
338 Tokenizer p1(a); in TEST()
[all …]
/dports/devel/android-tools-adb/platform_system_core-android-9.0.0_r3/init/parser/
H A Dtokenizer_test.cpp26 Tokenizer tokenizer(data); \
38 TEST(Tokenizer, Empty) { in TEST() argument
43 TEST(Tokenizer, Simple) { in TEST() argument
61 TEST(Tokenizer, WhiteSpace) { in TEST() argument
68 TEST(Tokenizer, TwoTokens) { in TEST() argument
76 TEST(Tokenizer, MultiToken) { in TEST() argument
87 TEST(Tokenizer, NewLine) { in TEST() argument
94 TEST(Tokenizer, TextNewLine) { in TEST() argument
125 TEST(Tokenizer, Comment) { in TEST() argument
215 TEST(Tokenizer, SimpleQuotes) { in TEST() argument
[all …]
/dports/devel/android-tools-fastboot/platform_system_core-platform-tools-29.0.5/init/parser/
H A Dtokenizer_test.cpp26 Tokenizer tokenizer(data); \
38 TEST(Tokenizer, Empty) { in TEST() argument
43 TEST(Tokenizer, Simple) { in TEST() argument
61 TEST(Tokenizer, WhiteSpace) { in TEST() argument
68 TEST(Tokenizer, TwoTokens) { in TEST() argument
76 TEST(Tokenizer, MultiToken) { in TEST() argument
87 TEST(Tokenizer, NewLine) { in TEST() argument
94 TEST(Tokenizer, TextNewLine) { in TEST() argument
125 TEST(Tokenizer, Comment) { in TEST() argument
215 TEST(Tokenizer, SimpleQuotes) { in TEST() argument
[all …]
/dports/mail/thunderbird/thunderbird-91.8.0/xpcom/tests/gtest/
H A DTestTokenizer.cpp26 Tokenizer::Token t; in TEST()
30 Tokenizer p( in TEST()
98 TEST(Tokenizer, Main) in TEST() argument
100 Tokenizer::Token t; in TEST()
110 Tokenizer::Token u; in TEST()
429 Tokenizer p1(a); in TEST()
436 Tokenizer p1(a); in TEST()
444 Tokenizer p1(a); in TEST()
562 Tokenizer p(str); in TEST()
848 Tokenizer p( in TEST()
[all …]
/dports/www/firefox/firefox-99.0/xpcom/tests/gtest/
H A DTestTokenizer.cpp26 Tokenizer::Token t; in TEST()
30 Tokenizer p( in TEST()
103 TEST(Tokenizer, Main) in TEST() argument
105 Tokenizer::Token t; in TEST()
115 Tokenizer::Token u; in TEST()
436 Tokenizer p1(a); in TEST()
443 Tokenizer p1(a); in TEST()
451 Tokenizer p1(a); in TEST()
569 Tokenizer p(str); in TEST()
855 Tokenizer p( in TEST()
[all …]
/dports/www/firefox-esr/firefox-91.8.0/xpcom/tests/gtest/
H A DTestTokenizer.cpp26 Tokenizer::Token t; in TEST()
30 Tokenizer p( in TEST()
98 TEST(Tokenizer, Main) in TEST() argument
100 Tokenizer::Token t; in TEST()
110 Tokenizer::Token u; in TEST()
429 Tokenizer p1(a); in TEST()
436 Tokenizer p1(a); in TEST()
444 Tokenizer p1(a); in TEST()
562 Tokenizer p(str); in TEST()
848 Tokenizer p( in TEST()
[all …]
/dports/lang/spidermonkey78/firefox-78.9.0/xpcom/tests/gtest/
H A DTestTokenizer.cpp26 Tokenizer::Token t; in TEST()
30 Tokenizer p( in TEST()
98 TEST(Tokenizer, Main) in TEST() argument
100 Tokenizer::Token t; in TEST()
110 Tokenizer::Token u; in TEST()
430 Tokenizer p1(a); in TEST()
437 Tokenizer p1(a); in TEST()
445 Tokenizer p1(a); in TEST()
565 Tokenizer p(str); in TEST()
852 Tokenizer p( in TEST()
[all …]
/dports/converters/wkhtmltopdf/qt-5db36ec/tools/porting/src/
H A Dtokenizer.cpp54 bool Tokenizer::s_initialized = false;
55 Tokenizer::scan_fun_ptr Tokenizer::s_scan_table[128 + 1];
56 int Tokenizer::s_attr_table[256];
58 Tokenizer::Tokenizer() in Tokenizer() function in Tokenizer
65 Tokenizer::~Tokenizer() in ~Tokenizer()
77 void Tokenizer::setupScanTable() in setupScanTable()
160 bool Tokenizer::nextToken(Token &tok) in nextToken()
174 void Tokenizer::scanChar(int *kind) in scanChar()
190 void Tokenizer::scanNewline(int *kind) in scanNewline()
303 void Tokenizer::scanComment(int *kind) in scanComment()
[all …]
/dports/textproc/py-mwparserfromhell/mwparserfromhell-0.6.3/src/mwparserfromhell/parser/ctokenizer/
H A Dtok_support.h29 int Tokenizer_push(Tokenizer *, uint64_t);
30 int Tokenizer_push_textbuffer(Tokenizer *);
31 void Tokenizer_delete_top_of_stack(Tokenizer *);
32 PyObject *Tokenizer_pop(Tokenizer *);
34 void Tokenizer_memoize_bad_route(Tokenizer *);
35 void *Tokenizer_fail_route(Tokenizer *);
36 int Tokenizer_check_route(Tokenizer *, uint64_t);
37 void Tokenizer_free_bad_route_tree(Tokenizer *);
41 int Tokenizer_emit_char(Tokenizer *, Py_UCS4);
44 int Tokenizer_emit_all(Tokenizer *, PyObject *);
[all …]
/dports/www/mediawiki137/mediawiki-1.37.1/vendor/wikimedia/remex-html/bin/
H A Dtest.php11 use Wikimedia\RemexHtml\Tokenizer; alias
17 $tokenizer = new Tokenizer\Tokenizer( $handler, $text, [] );
27 $tokenizer = new Tokenizer\Tokenizer( $handler, $text, [] );
36 reserializeState( $text, Tokenizer\Tokenizer::STATE_SCRIPT_DATA, 'script' );
40 reserializeState( $text, Tokenizer\Tokenizer::STATE_RCDATA, 'xmp' );
53 $tokenizer = new Tokenizer\Tokenizer( $dispatchTracer, $text, [] );
68 $tokenizer = new Tokenizer\Tokenizer( $dispatchTracer, $text, [] );
107 $tokenizer = new Tokenizer\Tokenizer( $dispatcher, $text, [] );
121 $tokenizer = new Tokenizer\Tokenizer( $dispatcher, $text, [] );
129 $tokenizer = new Tokenizer\Tokenizer( $handler, $text, $GLOBALS['tokenizerOptions'] );
[all …]
/dports/textproc/luceneplusplus/LucenePlusPlus-rel_3.0.8/src/core/analysis/
H A DTokenizer.cpp13 Tokenizer::Tokenizer() { in Tokenizer() function in Lucene::Tokenizer
16 Tokenizer::Tokenizer(const ReaderPtr& input) { in Tokenizer() function in Lucene::Tokenizer
21 Tokenizer::Tokenizer(const AttributeFactoryPtr& factory) : TokenStream(factory) { in Tokenizer() function in Lucene::Tokenizer
24 Tokenizer::Tokenizer(const AttributeFactoryPtr& factory, const ReaderPtr& input) : TokenStream(fact… in Tokenizer() function in Lucene::Tokenizer
29 Tokenizer::Tokenizer(const AttributeSourcePtr& source) : TokenStream(source) { in Tokenizer() function in Lucene::Tokenizer
32 Tokenizer::Tokenizer(const AttributeSourcePtr& source, const ReaderPtr& input) : TokenStream(source… in Tokenizer() function in Lucene::Tokenizer
37 Tokenizer::~Tokenizer() { in ~Tokenizer()
40 void Tokenizer::close() { in close()
47 int32_t Tokenizer::correctOffset(int32_t currentOff) { in correctOffset()
51 void Tokenizer::reset(const ReaderPtr& input) { in reset()
/dports/science/lammps/lammps-stable_29Sep2021/unittest/utils/
H A Dtest_tokenizer.cpp22 TEST(Tokenizer, empty_string) in TEST() argument
24 Tokenizer t("", " "); in TEST()
30 Tokenizer t(" ", " "); in TEST()
34 TEST(Tokenizer, single_word) in TEST() argument
36 Tokenizer t("test", " "); in TEST()
40 TEST(Tokenizer, two_words) in TEST() argument
46 TEST(Tokenizer, skip) in TEST() argument
89 Tokenizer u(t); in TEST()
112 Tokenizer v = u; in TEST()
146 TEST(Tokenizer, unix_paths) in TEST() argument
[all …]

12345678910>>...129