Lines Matching +refs:lex +refs:attr

20 namespace boost { namespace spirit { namespace lex  namespace
69 tokenize_and_parse(Iterator& first, Iterator last, Lexer const& lex, in tokenize_and_parse() argument
77 typename Lexer::iterator_type iter = lex.begin(first, last); in tokenize_and_parse()
79 iter, lex.end(), unused, unused, unused); in tokenize_and_parse()
86 tokenize_and_parse(Iterator& first, Iterator last, Lexer const& lex in tokenize_and_parse() argument
87 , ParserExpr const& xpr, Attribute& attr) in tokenize_and_parse() argument
94 typename Lexer::iterator_type iter = lex.begin(first, last); in tokenize_and_parse()
96 iter, lex.end(), unused, unused, attr); in tokenize_and_parse()
153 , Lexer const& lex, ParserExpr const& xpr, Skipper const& skipper in tokenize_and_phrase_parse() argument
167 typename Lexer::iterator_type iter = lex.begin(first, last); in tokenize_and_phrase_parse()
168 typename Lexer::iterator_type end = lex.end(); in tokenize_and_phrase_parse()
183 , Lexer const& lex, ParserExpr const& xpr, Skipper const& skipper in tokenize_and_phrase_parse() argument
184 , BOOST_SCOPED_ENUM(skip_flag) post_skip, Attribute& attr) in tokenize_and_phrase_parse() argument
197 typename Lexer::iterator_type iter = lex.begin(first, last); in tokenize_and_phrase_parse()
198 typename Lexer::iterator_type end = lex.end(); in tokenize_and_phrase_parse()
200 iter, end, unused, skipper_, attr)) in tokenize_and_phrase_parse()
214 , Lexer const& lex, ParserExpr const& xpr, Skipper const& skipper in tokenize_and_phrase_parse() argument
215 , Attribute& attr) in tokenize_and_phrase_parse() argument
217 return tokenize_and_phrase_parse(first, last, lex, xpr, skipper in tokenize_and_phrase_parse()
218 , skip_flag::postskip, attr); in tokenize_and_phrase_parse()
291 tokenize(Iterator& first, Iterator last, Lexer const& lex, F f in tokenize() argument
296 iterator_type iter = lex.begin(first, last, initial_state); in tokenize()
297 iterator_type end = lex.end(); in tokenize()
309 tokenize(Iterator& first, Iterator last, Lexer const& lex in tokenize() argument
314 iterator_type iter = lex.begin(first, last, initial_state); in tokenize()
315 iterator_type end = lex.end(); in tokenize()