/* * Copyright 2017 Google Inc. * * Use of this source code is governed by a BSD-style license that can be * found in the LICENSE file. */ #include "src/sksl/lex/NFAtoDFA.h" #include "src/sksl/lex/RegexParser.h" #include #include #include /** * Processes a .lex file and produces .h and .cpp files which implement a lexical analyzer. The .lex * file is a text file with one token definition per line. Each line is of the form: * = * where is either a regular expression (e.g [0-9]) or a double-quoted literal string. */ static constexpr const char* HEADER = "/*\n" " * Copyright 2017 Google Inc.\n" " *\n" " * Use of this source code is governed by a BSD-style license that can be\n" " * found in the LICENSE file.\n" " */\n" "/*****************************************************************************************\n" " ******************** This file was generated by sksllex. Do not edit. *******************\n" " *****************************************************************************************/\n"; void writeH(const DFA& dfa, const char* lexer, const char* token, const std::vector& tokens, const char* hPath) { std::ofstream out(hPath); SkASSERT(out.good()); out << HEADER; out << "#ifndef SKSL_" << lexer << "\n"; out << "#define SKSL_" << lexer << "\n"; out << "#include \n"; out << "#include \n"; out << "namespace SkSL {\n"; out << "\n"; out << "struct " << token << " {\n"; out << " enum Kind {\n"; for (const std::string& t : tokens) { out << " #undef " << t << "\n"; out << " " << t << ",\n"; } out << " };\n"; out << "\n"; out << " " << token << "()\n"; out << " : fKind(Kind::INVALID)\n"; out << " , fOffset(-1)\n"; out << " , fLength(-1) {}\n"; out << "\n"; out << " " << token << "(Kind kind, int32_t offset, int32_t length)\n"; out << " : fKind(kind)\n"; out << " , fOffset(offset)\n"; out << " , fLength(length) {}\n"; out << "\n"; out << " Kind fKind;\n"; out << " int fOffset;\n"; out << " int fLength;\n"; out << "};\n"; out << "\n"; out << "class " << lexer << " {\n"; out << "public:\n"; out << " void start(const char* text, int32_t length) {\n"; out << " fText = text;\n"; out << " fLength = length;\n"; out << " fOffset = 0;\n"; out << " }\n"; out << "\n"; out << " " << token << " next();\n"; out << "\n"; out << "private:\n"; out << " const char* fText;\n"; out << " int32_t fLength;\n"; out << " int32_t fOffset;\n"; out << "};\n"; out << "\n"; out << "} // namespace\n"; out << "#endif\n"; } void writeCPP(const DFA& dfa, const char* lexer, const char* token, const char* include, const char* cppPath) { std::ofstream out(cppPath); SkASSERT(out.good()); out << HEADER; out << "#include \"" << include << "\"\n"; out << "\n"; out << "namespace SkSL {\n"; out << "\n"; size_t states = 0; for (const auto& row : dfa.fTransitions) { states = std::max(states, row.size()); } // arbitrarily-chosen character which is greater than START_CHAR and should not appear in actual // input out << "static const uint8_t INVALID_CHAR = 18;"; out << "static int8_t mappings[" << dfa.fCharMappings.size() << "] = {\n "; const char* separator = ""; for (int m : dfa.fCharMappings) { out << separator << std::to_string(m); separator = ", "; } out << "\n};\n"; out << "static int16_t transitions[" << dfa.fTransitions.size() << "][" << states << "] = {\n"; for (size_t c = 0; c < dfa.fTransitions.size(); ++c) { out << " {"; for (size_t j = 0; j < states; ++j) { if ((size_t) c < dfa.fTransitions.size() && j < dfa.fTransitions[c].size()) { out << " " << dfa.fTransitions[c][j] << ","; } else { out << " 0,"; } } out << " },\n"; } out << "};\n"; out << "\n"; out << "static int8_t accepts[" << states << "] = {"; for (size_t i = 0; i < states; ++i) { if (i < dfa.fAccepts.size()) { out << " " << dfa.fAccepts[i] << ","; } else { out << " " << INVALID << ","; } } out << " };\n"; out << "\n"; out << token << " " << lexer << "::next() {\n"; out << " // note that we cheat here: normally a lexer needs to worry about the case\n"; out << " // where a token has a prefix which is not itself a valid token - for instance, \n"; out << " // maybe we have a valid token 'while', but 'w', 'wh', etc. are not valid\n"; out << " // tokens. Our grammar doesn't have this property, so we can simplify the logic\n"; out << " // a bit.\n"; out << " int32_t startOffset = fOffset;\n"; out << " if (startOffset == fLength) {\n"; out << " return " << token << "(" << token << "::END_OF_FILE, startOffset, 0);\n"; out << " }\n"; out << " int16_t state = 1;\n"; out << " for (;;) {\n"; out << " if (fOffset >= fLength) {\n"; out << " if (accepts[state] == -1) {\n"; out << " return Token(Token::END_OF_FILE, startOffset, 0);\n"; out << " }\n"; out << " break;\n"; out << " }\n"; out << " uint8_t c = (uint8_t) fText[fOffset];"; out << " if (c <= 8 || c >= " << dfa.fCharMappings.size() << ") {"; out << " c = INVALID_CHAR;"; out << " }"; out << " int16_t newState = transitions[mappings[c]][state];\n"; out << " if (!newState) {\n"; out << " break;\n"; out << " }\n"; out << " state = newState;"; out << " ++fOffset;\n"; out << " }\n"; out << " Token::Kind kind = (" << token << "::Kind) accepts[state];\n"; out << " return " << token << "(kind, startOffset, fOffset - startOffset);\n"; out << "}\n"; out << "\n"; out << "} // namespace\n"; } void process(const char* inPath, const char* lexer, const char* token, const char* hPath, const char* cppPath) { NFA nfa; std::vector tokens; tokens.push_back("END_OF_FILE"); std::string line; std::ifstream in(inPath); while (std::getline(in, line)) { std::istringstream split(line); std::string name, delimiter, pattern; if (split >> name >> delimiter >> pattern) { SkASSERT(split.eof()); SkASSERT(name != ""); SkASSERT(delimiter == "="); SkASSERT(pattern != ""); tokens.push_back(name); if (pattern[0] == '"') { SkASSERT(pattern.size() > 2 && pattern[pattern.size() - 1] == '"'); RegexNode node = RegexNode(RegexNode::kChar_Kind, pattern[1]); for (size_t i = 2; i < pattern.size() - 1; ++i) { node = RegexNode(RegexNode::kConcat_Kind, node, RegexNode(RegexNode::kChar_Kind, pattern[i])); } nfa.addRegex(node); } else { nfa.addRegex(RegexParser().parse(pattern)); } } } NFAtoDFA converter(&nfa); DFA dfa = converter.convert(); writeH(dfa, lexer, token, tokens, hPath); writeCPP(dfa, lexer, token, (std::string("src/sksl/SkSL") + lexer + ".h").c_str(), cppPath); } int main(int argc, const char** argv) { if (argc != 6) { printf("usage: sksllex \n"); exit(1); } process(argv[1], argv[2], argv[3], argv[4], argv[5]); return 0; }