/* * 02/15/2005 * * SQLTokenMaker.java - Scanner for SQL. * * This library is distributed under a modified BSD license. See the included * RSyntaxTextArea.License.txt file for details. */ package org.fife.ui.rsyntaxtextarea.modes; import java.io.*; import javax.swing.text.Segment; import org.fife.ui.rsyntaxtextarea.*; /** * This class generates tokens representing a text stream as SQL.

* * This implementation was created using * JFlex 1.4.1; however, the generated file * was modified for performance. Memory allocation needs to be almost * completely removed to be competitive with the handwritten lexers (subclasses * of AbstractTokenMaker, so this class has been modified so that * Strings are never allocated (via yytext()), and the scanner never has to * worry about refilling its buffer (needlessly copying chars around). * We can achieve this because RText always scans exactly 1 line of tokens at a * time, and hands the scanner this line as an array of characters (a Segment * really). Since tokens contain pointers to char arrays instead of Strings * holding their contents, there is no need for allocating new memory for * Strings.

* * The actual algorithm generated for scanning has, of course, not been * modified.

* * If you wish to regenerate this file yourself, keep in mind the following: *

* * @author Robert Futrell * @version 0.5 * */ %% %public %class SQLTokenMaker %extends AbstractJFlexTokenMaker %unicode %ignorecase %type org.fife.ui.rsyntaxtextarea.Token %{ /** * Constructor. This must be here because JFlex does not generate a * no-parameter constructor. */ public SQLTokenMaker() { super(); } /** * Adds the token specified to the current linked list of tokens. * * @param tokenType The token's type. */ private void addToken(int tokenType) { addToken(zzStartRead, zzMarkedPos-1, tokenType); } /** * Adds the token specified to the current linked list of tokens. * * @param tokenType The token's type. */ private void addToken(int start, int end, int tokenType) { int so = start + offsetShift; addToken(zzBuffer, start,end, tokenType, so); } /** * Adds the token specified to the current linked list of tokens. * * @param array The character array. * @param start The starting offset in the array. * @param end The ending offset in the array. * @param tokenType The token's type. * @param startOffset The offset in the document at which this token * occurs. */ public void addToken(char[] array, int start, int end, int tokenType, int startOffset) { super.addToken(array, start,end, tokenType, startOffset); zzStartRead = zzMarkedPos; } /** * Returns the text to place at the beginning and end of a * line to "comment" it in a this programming language. * * @return The start and end strings to add to a line to "comment" * it out. */ public String[] getLineCommentStartAndEnd() { return new String[] { "--", null }; } /** * Returns the first token in the linked list of tokens generated * from text. This method must be implemented by * subclasses so they can correctly implement syntax highlighting. * * @param text The text from which to get tokens. * @param initialTokenType The token type we should start with. * @param startOffset The offset into the document at which * text starts. * @return The first Token in a linked list representing * the syntax highlighted text. */ public Token getTokenList(Segment text, int initialTokenType, int startOffset) { resetTokenList(); this.offsetShift = -text.offset + startOffset; // Start off in the proper state. int state = Token.NULL; switch (initialTokenType) { case Token.LITERAL_STRING_DOUBLE_QUOTE: state = STRING; start = text.offset; break; case Token.LITERAL_CHAR: state = CHAR; start = text.offset; break; case Token.COMMENT_MULTILINE: state = MLC; start = text.offset; break; default: state = Token.NULL; } s = text; try { yyreset(zzReader); yybegin(state); return yylex(); } catch (IOException ioe) { ioe.printStackTrace(); return new DefaultToken(); } } /** * Refills the input buffer. * * @return true if EOF was reached, otherwise * false. */ private boolean zzRefill() { return zzCurrentPos>=s.offset+s.count; } /** * Resets the scanner to read from a new input stream. * Does not close the old reader. * * All internal variables are reset, the old input stream * cannot be reused (internal buffer is discarded and lost). * Lexical state is set to YY_INITIAL. * * @param reader the new input stream */ public final void yyreset(java.io.Reader reader) { // 's' has been updated. zzBuffer = s.array; /* * We replaced the line below with the two below it because zzRefill * no longer "refills" the buffer (since the way we do it, it's always * "full" the first time through, since it points to the segment's * array). So, we assign zzEndRead here. */ //zzStartRead = zzEndRead = s.offset; zzStartRead = s.offset; zzEndRead = zzStartRead + s.count - 1; zzCurrentPos = zzMarkedPos = zzPushbackPos = s.offset; zzLexicalState = YYINITIAL; zzReader = reader; zzAtBOL = true; zzAtEOF = false; } %} LineTerminator = ([\n]) Letter = ([A-Za-z]) Digit = ([0-9]) Whitespace = ([ \t]+) IdentifierStart = ({Letter}) IdentifierPart = ({IdentifierStart}|{Digit}|[_]) Identifier = ({IdentifierStart}{IdentifierPart}*) Operator = (">="|"<="|"<>"|">"|"<"|"="|"+"|"-"|"*"|"/") Separator = ([\(\)]) Parameter = ([:]{Identifier}) Integer = ({Digit}+) Float = (({Digit}+[.]{Digit}*)|([.]{Digit}*)) ApproxNum = (({Digit}+[eE][+-]?{Digit}+)|({Digit}+[.]{Digit}*[eE][+-]?[0-9]+)|([.][0-9]*[eE][+-]?[0-9]+)) CommentBegin = ("--") Comment = ({CommentBegin}.*) MLCBegin = "/*" MLCEnd = "*/" %state STRING %state CHAR %state MLC %% { /* Keywords */ "ADD" | "ALL" | "ALTER" | "AND" | "ANY" | "AS" | "ASC" | "AUTOINCREMENT" | "AVA" | "BETWEEN" | "BINARY" | "BIT" | "BOOLEAN" | "BY" | "BYTE" | "CHAR" | "CHARACTER" | "COLUMN" | "CONSTRAINT" | "COUNT" | "COUNTER" | "CREATE" | "CURRENCY" | "DATABASE" | "DATE" | "DATETIME" | "DELETE" | "DESC" | "DISALLOW" | "DISTINCT" | "DISTINCTROW" | "DOUBLE" | "DROP" | "EXISTS" | "FLOAT" | "FLOAT4" | "FLOAT8" | "FOREIGN" | "FROM" | "GENERAL" | "GROUP" | "GUID" | "HAVING" | "INNER" | "INSERT" | "IGNORE" | "IMP" | "IN" | "INDEX" | "INT" | "INTEGER" | "INTEGER1" | "INTEGER2" | "INTEGER4" | "INTO" | "IS" | "JOIN" | "KEY" | "LEFT" | "LEVEL" | "LIKE" | "LOGICAL" | "LONG" | "LONGBINARY" | "LONGTEXT" | "MAX" | "MEMO" | "MIN" | "MOD" | "MONEY" | "NOT" | "NULL" | "NUMBER" | "NUMERIC" | "OLEOBJECT" | "ON" | "OPTION" | "OR" | "ORDER" | "OUTER" | "OWNERACCESS" | "PARAMETERS" | "PASSWORD" | "PERCENT" | "PIVOT" | "PRIMARY" | "REAL" | "REFERENCES" | "RIGHT" | "SELECT" | "SET" | "SHORT" | "SINGLE" | "SMALLINT" | "SOME" | "STDEV" | "STDEVP" | "STRING" | "SUM" | "TABLE" | "TABLEID" | "TEXT" | "TIME" | "TIMESTAMP" | "TOP" | "TRANSFORM" | "TYPE" | "UNION" | "UNIQUE" | "UPDATE" | "USER" | "VALUE" | "VALUES" | "VAR" | "VARBINARY" | "VARCHAR" | "VARP" | "WHERE" | "WITH" | "YESNO" { addToken(Token.RESERVED_WORD); } /* SQL99 aggregate functions */ "AVG" | "COUNT" | "MIN" | "MAX" | "SUM" { addToken(Token.FUNCTION); } /* SQL99 built-in scalar functions */ "CURRENT_DATE" | "CURRENT_TIME" | "CURRENT_TIMESTAMP" | "CURRENT_USER" | "SESSION_USER" | "SYSTEM_USER" { addToken(Token.FUNCTION); } /* SQL99 numeric scalar functions */ "BIT_LENGTH" | "CHAR_LENGTH" | "EXTRACT" | "OCTET_LENGTH" | "POSITION" { addToken(Token.FUNCTION); } /* SQL99 string functions */ "CONCATENATE" | "CONVERT" | "LOWER" | "SUBSTRING" | "TRANSLATE" | "TRIM" | "UPPER" { addToken(Token.FUNCTION); } {LineTerminator} { addNullToken(); return firstToken; } {Identifier} { addToken(Token.IDENTIFIER); } ";" { addToken(Token.IDENTIFIER); } {Parameter} { addToken(Token.IDENTIFIER); } {Comment} { addToken(Token.COMMENT_EOL); } {MLCBegin} { start = zzMarkedPos-2; yybegin(MLC); } {Whitespace} { addToken(Token.WHITESPACE); } {Operator} { addToken(Token.OPERATOR); } {Separator} { addToken(Token.SEPARATOR); } {Integer} { addToken(Token.LITERAL_NUMBER_DECIMAL_INT); } {Float} { addToken(Token.LITERAL_NUMBER_FLOAT); } {ApproxNum} { addToken(Token.LITERAL_NUMBER_FLOAT); } "\"" { start = zzMarkedPos-1; yybegin(STRING); } "\'" { start = zzMarkedPos-1; yybegin(CHAR); } "["[^\]]*"]" { addToken(Token.PREPROCESSOR); } "["[^\]]* { addToken(Token.ERROR_IDENTIFIER); addNullToken(); return firstToken; } <> { addNullToken(); return firstToken; } /* Catch any other (unhandled) characters and flag them as OK; */ /* I don't know enough about SQL to know what's really invalid. */ . { addToken(Token.IDENTIFIER); } } { [^\n\"]+ {} \n { addToken(start,zzStartRead-1, Token.LITERAL_STRING_DOUBLE_QUOTE); return firstToken; } "\"\"" {} "\"" { yybegin(YYINITIAL); addToken(start,zzStartRead, Token.LITERAL_STRING_DOUBLE_QUOTE); } <> { addToken(start,zzStartRead-1, Token.LITERAL_STRING_DOUBLE_QUOTE); return firstToken; } } { [^\n\']+ {} \n { addToken(start,zzStartRead-1, Token.LITERAL_CHAR); return firstToken; } "\'\'" {} "\'" { yybegin(YYINITIAL); addToken(start,zzStartRead, Token.LITERAL_CHAR); } <> { addToken(start,zzStartRead-1, Token.LITERAL_CHAR); return firstToken; } } { [^\n\*]+ {} \n { addToken(start,zzStartRead-1, Token.COMMENT_MULTILINE); return firstToken; } {MLCEnd} { yybegin(YYINITIAL); addToken(start,zzStartRead+1, Token.COMMENT_MULTILINE); } \* {} <> { addToken(start,zzStartRead-1, Token.COMMENT_MULTILINE); return firstToken; } }