Class: Sparkql::Lexer
- Inherits:
-
StringScanner
- Object
- StringScanner
- Sparkql::Lexer
- Includes:
- Token
- Defined in:
- lib/sparkql/lexer.rb
Constant Summary
Constants included from Token
Token::ADD, Token::BOOLEAN, Token::CHARACTER, Token::CONJUNCTIONS, Token::CUSTOM_FIELD, Token::DATE, Token::DATETIME, Token::DECIMAL, Token::DIV, Token::EQUALITY_OPERATORS, Token::INTEGER, Token::KEYWORD, Token::LPAREN, Token::MOD, Token::MUL, Token::NEWLINE, Token::NULL, Token::OPERATORS, Token::RANGE_OPERATOR, Token::RPAREN, Token::SPACE, Token::STANDARD_FIELD, Token::SUB, Token::TIME, Token::UNARY_CONJUNCTIONS
Instance Attribute Summary collapse
-
#block_group_identifier ⇒ Object
Returns the value of attribute block_group_identifier.
-
#current_token_value ⇒ Object
readonly
Returns the value of attribute current_token_value.
-
#last_field ⇒ Object
readonly
Returns the value of attribute last_field.
-
#level ⇒ Object
Returns the value of attribute level.
-
#token_index ⇒ Object
readonly
Returns the value of attribute token_index.
Instance Method Summary collapse
- #check_keywords(value) ⇒ Object
- #check_reserved_words(value) ⇒ Object
- #check_standard_fields(value) ⇒ Object
-
#initialize(str) ⇒ Lexer
constructor
A new instance of Lexer.
- #leveldown ⇒ Object
- #levelup ⇒ Object
- #literal(symbol, value) ⇒ Object
-
#shift ⇒ Object
Lookup the next matching token.
Constructor Details
#initialize(str) ⇒ Lexer
Returns a new instance of Lexer.
10 11 12 13 14 15 16 |
# File 'lib/sparkql/lexer.rb', line 10 def initialize(str) str.freeze super(str, false) # DO NOT dup str @level = 0 @block_group_identifier = 0 @expression_count = 0 end |
Instance Attribute Details
#block_group_identifier ⇒ Object
Returns the value of attribute block_group_identifier.
6 7 8 |
# File 'lib/sparkql/lexer.rb', line 6 def block_group_identifier @block_group_identifier end |
#current_token_value ⇒ Object (readonly)
Returns the value of attribute current_token_value.
8 9 10 |
# File 'lib/sparkql/lexer.rb', line 8 def current_token_value @current_token_value end |
#last_field ⇒ Object (readonly)
Returns the value of attribute last_field.
8 9 10 |
# File 'lib/sparkql/lexer.rb', line 8 def last_field @last_field end |
#level ⇒ Object
Returns the value of attribute level.
6 7 8 |
# File 'lib/sparkql/lexer.rb', line 6 def level @level end |
#token_index ⇒ Object (readonly)
Returns the value of attribute token_index.
8 9 10 |
# File 'lib/sparkql/lexer.rb', line 8 def token_index @token_index end |
Instance Method Details
#check_keywords(value) ⇒ Object
99 100 101 102 103 104 105 |
# File 'lib/sparkql/lexer.rb', line 99 def check_keywords(value) result = check_reserved_words(value) if result.first == :UNKNOWN result = [:KEYWORD, value] end result end |
#check_reserved_words(value) ⇒ Object
65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 |
# File 'lib/sparkql/lexer.rb', line 65 def check_reserved_words(value) u_value = value.capitalize if OPERATORS.include?(u_value) [:OPERATOR, u_value] elsif RANGE_OPERATOR == u_value [:RANGE_OPERATOR, u_value] elsif CONJUNCTIONS.include?(u_value) [:CONJUNCTION, u_value] elsif UNARY_CONJUNCTIONS.include?(u_value) [:UNARY_CONJUNCTION, u_value] elsif ADD == u_value [:ADD, u_value] elsif SUB == u_value [:SUB, u_value] elsif MUL == u_value [:MUL, u_value] elsif DIV == u_value [:DIV, u_value] elsif MOD == u_value [:MOD, u_value] else [:UNKNOWN, "ERROR: '#{self.string}'"] end end |
#check_standard_fields(value) ⇒ Object
90 91 92 93 94 95 96 97 |
# File 'lib/sparkql/lexer.rb', line 90 def check_standard_fields(value) result = check_reserved_words(value) if result.first == :UNKNOWN @last_field = value result = [:STANDARD_FIELD, value] end result end |
#leveldown ⇒ Object
112 113 114 |
# File 'lib/sparkql/lexer.rb', line 112 def leveldown @level -= 1 end |
#levelup ⇒ Object
107 108 109 110 |
# File 'lib/sparkql/lexer.rb', line 107 def levelup @level += 1 @block_group_identifier += 1 end |
#literal(symbol, value) ⇒ Object
116 117 118 119 120 121 122 |
# File 'lib/sparkql/lexer.rb', line 116 def literal(symbol, value) node = { type: symbol.to_s.downcase.to_sym, value: value } [symbol, node] end |
#shift ⇒ Object
Lookup the next matching token
19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 |
# File 'lib/sparkql/lexer.rb', line 19 def shift @token_index = self.pos token = if (@current_token_value = scan(SPACE)) [:SPACE, @current_token_value] elsif (@current_token_value = scan(LPAREN)) levelup [:LPAREN, @current_token_value] elsif (@current_token_value = scan(RPAREN)) # leveldown: do this after parsing group [:RPAREN, @current_token_value] elsif (@current_token_value = scan(/,/)) [:COMMA, @current_token_value] elsif (@current_token_value = scan(NULL)) literal :NULL, "NULL" elsif (@current_token_value = scan(STANDARD_FIELD)) check_standard_fields(@current_token_value) elsif (@current_token_value = scan(DATETIME)) literal :DATETIME, @current_token_value elsif (@current_token_value = scan(DATE)) literal :DATE, @current_token_value elsif (@current_token_value = scan(TIME)) literal :TIME, @current_token_value elsif (@current_token_value = scan(DECIMAL)) literal :DECIMAL, @current_token_value elsif (@current_token_value = scan(INTEGER)) literal :INTEGER, @current_token_value elsif (@current_token_value = scan(/-/)) [:UMINUS, @current_token_value] elsif (@current_token_value = scan(CHARACTER)) literal :CHARACTER, @current_token_value elsif (@current_token_value = scan(BOOLEAN)) literal :BOOLEAN, @current_token_value elsif (@current_token_value = scan(KEYWORD)) check_keywords(@current_token_value) elsif (@current_token_value = scan(CUSTOM_FIELD)) [:CUSTOM_FIELD, @current_token_value] elsif eos? [false, false] # end of file, \Z don't work with StringScanner else [:UNKNOWN, "ERROR: '#{self.string}'"] end token.freeze end |