Class: Sparkql::Lexer
- Inherits:
-
StringScanner
- Object
- StringScanner
- Sparkql::Lexer
- Includes:
- Token
- Defined in:
- lib/sparkql/lexer.rb
Constant Summary
Constants included from Token
Token::BOOLEAN, Token::CHARACTER, Token::CONJUNCTIONS, Token::CUSTOM_FIELD, Token::DATE, Token::DATETIME, Token::DECIMAL, Token::EQUALITY_OPERATORS, Token::INTEGER, Token::KEYWORD, Token::LPAREN, Token::NEWLINE, Token::NULL, Token::OPERATORS, Token::RANGE_OPERATOR, Token::RPAREN, Token::SPACE, Token::STANDARD_FIELD, Token::TIME, Token::UNARY_CONJUNCTIONS
Instance Attribute Summary collapse
-
#block_group_identifier ⇒ Object
Returns the value of attribute block_group_identifier.
-
#current_token_value ⇒ Object
readonly
Returns the value of attribute current_token_value.
-
#last_field ⇒ Object
readonly
Returns the value of attribute last_field.
-
#level ⇒ Object
Returns the value of attribute level.
-
#token_index ⇒ Object
readonly
Returns the value of attribute token_index.
Instance Method Summary collapse
- #check_keywords(value) ⇒ Object
- #check_reserved_words(value) ⇒ Object
- #check_standard_fields(value) ⇒ Object
-
#initialize(str) ⇒ Lexer
constructor
A new instance of Lexer.
- #leveldown ⇒ Object
- #levelup ⇒ Object
- #literal(symbol, value) ⇒ Object
-
#shift ⇒ Object
Lookup the next matching token.
Constructor Details
#initialize(str) ⇒ Lexer
Returns a new instance of Lexer.
10 11 12 13 14 15 16 |
# File 'lib/sparkql/lexer.rb', line 10 def initialize(str) str.freeze super(str, false) # DO NOT dup str @level = 0 @block_group_identifier = 0 @expression_count = 0 end |
Instance Attribute Details
#block_group_identifier ⇒ Object
Returns the value of attribute block_group_identifier.
6 7 8 |
# File 'lib/sparkql/lexer.rb', line 6 def block_group_identifier @block_group_identifier end |
#current_token_value ⇒ Object (readonly)
Returns the value of attribute current_token_value.
8 9 10 |
# File 'lib/sparkql/lexer.rb', line 8 def current_token_value @current_token_value end |
#last_field ⇒ Object (readonly)
Returns the value of attribute last_field.
8 9 10 |
# File 'lib/sparkql/lexer.rb', line 8 def last_field @last_field end |
#level ⇒ Object
Returns the value of attribute level.
6 7 8 |
# File 'lib/sparkql/lexer.rb', line 6 def level @level end |
#token_index ⇒ Object (readonly)
Returns the value of attribute token_index.
8 9 10 |
# File 'lib/sparkql/lexer.rb', line 8 def token_index @token_index end |
Instance Method Details
#check_keywords(value) ⇒ Object
93 94 95 96 97 98 99 |
# File 'lib/sparkql/lexer.rb', line 93 def check_keywords(value) result = check_reserved_words(value) if result.first == :UNKNOWN result = [:KEYWORD,value] end result end |
#check_reserved_words(value) ⇒ Object
69 70 71 72 73 74 75 76 77 78 79 80 81 82 |
# File 'lib/sparkql/lexer.rb', line 69 def check_reserved_words(value) u_value = value.capitalize if OPERATORS.include?(u_value) [:OPERATOR,u_value] elsif RANGE_OPERATOR == u_value [:RANGE_OPERATOR,u_value] elsif CONJUNCTIONS.include?(u_value) [:CONJUNCTION,u_value] elsif UNARY_CONJUNCTIONS.include?(u_value) [:UNARY_CONJUNCTION,u_value] else [:UNKNOWN, "ERROR: '#{self.string}'"] end end |
#check_standard_fields(value) ⇒ Object
84 85 86 87 88 89 90 91 |
# File 'lib/sparkql/lexer.rb', line 84 def check_standard_fields(value) result = check_reserved_words(value) if result.first == :UNKNOWN @last_field = value result = [:STANDARD_FIELD,value] end result end |
#leveldown ⇒ Object
106 107 108 |
# File 'lib/sparkql/lexer.rb', line 106 def leveldown @level -= 1 end |
#levelup ⇒ Object
101 102 103 104 |
# File 'lib/sparkql/lexer.rb', line 101 def levelup @level += 1 @block_group_identifier += 1 end |
#literal(symbol, value) ⇒ Object
110 111 112 113 114 115 116 |
# File 'lib/sparkql/lexer.rb', line 110 def literal(symbol, value) node = { :type => symbol.to_s.downcase.to_sym, :value => value } [symbol, node] end |
#shift ⇒ Object
Lookup the next matching token
TODO the old implementation did value type detection conversion at a later date, we can perform this at parse time if we want!!!!
22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 |
# File 'lib/sparkql/lexer.rb', line 22 def shift @token_index = self.pos token = case when @current_token_value = scan(SPACE) [:SPACE, @current_token_value] when @current_token_value = scan(LPAREN) levelup [:LPAREN, @current_token_value] when @current_token_value = scan(RPAREN) # leveldown: do this after parsing group [:RPAREN, @current_token_value] when @current_token_value = scan(/\,/) [:COMMA,@current_token_value] when @current_token_value = scan(NULL) literal :NULL, "NULL" when @current_token_value = scan(STANDARD_FIELD) check_standard_fields(@current_token_value) when @current_token_value = scan(DATETIME) literal :DATETIME, @current_token_value when @current_token_value = scan(DATE) literal :DATE, @current_token_value when @current_token_value = scan(TIME) literal :TIME, @current_token_value when @current_token_value = scan(DECIMAL) literal :DECIMAL, @current_token_value when @current_token_value = scan(INTEGER) literal :INTEGER, @current_token_value when @current_token_value = scan(/\-/) [:UMINUS, @current_token_value] when @current_token_value = scan(CHARACTER) literal :CHARACTER, @current_token_value when @current_token_value = scan(BOOLEAN) literal :BOOLEAN, @current_token_value when @current_token_value = scan(KEYWORD) check_keywords(@current_token_value) when @current_token_value = scan(CUSTOM_FIELD) [:CUSTOM_FIELD,@current_token_value] when eos? [false, false] # end of file, \Z don't work with StringScanner else [:UNKNOWN, "ERROR: '#{self.string}'"] end token.freeze end |