sqlglot.dialects.teradata
1from __future__ import annotations 2 3from sqlglot import tokens 4from sqlglot.dialects.dialect import Dialect 5from sqlglot.generators.teradata import TeradataGenerator 6from sqlglot.parsers.teradata import TeradataParser 7from sqlglot.tokens import TokenType 8 9 10class Teradata(Dialect): 11 TYPED_DIVISION = True 12 13 TIME_MAPPING = { 14 "YY": "%y", 15 "Y4": "%Y", 16 "YYYY": "%Y", 17 "M4": "%B", 18 "M3": "%b", 19 "M": "%-M", 20 "MI": "%M", 21 "MM": "%m", 22 "MMM": "%b", 23 "MMMM": "%B", 24 "D": "%-d", 25 "DD": "%d", 26 "D3": "%j", 27 "DDD": "%j", 28 "H": "%-H", 29 "HH": "%H", 30 "HH24": "%H", 31 "S": "%-S", 32 "SS": "%S", 33 "SSSSSS": "%f", 34 "E": "%a", 35 "EE": "%a", 36 "E3": "%a", 37 "E4": "%A", 38 "EEE": "%a", 39 "EEEE": "%A", 40 } 41 42 class Tokenizer(tokens.Tokenizer): 43 # Tested each of these and they work, although there is no 44 # Teradata documentation explicitly mentioning them. 45 HEX_STRINGS = [("X'", "'"), ("x'", "'"), ("0x", "")] 46 # https://docs.teradata.com/r/Teradata-Database-SQL-Functions-Operators-Exprs-and-Predicates/March-2017/Comparison-Operators-and-Functions/Comparison-Operators/ANSI-Compliance 47 # https://docs.teradata.com/r/SQL-Functions-Operators-Exprs-and-Predicates/June-2017/Arithmetic-Trigonometric-Hyperbolic-Operators/Functions 48 KEYWORDS = { 49 **tokens.Tokenizer.KEYWORDS, 50 "**": TokenType.DSTAR, 51 "^=": TokenType.NEQ, 52 "BYTEINT": TokenType.SMALLINT, 53 "COLLECT": TokenType.COMMAND, 54 "DEL": TokenType.DELETE, 55 "EQ": TokenType.EQ, 56 "GE": TokenType.GTE, 57 "GT": TokenType.GT, 58 "HELP": TokenType.COMMAND, 59 "INS": TokenType.INSERT, 60 "LE": TokenType.LTE, 61 "LOCKING": TokenType.LOCK, 62 "LT": TokenType.LT, 63 "MINUS": TokenType.EXCEPT, 64 "MOD": TokenType.MOD, 65 "NE": TokenType.NEQ, 66 "NOT=": TokenType.NEQ, 67 "SAMPLE": TokenType.TABLE_SAMPLE, 68 "SEL": TokenType.SELECT, 69 "ST_GEOMETRY": TokenType.GEOMETRY, 70 "TOP": TokenType.TOP, 71 "UPD": TokenType.UPDATE, 72 } 73 KEYWORDS.pop("/*+") 74 75 # Teradata does not support % as a modulo operator 76 SINGLE_TOKENS = {**tokens.Tokenizer.SINGLE_TOKENS} 77 SINGLE_TOKENS.pop("%") 78 79 Parser = TeradataParser 80 81 Generator = TeradataGenerator
class
Teradata(sqlglot.dialects.dialect.Dialect):
11class Teradata(Dialect): 12 TYPED_DIVISION = True 13 14 TIME_MAPPING = { 15 "YY": "%y", 16 "Y4": "%Y", 17 "YYYY": "%Y", 18 "M4": "%B", 19 "M3": "%b", 20 "M": "%-M", 21 "MI": "%M", 22 "MM": "%m", 23 "MMM": "%b", 24 "MMMM": "%B", 25 "D": "%-d", 26 "DD": "%d", 27 "D3": "%j", 28 "DDD": "%j", 29 "H": "%-H", 30 "HH": "%H", 31 "HH24": "%H", 32 "S": "%-S", 33 "SS": "%S", 34 "SSSSSS": "%f", 35 "E": "%a", 36 "EE": "%a", 37 "E3": "%a", 38 "E4": "%A", 39 "EEE": "%a", 40 "EEEE": "%A", 41 } 42 43 class Tokenizer(tokens.Tokenizer): 44 # Tested each of these and they work, although there is no 45 # Teradata documentation explicitly mentioning them. 46 HEX_STRINGS = [("X'", "'"), ("x'", "'"), ("0x", "")] 47 # https://docs.teradata.com/r/Teradata-Database-SQL-Functions-Operators-Exprs-and-Predicates/March-2017/Comparison-Operators-and-Functions/Comparison-Operators/ANSI-Compliance 48 # https://docs.teradata.com/r/SQL-Functions-Operators-Exprs-and-Predicates/June-2017/Arithmetic-Trigonometric-Hyperbolic-Operators/Functions 49 KEYWORDS = { 50 **tokens.Tokenizer.KEYWORDS, 51 "**": TokenType.DSTAR, 52 "^=": TokenType.NEQ, 53 "BYTEINT": TokenType.SMALLINT, 54 "COLLECT": TokenType.COMMAND, 55 "DEL": TokenType.DELETE, 56 "EQ": TokenType.EQ, 57 "GE": TokenType.GTE, 58 "GT": TokenType.GT, 59 "HELP": TokenType.COMMAND, 60 "INS": TokenType.INSERT, 61 "LE": TokenType.LTE, 62 "LOCKING": TokenType.LOCK, 63 "LT": TokenType.LT, 64 "MINUS": TokenType.EXCEPT, 65 "MOD": TokenType.MOD, 66 "NE": TokenType.NEQ, 67 "NOT=": TokenType.NEQ, 68 "SAMPLE": TokenType.TABLE_SAMPLE, 69 "SEL": TokenType.SELECT, 70 "ST_GEOMETRY": TokenType.GEOMETRY, 71 "TOP": TokenType.TOP, 72 "UPD": TokenType.UPDATE, 73 } 74 KEYWORDS.pop("/*+") 75 76 # Teradata does not support % as a modulo operator 77 SINGLE_TOKENS = {**tokens.Tokenizer.SINGLE_TOKENS} 78 SINGLE_TOKENS.pop("%") 79 80 Parser = TeradataParser 81 82 Generator = TeradataGenerator
TYPED_DIVISION =
True
Whether the behavior of a / b depends on the types of a and b.
False means a / b is always float division.
True means a / b is integer division if both a and b are integers.
TIME_MAPPING: dict[str, str] =
{'YY': '%y', 'Y4': '%Y', 'YYYY': '%Y', 'M4': '%B', 'M3': '%b', 'M': '%-M', 'MI': '%M', 'MM': '%m', 'MMM': '%b', 'MMMM': '%B', 'D': '%-d', 'DD': '%d', 'D3': '%j', 'DDD': '%j', 'H': '%-H', 'HH': '%H', 'HH24': '%H', 'S': '%-S', 'SS': '%S', 'SSSSSS': '%f', 'E': '%a', 'EE': '%a', 'E3': '%a', 'E4': '%A', 'EEE': '%a', 'EEEE': '%A'}
Associates this dialect's time formats with their equivalent Python strftime formats.
Parser =
<class 'sqlglot.parsers.teradata.TeradataParser'>
Generator =
<class 'sqlglot.generators.teradata.TeradataGenerator'>
STRINGS_SUPPORT_ESCAPED_SEQUENCES: bool =
False
Whether string literals support escape sequences (e.g. \n). Set by the metaclass based on the tokenizer's STRING_ESCAPES.
BYTE_STRINGS_SUPPORT_ESCAPED_SEQUENCES: bool =
False
Whether byte string literals support escape sequences. Set by the metaclass based on the tokenizer's BYTE_STRING_ESCAPES.
tokenizer_class =
<class 'Teradata.Tokenizer'>
parser_class =
<class 'sqlglot.parsers.teradata.TeradataParser'>
generator_class =
<class 'sqlglot.generators.teradata.TeradataGenerator'>
TIME_TRIE: dict =
{'Y': {'Y': {0: True, 'Y': {'Y': {0: True}}}, '4': {0: True}}, 'M': {'4': {0: True}, '3': {0: True}, 0: True, 'I': {0: True}, 'M': {0: True, 'M': {0: True, 'M': {0: True}}}}, 'D': {0: True, 'D': {0: True, 'D': {0: True}}, '3': {0: True}}, 'H': {0: True, 'H': {0: True, '2': {'4': {0: True}}}}, 'S': {0: True, 'S': {0: True, 'S': {'S': {'S': {'S': {0: True}}}}}}, 'E': {0: True, 'E': {0: True, 'E': {0: True, 'E': {0: True}}}, '3': {0: True}, '4': {0: True}}}
FORMAT_TRIE: dict =
{'Y': {'Y': {0: True, 'Y': {'Y': {0: True}}}, '4': {0: True}}, 'M': {'4': {0: True}, '3': {0: True}, 0: True, 'I': {0: True}, 'M': {0: True, 'M': {0: True, 'M': {0: True}}}}, 'D': {0: True, 'D': {0: True, 'D': {0: True}}, '3': {0: True}}, 'H': {0: True, 'H': {0: True, '2': {'4': {0: True}}}}, 'S': {0: True, 'S': {0: True, 'S': {'S': {'S': {'S': {0: True}}}}}}, 'E': {0: True, 'E': {0: True, 'E': {0: True, 'E': {0: True}}}, '3': {0: True}, '4': {0: True}}}
INVERSE_TIME_MAPPING: dict[str, str] =
{'%y': 'YY', '%Y': 'YYYY', '%B': 'MMMM', '%b': 'MMM', '%-M': 'M', '%M': 'MI', '%m': 'MM', '%-d': 'D', '%d': 'DD', '%j': 'DDD', '%-H': 'H', '%H': 'HH24', '%-S': 'S', '%S': 'SS', '%f': 'SSSSSS', '%a': 'EEE', '%A': 'EEEE'}
INVERSE_TIME_TRIE: dict =
{'%': {'y': {0: True}, 'Y': {0: True}, 'B': {0: True}, 'b': {0: True}, '-': {'M': {0: True}, 'd': {0: True}, 'H': {0: True}, 'S': {0: True}}, 'M': {0: True}, 'm': {0: True}, 'd': {0: True}, 'j': {0: True}, 'H': {0: True}, 'S': {0: True}, 'f': {0: True}, 'a': {0: True}, 'A': {0: True}}}
VALID_INTERVAL_UNITS: set[str] =
{'QTRS', 'TZH', 'HR', 'DAYS', 'US', 'NSEC', 'MICROSECS', 'WK', 'NSECOND', 'MSECONDS', 'WOY', 'MINUTE', 'WEEKDAY', 'MILLISEC', 'MILLISECS', 'MONS', 'MONTHS', 'NSECONDS', 'USECS', 'Y', 'YEAR', 'DAYOFWEEK', 'SECONDS', 'QTR', 'YYYY', 'MILLISECONDS', 'DECADE', 'MM', 'MSECS', 'HOUR', 'DAYOFWEEK_ISO', 'MICROSEC', 'CENTS', 'MS', 'DAYOFWEEKISO', 'MICROSECONDS', 'MILLISECOND', 'USECONDS', 'SECS', 'EPOCH_NANOSECONDS', 'QUARTER', 'USEC', 'HH', 'YRS', 'WEEK', 'EPOCH_MICROSECOND', 'DECADES', 'WEEKOFYEAR', 'WY', 'NANOSECOND', 'DOW', 'SECOND', 'SEC', 'MIL', 'CENTURY', 'HRS', 'DAYOFYEAR', 'YY', 'NANOSEC', 'NANOSECS', 'M', 'WEEKOFYEARISO', 'YEARS', 'Q', 'MILLENNIUM', 'WEEKISO', 'DAY OF WEEK', 'WEEK_ISO', 'MILS', 'MINS', 'NS', 'W', 'EPOCH', 'D', 'DAYOFMONTH', 'DAY', 'MILLISECON', 'CENTURIES', 'S', 'QUARTERS', 'WEEKDAY_ISO', 'WEEKOFYEAR_ISO', 'DEC', 'DW_ISO', 'DY', 'C', 'MONTH', 'YR', 'DW', 'YYY', 'EPOCH_SECOND', 'MIN', 'DD', 'DOW_ISO', 'MI', 'EPOCH_MICROSECONDS', 'MSEC', 'DAY OF YEAR', 'TZM', 'MINUTES', 'MON', 'MSECOND', 'DOY', 'EPOCH_MILLISECONDS', 'MILLENIA', 'EPOCH_NANOSECOND', 'EPOCH_MILLISECOND', 'MICROSECOND', 'EPOCH_SECONDS', 'TIMEZONE_HOUR', 'H', 'HOURS', 'TIMEZONE_MINUTE', 'CENT', 'DECS', 'USECOND'}
43 class Tokenizer(tokens.Tokenizer): 44 # Tested each of these and they work, although there is no 45 # Teradata documentation explicitly mentioning them. 46 HEX_STRINGS = [("X'", "'"), ("x'", "'"), ("0x", "")] 47 # https://docs.teradata.com/r/Teradata-Database-SQL-Functions-Operators-Exprs-and-Predicates/March-2017/Comparison-Operators-and-Functions/Comparison-Operators/ANSI-Compliance 48 # https://docs.teradata.com/r/SQL-Functions-Operators-Exprs-and-Predicates/June-2017/Arithmetic-Trigonometric-Hyperbolic-Operators/Functions 49 KEYWORDS = { 50 **tokens.Tokenizer.KEYWORDS, 51 "**": TokenType.DSTAR, 52 "^=": TokenType.NEQ, 53 "BYTEINT": TokenType.SMALLINT, 54 "COLLECT": TokenType.COMMAND, 55 "DEL": TokenType.DELETE, 56 "EQ": TokenType.EQ, 57 "GE": TokenType.GTE, 58 "GT": TokenType.GT, 59 "HELP": TokenType.COMMAND, 60 "INS": TokenType.INSERT, 61 "LE": TokenType.LTE, 62 "LOCKING": TokenType.LOCK, 63 "LT": TokenType.LT, 64 "MINUS": TokenType.EXCEPT, 65 "MOD": TokenType.MOD, 66 "NE": TokenType.NEQ, 67 "NOT=": TokenType.NEQ, 68 "SAMPLE": TokenType.TABLE_SAMPLE, 69 "SEL": TokenType.SELECT, 70 "ST_GEOMETRY": TokenType.GEOMETRY, 71 "TOP": TokenType.TOP, 72 "UPD": TokenType.UPDATE, 73 } 74 KEYWORDS.pop("/*+") 75 76 # Teradata does not support % as a modulo operator 77 SINGLE_TOKENS = {**tokens.Tokenizer.SINGLE_TOKENS} 78 SINGLE_TOKENS.pop("%")
KEYWORDS =
{'{%': <TokenType.BLOCK_START: 71>, '{%+': <TokenType.BLOCK_START: 71>, '{%-': <TokenType.BLOCK_START: 71>, '%}': <TokenType.BLOCK_END: 72>, '+%}': <TokenType.BLOCK_END: 72>, '-%}': <TokenType.BLOCK_END: 72>, '{{+': <TokenType.BLOCK_START: 71>, '{{-': <TokenType.BLOCK_START: 71>, '+}}': <TokenType.BLOCK_END: 72>, '-}}': <TokenType.BLOCK_END: 72>, '&<': <TokenType.AMP_LT: 61>, '&>': <TokenType.AMP_GT: 62>, '==': <TokenType.EQ: 28>, '::': <TokenType.DCOLON: 14>, '?::': <TokenType.QDCOLON: 366>, '||': <TokenType.DPIPE: 37>, '|>': <TokenType.PIPE_GT: 38>, '>=': <TokenType.GTE: 26>, '<=': <TokenType.LTE: 24>, '<>': <TokenType.NEQ: 29>, '!=': <TokenType.NEQ: 29>, ':=': <TokenType.COLON_EQ: 31>, '<=>': <TokenType.NULLSAFE_EQ: 30>, '->': <TokenType.ARROW: 45>, '->>': <TokenType.DARROW: 46>, '=>': <TokenType.FARROW: 47>, '#>': <TokenType.HASH_ARROW: 49>, '#>>': <TokenType.DHASH_ARROW: 50>, '<->': <TokenType.LR_ARROW: 51>, '&&': <TokenType.DAMP: 60>, '??': <TokenType.DQMARK: 18>, '~~~': <TokenType.GLOB: 284>, '~~': <TokenType.LIKE: 315>, '~~*': <TokenType.ILIKE: 292>, '~*': <TokenType.IRLIKE: 304>, '-|-': <TokenType.ADJACENT: 63>, 'ALL': <TokenType.ALL: 218>, 'AND': <TokenType.AND: 34>, 'ANTI': <TokenType.ANTI: 219>, 'ANY': <TokenType.ANY: 220>, 'ASC': <TokenType.ASC: 223>, 'AS': <TokenType.ALIAS: 216>, 'ASOF': <TokenType.ASOF: 224>, 'AUTOINCREMENT': <TokenType.AUTO_INCREMENT: 226>, 'AUTO_INCREMENT': <TokenType.AUTO_INCREMENT: 226>, 'BEGIN': <TokenType.BEGIN: 227>, 'BETWEEN': <TokenType.BETWEEN: 228>, 'CACHE': <TokenType.CACHE: 230>, 'UNCACHE': <TokenType.UNCACHE: 409>, 'CASE': <TokenType.CASE: 231>, 'CHARACTER SET': <TokenType.CHARACTER_SET: 232>, 'CLUSTER BY': <TokenType.CLUSTER_BY: 233>, 'COLLATE': <TokenType.COLLATE: 234>, 'COLUMN': <TokenType.COLUMN: 79>, 'COMMIT': <TokenType.COMMIT: 237>, 'CONNECT BY': <TokenType.CONNECT_BY: 238>, 'CONSTRAINT': <TokenType.CONSTRAINT: 239>, 'COPY': <TokenType.COPY: 240>, 'CREATE': <TokenType.CREATE: 241>, 'CROSS': <TokenType.CROSS: 242>, 'CUBE': <TokenType.CUBE: 243>, 'CURRENT_DATE': <TokenType.CURRENT_DATE: 244>, 'CURRENT_SCHEMA': <TokenType.CURRENT_SCHEMA: 246>, 'CURRENT_TIME': <TokenType.CURRENT_TIME: 247>, 'CURRENT_TIMESTAMP': <TokenType.CURRENT_TIMESTAMP: 248>, 'CURRENT_USER': <TokenType.CURRENT_USER: 249>, 'CURRENT_CATALOG': <TokenType.CURRENT_CATALOG: 251>, 'DATABASE': <TokenType.DATABASE: 78>, 'DEFAULT': <TokenType.DEFAULT: 253>, 'DELETE': <TokenType.DELETE: 254>, 'DESC': <TokenType.DESC: 255>, 'DESCRIBE': <TokenType.DESCRIBE: 256>, 'DISTINCT': <TokenType.DISTINCT: 259>, 'DISTRIBUTE BY': <TokenType.DISTRIBUTE_BY: 260>, 'DIV': <TokenType.DIV: 261>, 'DROP': <TokenType.DROP: 262>, 'ELSE': <TokenType.ELSE: 263>, 'END': <TokenType.END: 264>, 'ENUM': <TokenType.ENUM: 203>, 'ESCAPE': <TokenType.ESCAPE: 265>, 'EXCEPT': <TokenType.EXCEPT: 266>, 'EXECUTE': <TokenType.EXECUTE: 267>, 'EXISTS': <TokenType.EXISTS: 268>, 'FALSE': <TokenType.FALSE: 269>, 'FETCH': <TokenType.FETCH: 270>, 'FILTER': <TokenType.FILTER: 273>, 'FILE': <TokenType.FILE: 271>, 'FIRST': <TokenType.FIRST: 275>, 'FULL': <TokenType.FULL: 281>, 'FUNCTION': <TokenType.FUNCTION: 282>, 'FOR': <TokenType.FOR: 276>, 'FOREIGN KEY': <TokenType.FOREIGN_KEY: 278>, 'FORMAT': <TokenType.FORMAT: 279>, 'FROM': <TokenType.FROM: 280>, 'GEOGRAPHY': <TokenType.GEOGRAPHY: 170>, 'GEOMETRY': <TokenType.GEOMETRY: 173>, 'GLOB': <TokenType.GLOB: 284>, 'GROUP BY': <TokenType.GROUP_BY: 287>, 'GROUPING SETS': <TokenType.GROUPING_SETS: 288>, 'HAVING': <TokenType.HAVING: 289>, 'ILIKE': <TokenType.ILIKE: 292>, 'IN': <TokenType.IN: 293>, 'INDEX': <TokenType.INDEX: 294>, 'INET': <TokenType.INET: 198>, 'INNER': <TokenType.INNER: 296>, 'INSERT': <TokenType.INSERT: 297>, 'INTERVAL': <TokenType.INTERVAL: 301>, 'INTERSECT': <TokenType.INTERSECT: 300>, 'INTO': <TokenType.INTO: 302>, 'IS': <TokenType.IS: 305>, 'ISNULL': <TokenType.ISNULL: 306>, 'JOIN': <TokenType.JOIN: 307>, 'KEEP': <TokenType.KEEP: 309>, 'KILL': <TokenType.KILL: 311>, 'LATERAL': <TokenType.LATERAL: 313>, 'LEFT': <TokenType.LEFT: 314>, 'LIKE': <TokenType.LIKE: 315>, 'LIMIT': <TokenType.LIMIT: 316>, 'LOAD': <TokenType.LOAD: 318>, 'LOCALTIME': <TokenType.LOCALTIME: 177>, 'LOCALTIMESTAMP': <TokenType.LOCALTIMESTAMP: 178>, 'LOCK': <TokenType.LOCK: 319>, 'MERGE': <TokenType.MERGE: 325>, 'NAMESPACE': <TokenType.NAMESPACE: 436>, 'NATURAL': <TokenType.NATURAL: 328>, 'NEXT': <TokenType.NEXT: 329>, 'NOT': <TokenType.NOT: 27>, 'NOTNULL': <TokenType.NOTNULL: 331>, 'NULL': <TokenType.NULL: 332>, 'OBJECT': <TokenType.OBJECT: 197>, 'OFFSET': <TokenType.OFFSET: 334>, 'ON': <TokenType.ON: 335>, 'OR': <TokenType.OR: 35>, 'XOR': <TokenType.XOR: 64>, 'ORDER BY': <TokenType.ORDER_BY: 338>, 'ORDINALITY': <TokenType.ORDINALITY: 341>, 'OUT': <TokenType.OUT: 342>, 'OUTER': <TokenType.OUTER: 344>, 'OVER': <TokenType.OVER: 345>, 'OVERLAPS': <TokenType.OVERLAPS: 346>, 'OVERWRITE': <TokenType.OVERWRITE: 347>, 'PARTITION': <TokenType.PARTITION: 349>, 'PARTITION BY': <TokenType.PARTITION_BY: 350>, 'PARTITIONED BY': <TokenType.PARTITION_BY: 350>, 'PARTITIONED_BY': <TokenType.PARTITION_BY: 350>, 'PERCENT': <TokenType.PERCENT: 351>, 'PIVOT': <TokenType.PIVOT: 352>, 'PRAGMA': <TokenType.PRAGMA: 357>, 'PRIMARY KEY': <TokenType.PRIMARY_KEY: 359>, 'PROCEDURE': <TokenType.PROCEDURE: 360>, 'OPERATOR': <TokenType.OPERATOR: 337>, 'QUALIFY': <TokenType.QUALIFY: 364>, 'RANGE': <TokenType.RANGE: 367>, 'RECURSIVE': <TokenType.RECURSIVE: 368>, 'REGEXP': <TokenType.RLIKE: 376>, 'RENAME': <TokenType.RENAME: 370>, 'REPLACE': <TokenType.REPLACE: 371>, 'RETURNING': <TokenType.RETURNING: 372>, 'REFERENCES': <TokenType.REFERENCES: 374>, 'RIGHT': <TokenType.RIGHT: 375>, 'RLIKE': <TokenType.RLIKE: 376>, 'ROLLBACK': <TokenType.ROLLBACK: 378>, 'ROLLUP': <TokenType.ROLLUP: 379>, 'ROW': <TokenType.ROW: 380>, 'ROWS': <TokenType.ROWS: 381>, 'SCHEMA': <TokenType.SCHEMA: 81>, 'SELECT': <TokenType.SELECT: 383>, 'SEMI': <TokenType.SEMI: 384>, 'SESSION': <TokenType.SESSION: 57>, 'SESSION_USER': <TokenType.SESSION_USER: 59>, 'SET': <TokenType.SET: 388>, 'SETTINGS': <TokenType.SETTINGS: 389>, 'SHOW': <TokenType.SHOW: 390>, 'SIMILAR TO': <TokenType.SIMILAR_TO: 391>, 'SOME': <TokenType.SOME: 392>, 'SORT BY': <TokenType.SORT_BY: 393>, 'SQL SECURITY': <TokenType.SQL_SECURITY: 395>, 'START WITH': <TokenType.START_WITH: 396>, 'STRAIGHT_JOIN': <TokenType.STRAIGHT_JOIN: 398>, 'TABLE': <TokenType.TABLE: 82>, 'TABLESAMPLE': <TokenType.TABLE_SAMPLE: 401>, 'TEMP': <TokenType.TEMPORARY: 403>, 'TEMPORARY': <TokenType.TEMPORARY: 403>, 'THEN': <TokenType.THEN: 405>, 'TRUE': <TokenType.TRUE: 406>, 'TRUNCATE': <TokenType.TRUNCATE: 407>, 'TRIGGER': <TokenType.TRIGGER: 408>, 'UNION': <TokenType.UNION: 410>, 'UNKNOWN': <TokenType.UNKNOWN: 212>, 'UNNEST': <TokenType.UNNEST: 411>, 'UNPIVOT': <TokenType.UNPIVOT: 412>, 'UPDATE': <TokenType.UPDATE: 413>, 'USE': <TokenType.USE: 414>, 'USING': <TokenType.USING: 415>, 'UUID': <TokenType.UUID: 169>, 'VALUES': <TokenType.VALUES: 416>, 'VIEW': <TokenType.VIEW: 418>, 'VOLATILE': <TokenType.VOLATILE: 420>, 'WHEN': <TokenType.WHEN: 422>, 'WHERE': <TokenType.WHERE: 423>, 'WINDOW': <TokenType.WINDOW: 424>, 'WITH': <TokenType.WITH: 425>, 'APPLY': <TokenType.APPLY: 221>, 'ARRAY': <TokenType.ARRAY: 222>, 'BIT': <TokenType.BIT: 95>, 'BOOL': <TokenType.BOOLEAN: 96>, 'BOOLEAN': <TokenType.BOOLEAN: 96>, 'BYTE': <TokenType.TINYINT: 97>, 'MEDIUMINT': <TokenType.MEDIUMINT: 101>, 'INT1': <TokenType.TINYINT: 97>, 'TINYINT': <TokenType.TINYINT: 97>, 'INT16': <TokenType.SMALLINT: 99>, 'SHORT': <TokenType.SMALLINT: 99>, 'SMALLINT': <TokenType.SMALLINT: 99>, 'HUGEINT': <TokenType.INT128: 108>, 'UHUGEINT': <TokenType.UINT128: 109>, 'INT2': <TokenType.SMALLINT: 99>, 'INTEGER': <TokenType.INT: 103>, 'INT': <TokenType.INT: 103>, 'INT4': <TokenType.INT: 103>, 'INT32': <TokenType.INT: 103>, 'INT64': <TokenType.BIGINT: 105>, 'INT128': <TokenType.INT128: 108>, 'INT256': <TokenType.INT256: 110>, 'LONG': <TokenType.BIGINT: 105>, 'BIGINT': <TokenType.BIGINT: 105>, 'INT8': <TokenType.TINYINT: 97>, 'UINT': <TokenType.UINT: 104>, 'UINT128': <TokenType.UINT128: 109>, 'UINT256': <TokenType.UINT256: 111>, 'DEC': <TokenType.DECIMAL: 115>, 'DECIMAL': <TokenType.DECIMAL: 115>, 'DECIMAL32': <TokenType.DECIMAL32: 116>, 'DECIMAL64': <TokenType.DECIMAL64: 117>, 'DECIMAL128': <TokenType.DECIMAL128: 118>, 'DECIMAL256': <TokenType.DECIMAL256: 119>, 'DECFLOAT': <TokenType.DECFLOAT: 120>, 'BIGDECIMAL': <TokenType.BIGDECIMAL: 122>, 'BIGNUMERIC': <TokenType.BIGDECIMAL: 122>, 'BIGNUM': <TokenType.BIGNUM: 107>, 'LIST': <TokenType.LIST: 317>, 'MAP': <TokenType.MAP: 320>, 'NULLABLE': <TokenType.NULLABLE: 172>, 'NUMBER': <TokenType.DECIMAL: 115>, 'NUMERIC': <TokenType.DECIMAL: 115>, 'FIXED': <TokenType.DECIMAL: 115>, 'REAL': <TokenType.FLOAT: 112>, 'FLOAT': <TokenType.FLOAT: 112>, 'FLOAT4': <TokenType.FLOAT: 112>, 'FLOAT8': <TokenType.DOUBLE: 113>, 'DOUBLE': <TokenType.DOUBLE: 113>, 'DOUBLE PRECISION': <TokenType.DOUBLE: 113>, 'JSON': <TokenType.JSON: 139>, 'JSONB': <TokenType.JSONB: 140>, 'CHAR': <TokenType.CHAR: 123>, 'CHARACTER': <TokenType.CHAR: 123>, 'CHAR VARYING': <TokenType.VARCHAR: 125>, 'CHARACTER VARYING': <TokenType.VARCHAR: 125>, 'NCHAR': <TokenType.NCHAR: 124>, 'VARCHAR': <TokenType.VARCHAR: 125>, 'VARCHAR2': <TokenType.VARCHAR: 125>, 'NVARCHAR': <TokenType.NVARCHAR: 126>, 'NVARCHAR2': <TokenType.NVARCHAR: 126>, 'BPCHAR': <TokenType.BPCHAR: 127>, 'STR': <TokenType.TEXT: 128>, 'STRING': <TokenType.TEXT: 128>, 'TEXT': <TokenType.TEXT: 128>, 'LONGTEXT': <TokenType.LONGTEXT: 130>, 'MEDIUMTEXT': <TokenType.MEDIUMTEXT: 129>, 'TINYTEXT': <TokenType.TINYTEXT: 135>, 'CLOB': <TokenType.TEXT: 128>, 'LONGVARCHAR': <TokenType.TEXT: 128>, 'BINARY': <TokenType.BINARY: 137>, 'BLOB': <TokenType.VARBINARY: 138>, 'LONGBLOB': <TokenType.LONGBLOB: 133>, 'MEDIUMBLOB': <TokenType.MEDIUMBLOB: 132>, 'TINYBLOB': <TokenType.TINYBLOB: 134>, 'BYTEA': <TokenType.VARBINARY: 138>, 'VARBINARY': <TokenType.VARBINARY: 138>, 'TIME': <TokenType.TIME: 141>, 'TIMETZ': <TokenType.TIMETZ: 142>, 'TIME_NS': <TokenType.TIME_NS: 143>, 'TIMESTAMP': <TokenType.TIMESTAMP: 144>, 'TIMESTAMPTZ': <TokenType.TIMESTAMPTZ: 145>, 'TIMESTAMPLTZ': <TokenType.TIMESTAMPLTZ: 146>, 'TIMESTAMP_LTZ': <TokenType.TIMESTAMPLTZ: 146>, 'TIMESTAMPNTZ': <TokenType.TIMESTAMPNTZ: 147>, 'TIMESTAMP_NTZ': <TokenType.TIMESTAMPNTZ: 147>, 'DATE': <TokenType.DATE: 155>, 'DATETIME': <TokenType.DATETIME: 151>, 'INT4RANGE': <TokenType.INT4RANGE: 157>, 'INT4MULTIRANGE': <TokenType.INT4MULTIRANGE: 158>, 'INT8RANGE': <TokenType.INT8RANGE: 159>, 'INT8MULTIRANGE': <TokenType.INT8MULTIRANGE: 160>, 'NUMRANGE': <TokenType.NUMRANGE: 161>, 'NUMMULTIRANGE': <TokenType.NUMMULTIRANGE: 162>, 'TSRANGE': <TokenType.TSRANGE: 163>, 'TSMULTIRANGE': <TokenType.TSMULTIRANGE: 164>, 'TSTZRANGE': <TokenType.TSTZRANGE: 165>, 'TSTZMULTIRANGE': <TokenType.TSTZMULTIRANGE: 166>, 'DATERANGE': <TokenType.DATERANGE: 167>, 'DATEMULTIRANGE': <TokenType.DATEMULTIRANGE: 168>, 'UNIQUE': <TokenType.UNIQUE: 426>, 'VECTOR': <TokenType.VECTOR: 213>, 'STRUCT': <TokenType.STRUCT: 399>, 'SEQUENCE': <TokenType.SEQUENCE: 386>, 'VARIANT': <TokenType.VARIANT: 196>, 'ALTER': <TokenType.ALTER: 217>, 'ANALYZE': <TokenType.ANALYZE: 435>, 'CALL': <TokenType.COMMAND: 235>, 'COMMENT': <TokenType.COMMENT: 236>, 'EXPLAIN': <TokenType.COMMAND: 235>, 'GRANT': <TokenType.GRANT: 286>, 'REVOKE': <TokenType.REVOKE: 373>, 'OPTIMIZE': <TokenType.COMMAND: 235>, 'PREPARE': <TokenType.COMMAND: 235>, 'VACUUM': <TokenType.COMMAND: 235>, 'USER-DEFINED': <TokenType.USERDEFINED: 191>, 'FOR VERSION': <TokenType.VERSION_SNAPSHOT: 430>, 'FOR TIMESTAMP': <TokenType.TIMESTAMP_SNAPSHOT: 431>, '**': <TokenType.DSTAR: 65>, '^=': <TokenType.NEQ: 29>, 'BYTEINT': <TokenType.SMALLINT: 99>, 'COLLECT': <TokenType.COMMAND: 235>, 'DEL': <TokenType.DELETE: 254>, 'EQ': <TokenType.EQ: 28>, 'GE': <TokenType.GTE: 26>, 'GT': <TokenType.GT: 25>, 'HELP': <TokenType.COMMAND: 235>, 'INS': <TokenType.INSERT: 297>, 'LE': <TokenType.LTE: 24>, 'LOCKING': <TokenType.LOCK: 319>, 'LT': <TokenType.LT: 23>, 'MINUS': <TokenType.EXCEPT: 266>, 'MOD': <TokenType.MOD: 326>, 'NE': <TokenType.NEQ: 29>, 'NOT=': <TokenType.NEQ: 29>, 'SAMPLE': <TokenType.TABLE_SAMPLE: 401>, 'SEL': <TokenType.SELECT: 383>, 'ST_GEOMETRY': <TokenType.GEOMETRY: 173>, 'TOP': <TokenType.TOP: 404>, 'UPD': <TokenType.UPDATE: 413>}
SINGLE_TOKENS =
{'(': <TokenType.L_PAREN: 1>, ')': <TokenType.R_PAREN: 2>, '[': <TokenType.L_BRACKET: 3>, ']': <TokenType.R_BRACKET: 4>, '{': <TokenType.L_BRACE: 5>, '}': <TokenType.R_BRACE: 6>, '&': <TokenType.AMP: 36>, '^': <TokenType.CARET: 42>, ':': <TokenType.COLON: 11>, ',': <TokenType.COMMA: 7>, '.': <TokenType.DOT: 8>, '-': <TokenType.DASH: 9>, '=': <TokenType.EQ: 28>, '>': <TokenType.GT: 25>, '<': <TokenType.LT: 23>, '!': <TokenType.NOT: 27>, '|': <TokenType.PIPE: 39>, '+': <TokenType.PLUS: 10>, ';': <TokenType.SEMICOLON: 19>, '/': <TokenType.SLASH: 22>, '\\': <TokenType.BACKSLASH: 21>, '*': <TokenType.STAR: 20>, '~': <TokenType.TILDE: 44>, '?': <TokenType.PLACEHOLDER: 353>, '@': <TokenType.PARAMETER: 56>, '#': <TokenType.HASH: 48>, "'": <TokenType.UNKNOWN: 212>, '`': <TokenType.UNKNOWN: 212>, '"': <TokenType.UNKNOWN: 212>}
Inherited Members
- sqlglot.tokens.Tokenizer
- Tokenizer
- BIT_STRINGS
- BYTE_STRINGS
- RAW_STRINGS
- HEREDOC_STRINGS
- UNICODE_STRINGS
- IDENTIFIERS
- QUOTES
- STRING_ESCAPES
- VAR_SINGLE_TOKENS
- ESCAPE_FOLLOW_CHARS
- IDENTIFIER_ESCAPES
- HEREDOC_TAG_IS_IDENTIFIER
- HEREDOC_STRING_ALTERNATIVE
- STRING_ESCAPES_ALLOWED_IN_RAW_STRINGS
- NESTED_COMMENTS
- HINT_START
- TOKENS_PRECEDING_HINT
- COMMANDS
- COMMAND_PREFIX_TOKENS
- NUMERIC_LITERALS
- NUMBERS_CAN_HAVE_DECIMALS
- COMMENTS
- dialect
- tokenize
- sql
- size
- tokens