SQLGlot is a no-dependency SQL parser, transpiler, optimizer, and engine. It can be used to format SQL or translate between 31 different dialects like DuckDB, Presto / Trino, Spark / Databricks, Snowflake, and BigQuery. It aims to read a wide variety of SQL inputs and output syntactically and semantically correct SQL in the targeted dialects.
It is a very comprehensive generic SQL parser with a robust test suite. It is also quite performant, while being written purely in Python.
You can easily customize the parser, analyze queries, traverse expression trees, and programmatically build SQL.
SQLGlot can detect a variety of syntax errors, such as unbalanced parentheses, incorrect usage of reserved keywords, and so on. These errors are highlighted and dialect incompatibilities can warn or raise depending on configurations.
Learn more about SQLGlot in the API documentation and the expression tree primer.
Contributions are very welcome in SQLGlot; read the contribution guide and the onboarding document to get started!
Table of Contents
- Install
- Versioning
- Get in Touch
- FAQ
- Examples
- Used By
- Documentation
- Run Tests and Lint
- Deployment
- Benchmarks
- Optional Dependencies
- Supported Dialects
Install
From PyPI:
# Pure python version
pip3 install sqlglot
# C extensions compiled with mypyc
# prebuilt wheel if available for your platform, otherwise builds from source
pip3 install "sqlglot[c]"
Or with a local checkout:
# Optionally prefix with UV=1 to use uv for the installation
make install
Requirements for development (optional):
# Optionally prefix with UV=1 to use uv for the installation
make install-dev
Versioning
Given a version number MAJOR.MINOR.PATCH, SQLGlot uses the following versioning strategy:
- The
PATCHversion is incremented when there are backwards-compatible fixes or feature additions. - The
MINORversion is incremented when there are backwards-incompatible fixes or feature additions. - The
MAJORversion is incremented when there are significant backwards-incompatible fixes or feature additions.
Get in Touch
We'd love to hear from you. Join our community Slack channel!
FAQ
I tried to parse SQL that should be valid but it failed, why did that happen?
- Most of the time, issues like this occur because the "source" dialect is omitted during parsing. For example, this is how to correctly parse a SQL query written in Spark SQL:
parse_one(sql, dialect="spark")(alternatively:read="spark"). If no dialect is specified,parse_onewill attempt to parse the query according to the "SQLGlot dialect", which is designed to be a superset of all supported dialects. If you tried specifying the dialect and it still doesn't work, please file an issue.
I tried to output SQL but it's not in the correct dialect!
- Like parsing, generating SQL also requires the target dialect to be specified, otherwise the SQLGlot dialect will be used by default. For example, to transpile a query from Spark SQL to DuckDB, do
parse_one(sql, dialect="spark").sql(dialect="duckdb")(alternatively:transpile(sql, read="spark", write="duckdb")).
What happened to sqlglot.dataframe?
- The PySpark dataframe api was moved to a standalone library called SQLFrame in v24. It now allows you to run queries as opposed to just generate SQL.
Examples
Formatting and Transpiling
Easily translate from one dialect to another. For example, date/time functions vary between dialects and can be hard to deal with:
import sqlglot
sqlglot.transpile("SELECT EPOCH_MS(1618088028295)", read="duckdb", write="hive")[0]
'SELECT FROM_UNIXTIME(1618088028295 / POW(10, 3))'
SQLGlot can even translate custom time formats:
import sqlglot
sqlglot.transpile("SELECT STRFTIME(x, '%y-%-m-%S')", read="duckdb", write="hive")[0]
"SELECT DATE_FORMAT(x, 'yy-M-ss')"
Identifier delimiters and data types can be translated as well:
import sqlglot
# Spark SQL requires backticks (`) for delimited identifiers and uses `FLOAT` over `REAL`
sql = """WITH baz AS (SELECT a, c FROM foo WHERE a = 1) SELECT f.a, b.b, baz.c, CAST("b"."a" AS REAL) d FROM foo f JOIN bar b ON f.a = b.a LEFT JOIN baz ON f.a = baz.a"""
# Translates the query into Spark SQL, formats it, and delimits all of its identifiers
print(sqlglot.transpile(sql, write="spark", identify=True, pretty=True)[0])
WITH `baz` AS (
SELECT
`a`,
`c`
FROM `foo`
WHERE
`a` = 1
)
SELECT
`f`.`a`,
`b`.`b`,
`baz`.`c`,
CAST(`b`.`a` AS FLOAT) AS `d`
FROM `foo` AS `f`
JOIN `bar` AS `b`
ON `f`.`a` = `b`.`a`
LEFT JOIN `baz`
ON `f`.`a` = `baz`.`a`
Comments are also preserved on a best-effort basis:
sql = """
/* multi
line
comment
*/
SELECT
tbl.cola /* comment 1 */ + tbl.colb /* comment 2 */,
CAST(x AS SIGNED), # comment 3
y -- comment 4
FROM
bar /* comment 5 */,
tbl # comment 6
"""
# Note: MySQL-specific comments (`#`) are converted into standard syntax
print(sqlglot.transpile(sql, read='mysql', pretty=True)[0])
/* multi
line
comment
*/
SELECT
tbl.cola /* comment 1 */ + tbl.colb /* comment 2 */,
CAST(x AS INT), /* comment 3 */
y /* comment 4 */
FROM bar /* comment 5 */, tbl /* comment 6 */
Metadata
You can explore SQL with expression helpers to do things like find columns and tables in a query:
from sqlglot import parse_one, exp
# print all column references (a and b)
for column in parse_one("SELECT a, b + 1 AS c FROM d").find_all(exp.Column):
print(column.alias_or_name)
# find all projections in select statements (a and c)
for select in parse_one("SELECT a, b + 1 AS c FROM d").find_all(exp.Select):
for projection in select.expressions:
print(projection.alias_or_name)
# find all tables (x, y, z)
for table in parse_one("SELECT * FROM x JOIN y JOIN z").find_all(exp.Table):
print(table.name)
Read the ast primer to learn more about SQLGlot's internals.
Parser Errors
When the parser detects an error in the syntax, it raises a ParseError:
import sqlglot
sqlglot.transpile("SELECT foo FROM (SELECT baz FROM t")
sqlglot.errors.ParseError: Expecting ). Line 1, Col: 34.
SELECT foo FROM (SELECT baz FROM t
~
Structured syntax errors are accessible for programmatic use:
import sqlglot.errors
try:
sqlglot.transpile("SELECT foo FROM (SELECT baz FROM t")
except sqlglot.errors.ParseError as e:
print(e.errors)
[{
'description': 'Expecting )',
'line': 1,
'col': 34,
'start_context': 'SELECT foo FROM (SELECT baz FROM ',
'highlight': 't',
'end_context': '',
'into_expression': None
}]
Unsupported Errors
It may not be possible to translate some queries between certain dialects. For these cases, SQLGlot may emit a warning and will proceed to do a best-effort translation by default:
import sqlglot
sqlglot.transpile("SELECT APPROX_DISTINCT(a, 0.1) FROM foo", read="presto", write="hive")
APPROX_COUNT_DISTINCT does not support accuracy
'SELECT APPROX_COUNT_DISTINCT(a) FROM foo'
This behavior can be changed by setting the unsupported_level attribute. For example, we can set it to either RAISE or IMMEDIATE to ensure an exception is raised instead:
import sqlglot
sqlglot.transpile("SELECT APPROX_DISTINCT(a, 0.1) FROM foo", read="presto", write="hive", unsupported_level=sqlglot.ErrorLevel.RAISE)
sqlglot.errors.UnsupportedError: APPROX_COUNT_DISTINCT does not support accuracy
There are queries that require additional information to be accurately transpiled, such as the schemas of the tables referenced in them. This is because certain transformations are type-sensitive, meaning that type inference is needed in order to understand their semantics. Even though the qualify and annotate_types optimizer rules can help with this, they are not used by default because they add significant overhead and complexity.
Transpilation is generally a hard problem, so SQLGlot employs an "incremental" approach to solving it. This means that there may be dialect pairs that currently lack support for some inputs, but this is expected to improve over time. We highly appreciate well-documented and tested issues or PRs, so feel free to reach out if you need guidance!
Build and Modify SQL
SQLGlot supports incrementally building SQL expressions:
from sqlglot import select, condition
where = condition("x=1").and_("y=1")
select("*").from_("y").where(where).sql()
'SELECT * FROM y WHERE x = 1 AND y = 1'
It's possible to modify a parsed tree:
from sqlglot import parse_one
parse_one("SELECT x FROM y").from_("z").sql()
'SELECT x FROM z'
Parsed expressions can also be transformed recursively by applying a mapping function to each node in the tree:
from sqlglot import exp, parse_one
expression_tree = parse_one("SELECT a FROM x")
def transformer(node):
if isinstance(node, exp.Column) and node.name == "a":
return parse_one("FUN(a)")
return node
transformed_tree = expression_tree.transform(transformer)
transformed_tree.sql()
'SELECT FUN(a) FROM x'
SQL Optimizer
SQLGlot can rewrite queries into an "optimized" form. It performs a variety of techniques to create a new canonical AST. This AST can be used to standardize queries or provide the foundations for implementing an actual engine. For example:
import sqlglot
from sqlglot.optimizer import optimize
print(
optimize(
sqlglot.parse_one("""
SELECT A OR (B OR (C AND D))
FROM x
WHERE Z = date '2021-01-01' + INTERVAL '1' month OR 1 = 0
"""),
schema={"x": {"A": "INT", "B": "INT", "C": "INT", "D": "INT", "Z": "STRING"}}
).sql(pretty=True)
)
SELECT
(
"x"."a" <> 0 OR "x"."b" <> 0 OR "x"."c" <> 0
)
AND (
"x"."a" <> 0 OR "x"."b" <> 0 OR "x"."d" <> 0
) AS "_col_0"
FROM "x" AS "x"
WHERE
CAST("x"."z" AS DATE) = CAST('2021-02-01' AS DATE)
AST Introspection
You can see the AST version of the parsed SQL by calling repr:
from sqlglot import parse_one
print(repr(parse_one("SELECT a + 1 AS z")))
Select(
expressions=[
Alias(
this=Add(
this=Column(
this=Identifier(this=a, quoted=False)),
expression=Literal(this=1, is_string=False)),
alias=Identifier(this=z, quoted=False))])
AST Diff
SQLGlot can calculate the semantic difference between two expressions and output changes in a form of a sequence of actions needed to transform a source expression into a target one:
from sqlglot import diff, parse_one
diff(parse_one("SELECT a + b, c, d"), parse_one("SELECT c, a - b, d"))
[
Remove(expression=Add(
this=Column(
this=Identifier(this=a, quoted=False)),
expression=Column(
this=Identifier(this=b, quoted=False)))),
Insert(expression=Sub(
this=Column(
this=Identifier(this=a, quoted=False)),
expression=Column(
this=Identifier(this=b, quoted=False)))),
Keep(
source=Column(this=Identifier(this=a, quoted=False)),
target=Column(this=Identifier(this=a, quoted=False))),
...
]
See also: Semantic Diff for SQL.
Custom Dialects
Dialects can be added by subclassing Dialect:
from sqlglot import exp
from sqlglot.dialects.dialect import Dialect
from sqlglot.generator import Generator
from sqlglot.tokens import Tokenizer, TokenType
class Custom(Dialect):
class Tokenizer(Tokenizer):
QUOTES = ["'", '"']
IDENTIFIERS = ["`"]
KEYWORDS = {
**Tokenizer.KEYWORDS,
"INT64": TokenType.BIGINT,
"FLOAT64": TokenType.DOUBLE,
}
class Generator(Generator):
TRANSFORMS = {exp.Array: lambda self, e: f"[{self.expressions(e)}]"}
TYPE_MAPPING = {
exp.DataType.Type.TINYINT: "INT64",
exp.DataType.Type.SMALLINT: "INT64",
exp.DataType.Type.INT: "INT64",
exp.DataType.Type.BIGINT: "INT64",
exp.DataType.Type.DECIMAL: "NUMERIC",
exp.DataType.Type.FLOAT: "FLOAT64",
exp.DataType.Type.DOUBLE: "FLOAT64",
exp.DataType.Type.BOOLEAN: "BOOL",
exp.DataType.Type.TEXT: "STRING",
}
print(Dialect["custom"])
<class '__main__.Custom'>
SQL Execution
SQLGlot is able to interpret SQL queries, where the tables are represented as Python dictionaries. The engine is not supposed to be fast, but it can be useful for unit testing and running SQL natively across Python objects. Additionally, the foundation can be easily integrated with fast compute kernels, such as Arrow and Pandas.
The example below showcases the execution of a query that involves aggregations and joins:
from sqlglot.executor import execute
tables = {
"sushi": [
{"id": 1, "price": 1.0},
{"id": 2, "price": 2.0},
{"id": 3, "price": 3.0},
],
"order_items": [
{"sushi_id": 1, "order_id": 1},
{"sushi_id": 1, "order_id": 1},
{"sushi_id": 2, "order_id": 1},
{"sushi_id": 3, "order_id": 2},
],
"orders": [
{"id": 1, "user_id": 1},
{"id": 2, "user_id": 2},
],
}
execute(
"""
SELECT
o.user_id,
SUM(s.price) AS price
FROM orders o
JOIN order_items i
ON o.id = i.order_id
JOIN sushi s
ON i.sushi_id = s.id
GROUP BY o.user_id
""",
tables=tables
)
user_id price
1 4.0
2 3.0
See also: Writing a Python SQL engine from scratch.
Used By
Documentation
SQLGlot uses pdoc to serve its API documentation.
A hosted version is on the SQLGlot website, or you can build locally with:
make docs-serve
Run Tests and Lint
make style # Only linter checks
make unit # Only unit tests (pure Python)
make test # Unit and integration tests (pure Python)
make unitc # Only unit tests (mypyc compiled)
make testc # Unit and integration tests (mypyc compiled)
make check # Full test suite & linter checks
make clean # Remove compiled C artifacts (.so files, build dirs)
Deployment
To deploy a new SQLGlot version, follow these steps:
- Run
git pullto make sure the local git repo is at the head of the main branch - Do a
git tagoperation to bump the SQLGlot version, e.g.git tag v28.5.0 - Run
git push && git push --tagsto deploy the new version
Benchmarks
Benchmarks run on Python 3.14.3 in seconds.
sqlglot, sqltree, sqlparse, and sqlfluff are python based whereas sqloxide and polyglot-sql are rust bindings.
| Query | sqlglot | sqlglot[c] | sqltree | sqlparse | sqlfluff | sqloxide | polyglot-sql |
|---|---|---|---|---|---|---|---|
| tpch | 0.002709 (1.00) | 0.000740 (0.27) | 0.002172 (0.80) | 0.014152 (5.22) | 0.241027 (88.97) | 0.000655 (0.24) | 0.000698 (0.26) |
| short | 0.000226 (1.00) | 0.000075 (0.33) | 0.000184 (0.81) | 0.000938 (4.15) | 0.031542 (139.47) | 0.000041 (0.18) | 0.000174 (0.77) |
| deep_arithmetic | 0.007760 (1.00) | 0.002015 (0.26) | 0.005927 (0.76) | N/A | 1.359824 (175.22) | 0.003117 (0.40) | 0.002964 (0.38) |
| large_in | 0.407987 (1.00) | 0.101644 (0.25) | 0.467943 (1.15) | N/A | N/A | 0.147765 (0.36) | 0.105854 (0.26) |
| values | 0.466734 (1.00) | 0.113762 (0.24) | 0.522797 (1.12) | N/A | N/A | 0.117628 (0.25) | 0.117169 (0.25) |
| many_joins | 0.011943 (1.00) | 0.002701 (0.23) | 0.009887 (0.83) | 0.059303 (4.97) | 1.246253 (104.35) | 0.002918 (0.24) | 0.002964 (0.25) |
| many_unions | 0.041321 (1.00) | 0.008291 (0.20) | 0.038249 (0.93) | N/A | 1.826401 (44.20) | 0.012395 (0.30) | 0.013087 (0.32) |
| nested_subqueries | 0.001200 (1.00) | 0.000235 (0.20) | N/A | 0.003860 (3.22) | 0.089490 (74.56) | 0.000215 (0.18) | 0.000262 (0.22) |
| many_columns | 0.011821 (1.00) | 0.002825 (0.24) | 0.012722 (1.08) | 0.238510 (20.18) | 1.050386 (88.86) | 0.002515 (0.21) | 0.003765 (0.32) |
| large_case | 0.035822 (1.00) | 0.008593 (0.24) | 0.033578 (0.94) | N/A | 4.200220 (117.25) | 0.009870 (0.28) | 0.009442 (0.26) |
| complex_where | 0.032710 (1.00) | 0.006602 (0.20) | N/A | 0.136203 (4.16) | 2.492927 (76.21) | 0.006002 (0.18) | 0.007787 (0.24) |
| many_ctes | 0.017610 (1.00) | 0.003630 (0.21) | 0.012377 (0.70) | 0.123620 (7.02) | 0.657611 (37.34) | 0.004197 (0.24) | 0.003273 (0.19) |
| many_windows | 0.020790 (1.00) | 0.005751 (0.28) | N/A | 0.203144 (9.77) | 1.421216 (68.36) | 0.003941 (0.19) | 0.004570 (0.22) |
| nested_functions | 0.000703 (1.00) | 0.000189 (0.27) | 0.000754 (1.07) | 0.005082 (7.23) | 0.091007 (129.51) | 0.000168 (0.24) | 0.000225 (0.32) |
| large_strings | 0.005073 (1.00) | 0.001480 (0.29) | 0.014533 (2.86) | 0.049392 (9.74) | 0.320672 (63.22) | 0.001616 (0.32) | 0.002151 (0.42) |
| many_numbers | 0.103898 (1.00) | 0.024483 (0.24) | 0.120119 (1.16) | N/A | N/A | 0.031667 (0.30) | 0.026880 (0.26) |
make bench # Run parsing benchmark
make bench-optimize # Run optimization benchmark
Optional Dependencies
SQLGlot uses dateutil to simplify literal timedelta expressions. The optimizer will not simplify expressions like the following if the module cannot be found:
x + interval '1' month
Supported Dialects
| Dialect | Support Level |
|---|---|
| Athena | Official |
| BigQuery | Official |
| ClickHouse | Official |
| Databricks | Official |
| Doris | Community |
| Dremio | Community |
| Drill | Community |
| Druid | Community |
| DuckDB | Official |
| Exasol | Community |
| Fabric | Community |
| Hive | Official |
| Materialize | Community |
| MySQL | Official |
| Oracle | Official |
| Postgres | Official |
| Presto | Official |
| PRQL | Community |
| Redshift | Official |
| RisingWave | Community |
| SingleStore | Community |
| Snowflake | Official |
| Solr | Community |
| Spark | Official |
| SQLite | Official |
| StarRocks | Official |
| Tableau | Official |
| Teradata | Community |
| Trino | Official |
| TSQL | Official |
| YDB | Plugin |
| MaxCompute | Plugin |
Official Dialects are maintained by the core SQLGlot team with higher priority for bug fixes and feature additions.
Community Dialects are developed and maintained primarily through community contributions. These are fully functional but may receive lower priority for issue resolution compared to officially supported dialects. We welcome and encourage community contributions to improve these dialects.
Plugin Dialects (supported since v28.6.0) are third-party dialects developed and maintained in external repositories by independent contributors. These dialects are not part of the SQLGlot codebase and are distributed as separate packages. The SQLGlot team does not provide support or maintenance for plugin dialects — please direct any issues or feature requests to their respective repositories. See Creating a Dialect Plugin below for information on how to build your own.
Creating a Dialect Plugin
If your database isn't supported, you can create a plugin that registers a custom dialect via entry points. Create a package with your dialect class and register it in setup.py:
from setuptools import setup
setup(
name="mydb-sqlglot-dialect",
entry_points={
"sqlglot.dialects": [
"mydb = my_package.dialect:MyDB",
],
},
)
The dialect will be automatically discovered and can be used like any built-in dialect:
from sqlglot import transpile
transpile("SELECT * FROM t", read="mydb", write="postgres")
See the Custom Dialects section for implementation details.
1# ruff: noqa: F401, E402 2""" 3.. include:: ../README.md 4 5---- 6""" 7 8from __future__ import annotations 9 10# bootstrap mypyc runtime: compiled .so modules do a top-level `import HASH__mypyc`, 11# but the runtime .so lives inside sqlglot/. Pre-load it into sys.modules. 12# this is only needed for editable builds 13from collections.abc import Collection 14import sys 15from pathlib import Path 16from builtins import type as Type 17 18for path in Path(__file__).parent.glob("*__mypyc*.so"): 19 name = path.stem.split(".")[0] 20 if name not in sys.modules: 21 import importlib.util 22 23 spec = importlib.util.spec_from_file_location(name, path) 24 if spec and spec.loader: 25 mod = importlib.util.module_from_spec(spec) 26 sys.modules[name] = mod 27 spec.loader.exec_module(mod) 28 29import logging 30import typing as t 31 32from sqlglot import expressions as exp 33from sqlglot.dialects.dialect import Dialect as Dialect, Dialects as Dialects 34from sqlglot.diff import diff as diff 35from sqlglot.errors import ( 36 ErrorLevel as ErrorLevel, 37 ParseError as ParseError, 38 TokenError as TokenError, 39 UnsupportedError as UnsupportedError, 40) 41from sqlglot.expressions import ( 42 Expr as Expr, 43 alias_ as alias, 44 and_ as and_, 45 case as case, 46 cast as cast, 47 column as column, 48 condition as condition, 49 delete as delete, 50 except_ as except_, 51 find_tables as find_tables, 52 from_ as from_, 53 func as func, 54 insert as insert, 55 intersect as intersect, 56 maybe_parse as maybe_parse, 57 merge as merge, 58 not_ as not_, 59 or_ as or_, 60 select as select, 61 subquery as subquery, 62 table_ as table, 63 to_column as to_column, 64 to_identifier as to_identifier, 65 to_table as to_table, 66 union as union, 67) 68from sqlglot.generator import Generator as Generator 69from sqlglot.parser import Parser as Parser 70from sqlglot.schema import MappingSchema as MappingSchema, Schema as Schema 71from sqlglot.tokens import Token as Token, Tokenizer as Tokenizer, TokenType as TokenType 72 73if t.TYPE_CHECKING: 74 from sqlglot._typing import E, GeneratorArgs, ParserNoDialectArgs 75 from typing_extensions import Unpack 76 from sqlglot.dialects.dialect import DialectType as DialectType 77 78logger = logging.getLogger("sqlglot") 79 80 81try: 82 from sqlglot._version import __version__, __version_tuple__ # type: ignore[import-not-found] 83except ImportError: 84 logger.error( 85 "Unable to set __version__, run `pip install -e .` or `python setup.py develop` first." 86 ) 87 88 89pretty = False 90"""Whether to format generated SQL by default.""" 91 92 93def tokenize(sql: str, read: DialectType = None, dialect: DialectType = None) -> list[Token]: 94 """ 95 Tokenizes the given SQL string. 96 97 Args: 98 sql: the SQL code string to tokenize. 99 read: the SQL dialect to apply during tokenizing (eg. "spark", "hive", "presto", "mysql"). 100 dialect: the SQL dialect (alias for read). 101 102 Returns: 103 The resulting list of tokens. 104 """ 105 return Dialect.get_or_raise(read or dialect).tokenize(sql) 106 107 108def parse( 109 sql: str, 110 read: DialectType = None, 111 dialect: DialectType = None, 112 **opts: Unpack[ParserNoDialectArgs], 113) -> list[Expr | None]: 114 """ 115 Parses the given SQL string into a collection of syntax trees, one per parsed SQL statement. 116 117 Args: 118 sql: the SQL code string to parse. 119 read: the SQL dialect to apply during parsing (eg. "spark", "hive", "presto", "mysql"). 120 dialect: the SQL dialect (alias for read). 121 **opts: other `sqlglot.parser.Parser` options. 122 123 Returns: 124 The resulting syntax tree collection. 125 """ 126 return Dialect.get_or_raise(read or dialect).parse(sql, **opts) 127 128 129@t.overload 130def parse_one( 131 sql: str, 132 *, 133 read: DialectType = ..., 134 dialect: DialectType = ..., 135 into: Type[E], 136 **opts: Unpack[ParserNoDialectArgs], 137) -> E: ... 138 139 140@t.overload 141def parse_one( 142 sql: str, 143 read: DialectType = ..., 144 dialect: DialectType = ..., 145 into: exp.IntoType | None = ..., 146 **opts: Unpack[ParserNoDialectArgs], 147) -> Expr: ... 148 149 150def parse_one( 151 sql: str, 152 read: DialectType = None, 153 dialect: DialectType = None, 154 into: exp.IntoType | None = None, 155 **opts: Unpack[ParserNoDialectArgs], 156) -> Expr: 157 """ 158 Parses the given SQL string and returns a syntax tree. 159 160 Args: 161 sql: the SQL code string to parse. 162 read: the SQL dialect to apply during parsing (eg. "spark", "hive", "presto", "mysql"). 163 dialect: the SQL dialect (alias for read) 164 into: the SQLGlot Expr to parse into. 165 **opts: other `sqlglot.parser.Parser` options. 166 167 Returns: 168 A single syntax tree if one statement is parsed, otherwise a Block expression containing all parsed syntax trees. 169 """ 170 171 dialect = Dialect.get_or_raise(read or dialect) 172 173 if into: 174 result = dialect.parse_into(into, sql, **opts) 175 else: 176 result = dialect.parse(sql, **opts) 177 178 if not result or result[0] is None: 179 raise ParseError(f"No expression was parsed from '{sql}'") 180 181 return exp.Block(expressions=result) if len(result) > 1 else result[0] 182 183 184def transpile( 185 sql: str, 186 read: DialectType = None, 187 write: DialectType = None, 188 identity: bool = True, 189 error_level: ErrorLevel | None = None, 190 **opts: Unpack[GeneratorArgs], 191) -> list[str]: 192 """ 193 Parses the given SQL string in accordance with the source dialect and returns a list of SQL strings transformed 194 to conform to the target dialect. Each string in the returned list represents a single transformed SQL statement. 195 196 Args: 197 sql: the SQL code string to transpile. 198 read: the source dialect used to parse the input string (eg. "spark", "hive", "presto", "mysql"). 199 write: the target dialect into which the input should be transformed (eg. "spark", "hive", "presto", "mysql"). 200 identity: if set to `True` and if the target dialect is not specified the source dialect will be used as both: 201 the source and the target dialect. 202 error_level: the desired error level of the parser. 203 **opts: other `sqlglot.generator.Generator` options. 204 205 Returns: 206 The list of transpiled SQL statements. 207 """ 208 write = (read if write is None else write) if identity else write 209 write = Dialect.get_or_raise(write) 210 return [ 211 write.generate(expression, copy=False, **opts) if expression else "" 212 for expression in parse(sql, read, error_level=error_level) 213 ]
Whether to format generated SQL by default.
94def tokenize(sql: str, read: DialectType = None, dialect: DialectType = None) -> list[Token]: 95 """ 96 Tokenizes the given SQL string. 97 98 Args: 99 sql: the SQL code string to tokenize. 100 read: the SQL dialect to apply during tokenizing (eg. "spark", "hive", "presto", "mysql"). 101 dialect: the SQL dialect (alias for read). 102 103 Returns: 104 The resulting list of tokens. 105 """ 106 return Dialect.get_or_raise(read or dialect).tokenize(sql)
Tokenizes the given SQL string.
Arguments:
- sql: the SQL code string to tokenize.
- read: the SQL dialect to apply during tokenizing (eg. "spark", "hive", "presto", "mysql").
- dialect: the SQL dialect (alias for read).
Returns:
The resulting list of tokens.
109def parse( 110 sql: str, 111 read: DialectType = None, 112 dialect: DialectType = None, 113 **opts: Unpack[ParserNoDialectArgs], 114) -> list[Expr | None]: 115 """ 116 Parses the given SQL string into a collection of syntax trees, one per parsed SQL statement. 117 118 Args: 119 sql: the SQL code string to parse. 120 read: the SQL dialect to apply during parsing (eg. "spark", "hive", "presto", "mysql"). 121 dialect: the SQL dialect (alias for read). 122 **opts: other `sqlglot.parser.Parser` options. 123 124 Returns: 125 The resulting syntax tree collection. 126 """ 127 return Dialect.get_or_raise(read or dialect).parse(sql, **opts)
Parses the given SQL string into a collection of syntax trees, one per parsed SQL statement.
Arguments:
- sql: the SQL code string to parse.
- read: the SQL dialect to apply during parsing (eg. "spark", "hive", "presto", "mysql").
- dialect: the SQL dialect (alias for read).
- **opts: other
sqlglot.parser.Parseroptions.
Returns:
The resulting syntax tree collection.
151def parse_one( 152 sql: str, 153 read: DialectType = None, 154 dialect: DialectType = None, 155 into: exp.IntoType | None = None, 156 **opts: Unpack[ParserNoDialectArgs], 157) -> Expr: 158 """ 159 Parses the given SQL string and returns a syntax tree. 160 161 Args: 162 sql: the SQL code string to parse. 163 read: the SQL dialect to apply during parsing (eg. "spark", "hive", "presto", "mysql"). 164 dialect: the SQL dialect (alias for read) 165 into: the SQLGlot Expr to parse into. 166 **opts: other `sqlglot.parser.Parser` options. 167 168 Returns: 169 A single syntax tree if one statement is parsed, otherwise a Block expression containing all parsed syntax trees. 170 """ 171 172 dialect = Dialect.get_or_raise(read or dialect) 173 174 if into: 175 result = dialect.parse_into(into, sql, **opts) 176 else: 177 result = dialect.parse(sql, **opts) 178 179 if not result or result[0] is None: 180 raise ParseError(f"No expression was parsed from '{sql}'") 181 182 return exp.Block(expressions=result) if len(result) > 1 else result[0]
Parses the given SQL string and returns a syntax tree.
Arguments:
- sql: the SQL code string to parse.
- read: the SQL dialect to apply during parsing (eg. "spark", "hive", "presto", "mysql").
- dialect: the SQL dialect (alias for read)
- into: the SQLGlot Expr to parse into.
- **opts: other
sqlglot.parser.Parseroptions.
Returns:
A single syntax tree if one statement is parsed, otherwise a Block expression containing all parsed syntax trees.
185def transpile( 186 sql: str, 187 read: DialectType = None, 188 write: DialectType = None, 189 identity: bool = True, 190 error_level: ErrorLevel | None = None, 191 **opts: Unpack[GeneratorArgs], 192) -> list[str]: 193 """ 194 Parses the given SQL string in accordance with the source dialect and returns a list of SQL strings transformed 195 to conform to the target dialect. Each string in the returned list represents a single transformed SQL statement. 196 197 Args: 198 sql: the SQL code string to transpile. 199 read: the source dialect used to parse the input string (eg. "spark", "hive", "presto", "mysql"). 200 write: the target dialect into which the input should be transformed (eg. "spark", "hive", "presto", "mysql"). 201 identity: if set to `True` and if the target dialect is not specified the source dialect will be used as both: 202 the source and the target dialect. 203 error_level: the desired error level of the parser. 204 **opts: other `sqlglot.generator.Generator` options. 205 206 Returns: 207 The list of transpiled SQL statements. 208 """ 209 write = (read if write is None else write) if identity else write 210 write = Dialect.get_or_raise(write) 211 return [ 212 write.generate(expression, copy=False, **opts) if expression else "" 213 for expression in parse(sql, read, error_level=error_level) 214 ]
Parses the given SQL string in accordance with the source dialect and returns a list of SQL strings transformed to conform to the target dialect. Each string in the returned list represents a single transformed SQL statement.
Arguments:
- sql: the SQL code string to transpile.
- read: the source dialect used to parse the input string (eg. "spark", "hive", "presto", "mysql").
- write: the target dialect into which the input should be transformed (eg. "spark", "hive", "presto", "mysql").
- identity: if set to
Trueand if the target dialect is not specified the source dialect will be used as both: the source and the target dialect. - error_level: the desired error level of the parser.
- **opts: other
sqlglot.generator.Generatoroptions.
Returns:
The list of transpiled SQL statements.