Edit on GitHub

sqlglot.dialects.bigquery

  1from __future__ import annotations
  2
  3import logging
  4import re
  5import typing as t
  6
  7from sqlglot import exp, generator, parser, tokens, transforms
  8from sqlglot._typing import E
  9from sqlglot.dialects.dialect import (
 10    Dialect,
 11    NormalizationStrategy,
 12    arg_max_or_min_no_count,
 13    binary_from_function,
 14    date_add_interval_sql,
 15    datestrtodate_sql,
 16    format_time_lambda,
 17    if_sql,
 18    inline_array_sql,
 19    json_keyvalue_comma_sql,
 20    max_or_greatest,
 21    min_or_least,
 22    no_ilike_sql,
 23    parse_date_delta_with_interval,
 24    path_to_jsonpath,
 25    regexp_replace_sql,
 26    rename_func,
 27    timestrtotime_sql,
 28    ts_or_ds_add_cast,
 29)
 30from sqlglot.helper import seq_get, split_num_words
 31from sqlglot.tokens import TokenType
 32
 33logger = logging.getLogger("sqlglot")
 34
 35
 36def _derived_table_values_to_unnest(self: BigQuery.Generator, expression: exp.Values) -> str:
 37    if not expression.find_ancestor(exp.From, exp.Join):
 38        return self.values_sql(expression)
 39
 40    alias = expression.args.get("alias")
 41
 42    structs = [
 43        exp.Struct(
 44            expressions=[
 45                exp.alias_(value, column_name)
 46                for value, column_name in zip(
 47                    t.expressions,
 48                    alias.columns
 49                    if alias and alias.columns
 50                    else (f"_c{i}" for i in range(len(t.expressions))),
 51                )
 52            ]
 53        )
 54        for t in expression.find_all(exp.Tuple)
 55    ]
 56
 57    return self.unnest_sql(exp.Unnest(expressions=[exp.Array(expressions=structs)]))
 58
 59
 60def _returnsproperty_sql(self: BigQuery.Generator, expression: exp.ReturnsProperty) -> str:
 61    this = expression.this
 62    if isinstance(this, exp.Schema):
 63        this = f"{this.this} <{self.expressions(this)}>"
 64    else:
 65        this = self.sql(this)
 66    return f"RETURNS {this}"
 67
 68
 69def _create_sql(self: BigQuery.Generator, expression: exp.Create) -> str:
 70    kind = expression.args["kind"]
 71    returns = expression.find(exp.ReturnsProperty)
 72
 73    if kind.upper() == "FUNCTION" and returns and returns.args.get("is_table"):
 74        expression.set("kind", "TABLE FUNCTION")
 75
 76        if isinstance(expression.expression, (exp.Subquery, exp.Literal)):
 77            expression.set("expression", expression.expression.this)
 78
 79        return self.create_sql(expression)
 80
 81    return self.create_sql(expression)
 82
 83
 84def _unqualify_unnest(expression: exp.Expression) -> exp.Expression:
 85    """Remove references to unnest table aliases since bigquery doesn't allow them.
 86
 87    These are added by the optimizer's qualify_column step.
 88    """
 89    from sqlglot.optimizer.scope import find_all_in_scope
 90
 91    if isinstance(expression, exp.Select):
 92        unnest_aliases = {
 93            unnest.alias
 94            for unnest in find_all_in_scope(expression, exp.Unnest)
 95            if isinstance(unnest.parent, (exp.From, exp.Join))
 96        }
 97        if unnest_aliases:
 98            for column in expression.find_all(exp.Column):
 99                if column.table in unnest_aliases:
100                    column.set("table", None)
101                elif column.db in unnest_aliases:
102                    column.set("db", None)
103
104    return expression
105
106
107# https://issuetracker.google.com/issues/162294746
108# workaround for bigquery bug when grouping by an expression and then ordering
109# WITH x AS (SELECT 1 y)
110# SELECT y + 1 z
111# FROM x
112# GROUP BY x + 1
113# ORDER by z
114def _alias_ordered_group(expression: exp.Expression) -> exp.Expression:
115    if isinstance(expression, exp.Select):
116        group = expression.args.get("group")
117        order = expression.args.get("order")
118
119        if group and order:
120            aliases = {
121                select.this: select.args["alias"]
122                for select in expression.selects
123                if isinstance(select, exp.Alias)
124            }
125
126            for e in group.expressions:
127                alias = aliases.get(e)
128
129                if alias:
130                    e.replace(exp.column(alias))
131
132    return expression
133
134
135def _pushdown_cte_column_names(expression: exp.Expression) -> exp.Expression:
136    """BigQuery doesn't allow column names when defining a CTE, so we try to push them down."""
137    if isinstance(expression, exp.CTE) and expression.alias_column_names:
138        cte_query = expression.this
139
140        if cte_query.is_star:
141            logger.warning(
142                "Can't push down CTE column names for star queries. Run the query through"
143                " the optimizer or use 'qualify' to expand the star projections first."
144            )
145            return expression
146
147        column_names = expression.alias_column_names
148        expression.args["alias"].set("columns", None)
149
150        for name, select in zip(column_names, cte_query.selects):
151            to_replace = select
152
153            if isinstance(select, exp.Alias):
154                select = select.this
155
156            # Inner aliases are shadowed by the CTE column names
157            to_replace.replace(exp.alias_(select, name))
158
159    return expression
160
161
162def _parse_timestamp(args: t.List) -> exp.StrToTime:
163    this = format_time_lambda(exp.StrToTime, "bigquery")([seq_get(args, 1), seq_get(args, 0)])
164    this.set("zone", seq_get(args, 2))
165    return this
166
167
168def _parse_date(args: t.List) -> exp.Date | exp.DateFromParts:
169    expr_type = exp.DateFromParts if len(args) == 3 else exp.Date
170    return expr_type.from_arg_list(args)
171
172
173def _parse_to_hex(args: t.List) -> exp.Hex | exp.MD5:
174    # TO_HEX(MD5(..)) is common in BigQuery, so it's parsed into MD5 to simplify its transpilation
175    arg = seq_get(args, 0)
176    return exp.MD5(this=arg.this) if isinstance(arg, exp.MD5Digest) else exp.Hex(this=arg)
177
178
179def _array_contains_sql(self: BigQuery.Generator, expression: exp.ArrayContains) -> str:
180    return self.sql(
181        exp.Exists(
182            this=exp.select("1")
183            .from_(exp.Unnest(expressions=[expression.left]).as_("_unnest", table=["_col"]))
184            .where(exp.column("_col").eq(expression.right))
185        )
186    )
187
188
189def _ts_or_ds_add_sql(self: BigQuery.Generator, expression: exp.TsOrDsAdd) -> str:
190    return date_add_interval_sql("DATE", "ADD")(self, ts_or_ds_add_cast(expression))
191
192
193def _ts_or_ds_diff_sql(self: BigQuery.Generator, expression: exp.TsOrDsDiff) -> str:
194    expression.this.replace(exp.cast(expression.this, "TIMESTAMP", copy=True))
195    expression.expression.replace(exp.cast(expression.expression, "TIMESTAMP", copy=True))
196    unit = expression.args.get("unit") or "DAY"
197    return self.func("DATE_DIFF", expression.this, expression.expression, unit)
198
199
200def _unix_to_time_sql(self: BigQuery.Generator, expression: exp.UnixToTime) -> str:
201    scale = expression.args.get("scale")
202    timestamp = self.sql(expression, "this")
203    if scale in (None, exp.UnixToTime.SECONDS):
204        return f"TIMESTAMP_SECONDS({timestamp})"
205    if scale == exp.UnixToTime.MILLIS:
206        return f"TIMESTAMP_MILLIS({timestamp})"
207    if scale == exp.UnixToTime.MICROS:
208        return f"TIMESTAMP_MICROS({timestamp})"
209    if scale == exp.UnixToTime.NANOS:
210        # We need to cast to INT64 because that's what BQ expects
211        return f"TIMESTAMP_MICROS(CAST({timestamp} / 1000 AS INT64))"
212
213    self.unsupported(f"Unsupported scale for timestamp: {scale}.")
214    return ""
215
216
217def _parse_time(args: t.List) -> exp.Func:
218    if len(args) == 1:
219        return exp.TsOrDsToTime(this=args[0])
220    if len(args) == 3:
221        return exp.TimeFromParts.from_arg_list(args)
222
223    return exp.Anonymous(this="TIME", expressions=args)
224
225
226class BigQuery(Dialect):
227    WEEK_OFFSET = -1
228    UNNEST_COLUMN_ONLY = True
229    SUPPORTS_USER_DEFINED_TYPES = False
230    SUPPORTS_SEMI_ANTI_JOIN = False
231    LOG_BASE_FIRST = False
232
233    # https://cloud.google.com/bigquery/docs/reference/standard-sql/lexical#case_sensitivity
234    NORMALIZATION_STRATEGY = NormalizationStrategy.CASE_INSENSITIVE
235
236    # bigquery udfs are case sensitive
237    NORMALIZE_FUNCTIONS = False
238
239    TIME_MAPPING = {
240        "%D": "%m/%d/%y",
241    }
242
243    ESCAPE_SEQUENCES = {
244        "\\a": "\a",
245        "\\b": "\b",
246        "\\f": "\f",
247        "\\n": "\n",
248        "\\r": "\r",
249        "\\t": "\t",
250        "\\v": "\v",
251    }
252
253    FORMAT_MAPPING = {
254        "DD": "%d",
255        "MM": "%m",
256        "MON": "%b",
257        "MONTH": "%B",
258        "YYYY": "%Y",
259        "YY": "%y",
260        "HH": "%I",
261        "HH12": "%I",
262        "HH24": "%H",
263        "MI": "%M",
264        "SS": "%S",
265        "SSSSS": "%f",
266        "TZH": "%z",
267    }
268
269    # The _PARTITIONTIME and _PARTITIONDATE pseudo-columns are not returned by a SELECT * statement
270    # https://cloud.google.com/bigquery/docs/querying-partitioned-tables#query_an_ingestion-time_partitioned_table
271    PSEUDOCOLUMNS = {"_PARTITIONTIME", "_PARTITIONDATE"}
272
273    def normalize_identifier(self, expression: E) -> E:
274        if isinstance(expression, exp.Identifier):
275            parent = expression.parent
276            while isinstance(parent, exp.Dot):
277                parent = parent.parent
278
279            # In BigQuery, CTEs aren't case-sensitive, but table names are (by default, at least).
280            # The following check is essentially a heuristic to detect tables based on whether or
281            # not they're qualified. It also avoids normalizing UDFs, because they're case-sensitive.
282            if (
283                not isinstance(parent, exp.UserDefinedFunction)
284                and not (isinstance(parent, exp.Table) and parent.db)
285                and not expression.meta.get("is_table")
286            ):
287                expression.set("this", expression.this.lower())
288
289        return expression
290
291    class Tokenizer(tokens.Tokenizer):
292        QUOTES = ["'", '"', '"""', "'''"]
293        COMMENTS = ["--", "#", ("/*", "*/")]
294        IDENTIFIERS = ["`"]
295        STRING_ESCAPES = ["\\"]
296
297        HEX_STRINGS = [("0x", ""), ("0X", "")]
298
299        BYTE_STRINGS = [
300            (prefix + q, q) for q in t.cast(t.List[str], QUOTES) for prefix in ("b", "B")
301        ]
302
303        RAW_STRINGS = [
304            (prefix + q, q) for q in t.cast(t.List[str], QUOTES) for prefix in ("r", "R")
305        ]
306
307        KEYWORDS = {
308            **tokens.Tokenizer.KEYWORDS,
309            "ANY TYPE": TokenType.VARIANT,
310            "BEGIN": TokenType.COMMAND,
311            "BEGIN TRANSACTION": TokenType.BEGIN,
312            "BYTES": TokenType.BINARY,
313            "CURRENT_DATETIME": TokenType.CURRENT_DATETIME,
314            "DECLARE": TokenType.COMMAND,
315            "FLOAT64": TokenType.DOUBLE,
316            "FOR SYSTEM_TIME": TokenType.TIMESTAMP_SNAPSHOT,
317            "MODEL": TokenType.MODEL,
318            "NOT DETERMINISTIC": TokenType.VOLATILE,
319            "RECORD": TokenType.STRUCT,
320            "TIMESTAMP": TokenType.TIMESTAMPTZ,
321        }
322        KEYWORDS.pop("DIV")
323
324    class Parser(parser.Parser):
325        PREFIXED_PIVOT_COLUMNS = True
326
327        LOG_DEFAULTS_TO_LN = True
328
329        FUNCTIONS = {
330            **parser.Parser.FUNCTIONS,
331            "DATE": _parse_date,
332            "DATE_ADD": parse_date_delta_with_interval(exp.DateAdd),
333            "DATE_SUB": parse_date_delta_with_interval(exp.DateSub),
334            "DATE_TRUNC": lambda args: exp.DateTrunc(
335                unit=exp.Literal.string(str(seq_get(args, 1))),
336                this=seq_get(args, 0),
337            ),
338            "DATETIME_ADD": parse_date_delta_with_interval(exp.DatetimeAdd),
339            "DATETIME_SUB": parse_date_delta_with_interval(exp.DatetimeSub),
340            "DIV": binary_from_function(exp.IntDiv),
341            "FORMAT_DATE": lambda args: exp.TimeToStr(
342                this=exp.TsOrDsToDate(this=seq_get(args, 1)), format=seq_get(args, 0)
343            ),
344            "GENERATE_ARRAY": exp.GenerateSeries.from_arg_list,
345            "JSON_EXTRACT_SCALAR": lambda args: exp.JSONExtractScalar(
346                this=seq_get(args, 0), expression=seq_get(args, 1) or exp.Literal.string("$")
347            ),
348            "MD5": exp.MD5Digest.from_arg_list,
349            "TO_HEX": _parse_to_hex,
350            "PARSE_DATE": lambda args: format_time_lambda(exp.StrToDate, "bigquery")(
351                [seq_get(args, 1), seq_get(args, 0)]
352            ),
353            "PARSE_TIMESTAMP": _parse_timestamp,
354            "REGEXP_CONTAINS": exp.RegexpLike.from_arg_list,
355            "REGEXP_EXTRACT": lambda args: exp.RegexpExtract(
356                this=seq_get(args, 0),
357                expression=seq_get(args, 1),
358                position=seq_get(args, 2),
359                occurrence=seq_get(args, 3),
360                group=exp.Literal.number(1) if re.compile(args[1].name).groups == 1 else None,
361            ),
362            "SHA256": lambda args: exp.SHA2(this=seq_get(args, 0), length=exp.Literal.number(256)),
363            "SHA512": lambda args: exp.SHA2(this=seq_get(args, 0), length=exp.Literal.number(512)),
364            "SPLIT": lambda args: exp.Split(
365                # https://cloud.google.com/bigquery/docs/reference/standard-sql/string_functions#split
366                this=seq_get(args, 0),
367                expression=seq_get(args, 1) or exp.Literal.string(","),
368            ),
369            "TIME": _parse_time,
370            "TIME_ADD": parse_date_delta_with_interval(exp.TimeAdd),
371            "TIME_SUB": parse_date_delta_with_interval(exp.TimeSub),
372            "TIMESTAMP_ADD": parse_date_delta_with_interval(exp.TimestampAdd),
373            "TIMESTAMP_SUB": parse_date_delta_with_interval(exp.TimestampSub),
374            "TIMESTAMP_MICROS": lambda args: exp.UnixToTime(
375                this=seq_get(args, 0), scale=exp.UnixToTime.MICROS
376            ),
377            "TIMESTAMP_MILLIS": lambda args: exp.UnixToTime(
378                this=seq_get(args, 0), scale=exp.UnixToTime.MILLIS
379            ),
380            "TIMESTAMP_SECONDS": lambda args: exp.UnixToTime(
381                this=seq_get(args, 0), scale=exp.UnixToTime.SECONDS
382            ),
383            "TO_JSON_STRING": exp.JSONFormat.from_arg_list,
384        }
385
386        FUNCTION_PARSERS = {
387            **parser.Parser.FUNCTION_PARSERS,
388            "ARRAY": lambda self: self.expression(exp.Array, expressions=[self._parse_statement()]),
389        }
390        FUNCTION_PARSERS.pop("TRIM")
391
392        NO_PAREN_FUNCTIONS = {
393            **parser.Parser.NO_PAREN_FUNCTIONS,
394            TokenType.CURRENT_DATETIME: exp.CurrentDatetime,
395        }
396
397        NESTED_TYPE_TOKENS = {
398            *parser.Parser.NESTED_TYPE_TOKENS,
399            TokenType.TABLE,
400        }
401
402        ID_VAR_TOKENS = {
403            *parser.Parser.ID_VAR_TOKENS,
404            TokenType.VALUES,
405        }
406
407        PROPERTY_PARSERS = {
408            **parser.Parser.PROPERTY_PARSERS,
409            "NOT DETERMINISTIC": lambda self: self.expression(
410                exp.StabilityProperty, this=exp.Literal.string("VOLATILE")
411            ),
412            "OPTIONS": lambda self: self._parse_with_property(),
413        }
414
415        CONSTRAINT_PARSERS = {
416            **parser.Parser.CONSTRAINT_PARSERS,
417            "OPTIONS": lambda self: exp.Properties(expressions=self._parse_with_property()),
418        }
419
420        RANGE_PARSERS = parser.Parser.RANGE_PARSERS.copy()
421        RANGE_PARSERS.pop(TokenType.OVERLAPS, None)
422
423        NULL_TOKENS = {TokenType.NULL, TokenType.UNKNOWN}
424
425        STATEMENT_PARSERS = {
426            **parser.Parser.STATEMENT_PARSERS,
427            TokenType.END: lambda self: self._parse_as_command(self._prev),
428            TokenType.FOR: lambda self: self._parse_for_in(),
429        }
430
431        BRACKET_OFFSETS = {
432            "OFFSET": (0, False),
433            "ORDINAL": (1, False),
434            "SAFE_OFFSET": (0, True),
435            "SAFE_ORDINAL": (1, True),
436        }
437
438        def _parse_for_in(self) -> exp.ForIn:
439            this = self._parse_range()
440            self._match_text_seq("DO")
441            return self.expression(exp.ForIn, this=this, expression=self._parse_statement())
442
443        def _parse_table_part(self, schema: bool = False) -> t.Optional[exp.Expression]:
444            this = super()._parse_table_part(schema=schema) or self._parse_number()
445
446            # https://cloud.google.com/bigquery/docs/reference/standard-sql/lexical#table_names
447            if isinstance(this, exp.Identifier):
448                table_name = this.name
449                while self._match(TokenType.DASH, advance=False) and self._next:
450                    self._advance(2)
451                    table_name += f"-{self._prev.text}"
452
453                this = exp.Identifier(this=table_name, quoted=this.args.get("quoted"))
454            elif isinstance(this, exp.Literal):
455                table_name = this.name
456
457                if self._is_connected() and self._parse_var(any_token=True):
458                    table_name += self._prev.text
459
460                this = exp.Identifier(this=table_name, quoted=True)
461
462            return this
463
464        def _parse_table_parts(self, schema: bool = False) -> exp.Table:
465            table = super()._parse_table_parts(schema=schema)
466            if isinstance(table.this, exp.Identifier) and "." in table.name:
467                catalog, db, this, *rest = (
468                    t.cast(t.Optional[exp.Expression], exp.to_identifier(x))
469                    for x in split_num_words(table.name, ".", 3)
470                )
471
472                if rest and this:
473                    this = exp.Dot.build(t.cast(t.List[exp.Expression], [this, *rest]))
474
475                table = exp.Table(this=this, db=db, catalog=catalog)
476
477            return table
478
479        def _parse_json_object(self) -> exp.JSONObject:
480            json_object = super()._parse_json_object()
481            array_kv_pair = seq_get(json_object.expressions, 0)
482
483            # Converts BQ's "signature 2" of JSON_OBJECT into SQLGlot's canonical representation
484            # https://cloud.google.com/bigquery/docs/reference/standard-sql/json_functions#json_object_signature2
485            if (
486                array_kv_pair
487                and isinstance(array_kv_pair.this, exp.Array)
488                and isinstance(array_kv_pair.expression, exp.Array)
489            ):
490                keys = array_kv_pair.this.expressions
491                values = array_kv_pair.expression.expressions
492
493                json_object.set(
494                    "expressions",
495                    [exp.JSONKeyValue(this=k, expression=v) for k, v in zip(keys, values)],
496                )
497
498            return json_object
499
500        def _parse_bracket(self, this: t.Optional[exp.Expression]) -> t.Optional[exp.Expression]:
501            bracket = super()._parse_bracket(this)
502
503            if this is bracket:
504                return bracket
505
506            if isinstance(bracket, exp.Bracket):
507                for expression in bracket.expressions:
508                    name = expression.name.upper()
509
510                    if name not in self.BRACKET_OFFSETS:
511                        break
512
513                    offset, safe = self.BRACKET_OFFSETS[name]
514                    bracket.set("offset", offset)
515                    bracket.set("safe", safe)
516                    expression.replace(expression.expressions[0])
517
518            return bracket
519
520    class Generator(generator.Generator):
521        EXPLICIT_UNION = True
522        INTERVAL_ALLOWS_PLURAL_FORM = False
523        JOIN_HINTS = False
524        QUERY_HINTS = False
525        TABLE_HINTS = False
526        LIMIT_FETCH = "LIMIT"
527        RENAME_TABLE_WITH_DB = False
528        NVL2_SUPPORTED = False
529        UNNEST_WITH_ORDINALITY = False
530        COLLATE_IS_FUNC = True
531        LIMIT_ONLY_LITERALS = True
532        SUPPORTS_TABLE_ALIAS_COLUMNS = False
533        UNPIVOT_ALIASES_ARE_IDENTIFIERS = False
534
535        TRANSFORMS = {
536            **generator.Generator.TRANSFORMS,
537            exp.ApproxDistinct: rename_func("APPROX_COUNT_DISTINCT"),
538            exp.ArgMax: arg_max_or_min_no_count("MAX_BY"),
539            exp.ArgMin: arg_max_or_min_no_count("MIN_BY"),
540            exp.ArrayContains: _array_contains_sql,
541            exp.ArraySize: rename_func("ARRAY_LENGTH"),
542            exp.Cast: transforms.preprocess([transforms.remove_precision_parameterized_types]),
543            exp.CollateProperty: lambda self, e: f"DEFAULT COLLATE {self.sql(e, 'this')}"
544            if e.args.get("default")
545            else f"COLLATE {self.sql(e, 'this')}",
546            exp.CountIf: rename_func("COUNTIF"),
547            exp.Create: _create_sql,
548            exp.CTE: transforms.preprocess([_pushdown_cte_column_names]),
549            exp.DateAdd: date_add_interval_sql("DATE", "ADD"),
550            exp.DateDiff: lambda self, e: f"DATE_DIFF({self.sql(e, 'this')}, {self.sql(e, 'expression')}, {self.sql(e.args.get('unit', 'DAY'))})",
551            exp.DateFromParts: rename_func("DATE"),
552            exp.DateStrToDate: datestrtodate_sql,
553            exp.DateSub: date_add_interval_sql("DATE", "SUB"),
554            exp.DatetimeAdd: date_add_interval_sql("DATETIME", "ADD"),
555            exp.DatetimeSub: date_add_interval_sql("DATETIME", "SUB"),
556            exp.DateTrunc: lambda self, e: self.func("DATE_TRUNC", e.this, e.text("unit")),
557            exp.GenerateSeries: rename_func("GENERATE_ARRAY"),
558            exp.GetPath: path_to_jsonpath(),
559            exp.GroupConcat: rename_func("STRING_AGG"),
560            exp.Hex: rename_func("TO_HEX"),
561            exp.If: if_sql(false_value="NULL"),
562            exp.ILike: no_ilike_sql,
563            exp.IntDiv: rename_func("DIV"),
564            exp.JSONFormat: rename_func("TO_JSON_STRING"),
565            exp.JSONKeyValue: json_keyvalue_comma_sql,
566            exp.Max: max_or_greatest,
567            exp.MD5: lambda self, e: self.func("TO_HEX", self.func("MD5", e.this)),
568            exp.MD5Digest: rename_func("MD5"),
569            exp.Min: min_or_least,
570            exp.PartitionedByProperty: lambda self, e: f"PARTITION BY {self.sql(e, 'this')}",
571            exp.RegexpExtract: lambda self, e: self.func(
572                "REGEXP_EXTRACT",
573                e.this,
574                e.expression,
575                e.args.get("position"),
576                e.args.get("occurrence"),
577            ),
578            exp.RegexpReplace: regexp_replace_sql,
579            exp.RegexpLike: rename_func("REGEXP_CONTAINS"),
580            exp.ReturnsProperty: _returnsproperty_sql,
581            exp.Select: transforms.preprocess(
582                [
583                    transforms.explode_to_unnest(),
584                    _unqualify_unnest,
585                    transforms.eliminate_distinct_on,
586                    _alias_ordered_group,
587                    transforms.eliminate_semi_and_anti_joins,
588                ]
589            ),
590            exp.SHA2: lambda self, e: self.func(
591                f"SHA256" if e.text("length") == "256" else "SHA512", e.this
592            ),
593            exp.StabilityProperty: lambda self, e: f"DETERMINISTIC"
594            if e.name == "IMMUTABLE"
595            else "NOT DETERMINISTIC",
596            exp.StrToDate: lambda self, e: f"PARSE_DATE({self.format_time(e)}, {self.sql(e, 'this')})",
597            exp.StrToTime: lambda self, e: self.func(
598                "PARSE_TIMESTAMP", self.format_time(e), e.this, e.args.get("zone")
599            ),
600            exp.TimeAdd: date_add_interval_sql("TIME", "ADD"),
601            exp.TimeFromParts: rename_func("TIME"),
602            exp.TimeSub: date_add_interval_sql("TIME", "SUB"),
603            exp.TimestampAdd: date_add_interval_sql("TIMESTAMP", "ADD"),
604            exp.TimestampSub: date_add_interval_sql("TIMESTAMP", "SUB"),
605            exp.TimeStrToTime: timestrtotime_sql,
606            exp.Trim: lambda self, e: self.func(f"TRIM", e.this, e.expression),
607            exp.TsOrDsAdd: _ts_or_ds_add_sql,
608            exp.TsOrDsDiff: _ts_or_ds_diff_sql,
609            exp.TsOrDsToTime: rename_func("TIME"),
610            exp.Unhex: rename_func("FROM_HEX"),
611            exp.UnixDate: rename_func("UNIX_DATE"),
612            exp.UnixToTime: _unix_to_time_sql,
613            exp.Values: _derived_table_values_to_unnest,
614            exp.VariancePop: rename_func("VAR_POP"),
615        }
616
617        TYPE_MAPPING = {
618            **generator.Generator.TYPE_MAPPING,
619            exp.DataType.Type.BIGDECIMAL: "BIGNUMERIC",
620            exp.DataType.Type.BIGINT: "INT64",
621            exp.DataType.Type.BINARY: "BYTES",
622            exp.DataType.Type.BOOLEAN: "BOOL",
623            exp.DataType.Type.CHAR: "STRING",
624            exp.DataType.Type.DECIMAL: "NUMERIC",
625            exp.DataType.Type.DOUBLE: "FLOAT64",
626            exp.DataType.Type.FLOAT: "FLOAT64",
627            exp.DataType.Type.INT: "INT64",
628            exp.DataType.Type.NCHAR: "STRING",
629            exp.DataType.Type.NVARCHAR: "STRING",
630            exp.DataType.Type.SMALLINT: "INT64",
631            exp.DataType.Type.TEXT: "STRING",
632            exp.DataType.Type.TIMESTAMP: "DATETIME",
633            exp.DataType.Type.TIMESTAMPTZ: "TIMESTAMP",
634            exp.DataType.Type.TIMESTAMPLTZ: "TIMESTAMP",
635            exp.DataType.Type.TINYINT: "INT64",
636            exp.DataType.Type.VARBINARY: "BYTES",
637            exp.DataType.Type.VARCHAR: "STRING",
638            exp.DataType.Type.VARIANT: "ANY TYPE",
639        }
640
641        PROPERTIES_LOCATION = {
642            **generator.Generator.PROPERTIES_LOCATION,
643            exp.PartitionedByProperty: exp.Properties.Location.POST_SCHEMA,
644            exp.VolatileProperty: exp.Properties.Location.UNSUPPORTED,
645        }
646
647        # from: https://cloud.google.com/bigquery/docs/reference/standard-sql/lexical#reserved_keywords
648        RESERVED_KEYWORDS = {
649            *generator.Generator.RESERVED_KEYWORDS,
650            "all",
651            "and",
652            "any",
653            "array",
654            "as",
655            "asc",
656            "assert_rows_modified",
657            "at",
658            "between",
659            "by",
660            "case",
661            "cast",
662            "collate",
663            "contains",
664            "create",
665            "cross",
666            "cube",
667            "current",
668            "default",
669            "define",
670            "desc",
671            "distinct",
672            "else",
673            "end",
674            "enum",
675            "escape",
676            "except",
677            "exclude",
678            "exists",
679            "extract",
680            "false",
681            "fetch",
682            "following",
683            "for",
684            "from",
685            "full",
686            "group",
687            "grouping",
688            "groups",
689            "hash",
690            "having",
691            "if",
692            "ignore",
693            "in",
694            "inner",
695            "intersect",
696            "interval",
697            "into",
698            "is",
699            "join",
700            "lateral",
701            "left",
702            "like",
703            "limit",
704            "lookup",
705            "merge",
706            "natural",
707            "new",
708            "no",
709            "not",
710            "null",
711            "nulls",
712            "of",
713            "on",
714            "or",
715            "order",
716            "outer",
717            "over",
718            "partition",
719            "preceding",
720            "proto",
721            "qualify",
722            "range",
723            "recursive",
724            "respect",
725            "right",
726            "rollup",
727            "rows",
728            "select",
729            "set",
730            "some",
731            "struct",
732            "tablesample",
733            "then",
734            "to",
735            "treat",
736            "true",
737            "unbounded",
738            "union",
739            "unnest",
740            "using",
741            "when",
742            "where",
743            "window",
744            "with",
745            "within",
746        }
747
748        def timetostr_sql(self, expression: exp.TimeToStr) -> str:
749            if isinstance(expression.this, exp.TsOrDsToDate):
750                this: exp.Expression = expression.this
751            else:
752                this = expression
753
754            return f"FORMAT_DATE({self.format_time(expression)}, {self.sql(this, 'this')})"
755
756        def struct_sql(self, expression: exp.Struct) -> str:
757            args = []
758            for expr in expression.expressions:
759                if isinstance(expr, self.KEY_VALUE_DEFINITIONS):
760                    arg = f"{self.sql(expr, 'expression')} AS {expr.this.name}"
761                else:
762                    arg = self.sql(expr)
763
764                args.append(arg)
765
766            return self.func("STRUCT", *args)
767
768        def eq_sql(self, expression: exp.EQ) -> str:
769            # Operands of = cannot be NULL in BigQuery
770            if isinstance(expression.left, exp.Null) or isinstance(expression.right, exp.Null):
771                if not isinstance(expression.parent, exp.Update):
772                    return "NULL"
773
774            return self.binary(expression, "=")
775
776        def attimezone_sql(self, expression: exp.AtTimeZone) -> str:
777            parent = expression.parent
778
779            # BigQuery allows CAST(.. AS {STRING|TIMESTAMP} [FORMAT <fmt> [AT TIME ZONE <tz>]]).
780            # Only the TIMESTAMP one should use the below conversion, when AT TIME ZONE is included.
781            if not isinstance(parent, exp.Cast) or not parent.to.is_type("text"):
782                return self.func(
783                    "TIMESTAMP", self.func("DATETIME", expression.this, expression.args.get("zone"))
784                )
785
786            return super().attimezone_sql(expression)
787
788        def trycast_sql(self, expression: exp.TryCast) -> str:
789            return self.cast_sql(expression, safe_prefix="SAFE_")
790
791        def cte_sql(self, expression: exp.CTE) -> str:
792            if expression.alias_column_names:
793                self.unsupported("Column names in CTE definition are not supported.")
794            return super().cte_sql(expression)
795
796        def array_sql(self, expression: exp.Array) -> str:
797            first_arg = seq_get(expression.expressions, 0)
798            if isinstance(first_arg, exp.Subqueryable):
799                return f"ARRAY{self.wrap(self.sql(first_arg))}"
800
801            return inline_array_sql(self, expression)
802
803        def bracket_sql(self, expression: exp.Bracket) -> str:
804            this = self.sql(expression, "this")
805            expressions = expression.expressions
806
807            if len(expressions) == 1:
808                arg = expressions[0]
809                if arg.type is None:
810                    from sqlglot.optimizer.annotate_types import annotate_types
811
812                    arg = annotate_types(arg)
813
814                if arg.type and arg.type.this in exp.DataType.TEXT_TYPES:
815                    # BQ doesn't support bracket syntax with string values
816                    return f"{this}.{arg.name}"
817
818            expressions_sql = ", ".join(self.sql(e) for e in expressions)
819            offset = expression.args.get("offset")
820
821            if offset == 0:
822                expressions_sql = f"OFFSET({expressions_sql})"
823            elif offset == 1:
824                expressions_sql = f"ORDINAL({expressions_sql})"
825            elif offset is not None:
826                self.unsupported(f"Unsupported array offset: {offset}")
827
828            if expression.args.get("safe"):
829                expressions_sql = f"SAFE_{expressions_sql}"
830
831            return f"{this}[{expressions_sql}]"
832
833        def transaction_sql(self, *_) -> str:
834            return "BEGIN TRANSACTION"
835
836        def commit_sql(self, *_) -> str:
837            return "COMMIT TRANSACTION"
838
839        def rollback_sql(self, *_) -> str:
840            return "ROLLBACK TRANSACTION"
841
842        def in_unnest_op(self, expression: exp.Unnest) -> str:
843            return self.sql(expression)
844
845        def except_op(self, expression: exp.Except) -> str:
846            if not expression.args.get("distinct", False):
847                self.unsupported("EXCEPT without DISTINCT is not supported in BigQuery")
848            return f"EXCEPT{' DISTINCT' if expression.args.get('distinct') else ' ALL'}"
849
850        def intersect_op(self, expression: exp.Intersect) -> str:
851            if not expression.args.get("distinct", False):
852                self.unsupported("INTERSECT without DISTINCT is not supported in BigQuery")
853            return f"INTERSECT{' DISTINCT' if expression.args.get('distinct') else ' ALL'}"
854
855        def with_properties(self, properties: exp.Properties) -> str:
856            return self.properties(properties, prefix=self.seg("OPTIONS"))
857
858        def version_sql(self, expression: exp.Version) -> str:
859            if expression.name == "TIMESTAMP":
860                expression.set("this", "SYSTEM_TIME")
861            return super().version_sql(expression)
logger = <Logger sqlglot (WARNING)>
class BigQuery(sqlglot.dialects.dialect.Dialect):
227class BigQuery(Dialect):
228    WEEK_OFFSET = -1
229    UNNEST_COLUMN_ONLY = True
230    SUPPORTS_USER_DEFINED_TYPES = False
231    SUPPORTS_SEMI_ANTI_JOIN = False
232    LOG_BASE_FIRST = False
233
234    # https://cloud.google.com/bigquery/docs/reference/standard-sql/lexical#case_sensitivity
235    NORMALIZATION_STRATEGY = NormalizationStrategy.CASE_INSENSITIVE
236
237    # bigquery udfs are case sensitive
238    NORMALIZE_FUNCTIONS = False
239
240    TIME_MAPPING = {
241        "%D": "%m/%d/%y",
242    }
243
244    ESCAPE_SEQUENCES = {
245        "\\a": "\a",
246        "\\b": "\b",
247        "\\f": "\f",
248        "\\n": "\n",
249        "\\r": "\r",
250        "\\t": "\t",
251        "\\v": "\v",
252    }
253
254    FORMAT_MAPPING = {
255        "DD": "%d",
256        "MM": "%m",
257        "MON": "%b",
258        "MONTH": "%B",
259        "YYYY": "%Y",
260        "YY": "%y",
261        "HH": "%I",
262        "HH12": "%I",
263        "HH24": "%H",
264        "MI": "%M",
265        "SS": "%S",
266        "SSSSS": "%f",
267        "TZH": "%z",
268    }
269
270    # The _PARTITIONTIME and _PARTITIONDATE pseudo-columns are not returned by a SELECT * statement
271    # https://cloud.google.com/bigquery/docs/querying-partitioned-tables#query_an_ingestion-time_partitioned_table
272    PSEUDOCOLUMNS = {"_PARTITIONTIME", "_PARTITIONDATE"}
273
274    def normalize_identifier(self, expression: E) -> E:
275        if isinstance(expression, exp.Identifier):
276            parent = expression.parent
277            while isinstance(parent, exp.Dot):
278                parent = parent.parent
279
280            # In BigQuery, CTEs aren't case-sensitive, but table names are (by default, at least).
281            # The following check is essentially a heuristic to detect tables based on whether or
282            # not they're qualified. It also avoids normalizing UDFs, because they're case-sensitive.
283            if (
284                not isinstance(parent, exp.UserDefinedFunction)
285                and not (isinstance(parent, exp.Table) and parent.db)
286                and not expression.meta.get("is_table")
287            ):
288                expression.set("this", expression.this.lower())
289
290        return expression
291
292    class Tokenizer(tokens.Tokenizer):
293        QUOTES = ["'", '"', '"""', "'''"]
294        COMMENTS = ["--", "#", ("/*", "*/")]
295        IDENTIFIERS = ["`"]
296        STRING_ESCAPES = ["\\"]
297
298        HEX_STRINGS = [("0x", ""), ("0X", "")]
299
300        BYTE_STRINGS = [
301            (prefix + q, q) for q in t.cast(t.List[str], QUOTES) for prefix in ("b", "B")
302        ]
303
304        RAW_STRINGS = [
305            (prefix + q, q) for q in t.cast(t.List[str], QUOTES) for prefix in ("r", "R")
306        ]
307
308        KEYWORDS = {
309            **tokens.Tokenizer.KEYWORDS,
310            "ANY TYPE": TokenType.VARIANT,
311            "BEGIN": TokenType.COMMAND,
312            "BEGIN TRANSACTION": TokenType.BEGIN,
313            "BYTES": TokenType.BINARY,
314            "CURRENT_DATETIME": TokenType.CURRENT_DATETIME,
315            "DECLARE": TokenType.COMMAND,
316            "FLOAT64": TokenType.DOUBLE,
317            "FOR SYSTEM_TIME": TokenType.TIMESTAMP_SNAPSHOT,
318            "MODEL": TokenType.MODEL,
319            "NOT DETERMINISTIC": TokenType.VOLATILE,
320            "RECORD": TokenType.STRUCT,
321            "TIMESTAMP": TokenType.TIMESTAMPTZ,
322        }
323        KEYWORDS.pop("DIV")
324
325    class Parser(parser.Parser):
326        PREFIXED_PIVOT_COLUMNS = True
327
328        LOG_DEFAULTS_TO_LN = True
329
330        FUNCTIONS = {
331            **parser.Parser.FUNCTIONS,
332            "DATE": _parse_date,
333            "DATE_ADD": parse_date_delta_with_interval(exp.DateAdd),
334            "DATE_SUB": parse_date_delta_with_interval(exp.DateSub),
335            "DATE_TRUNC": lambda args: exp.DateTrunc(
336                unit=exp.Literal.string(str(seq_get(args, 1))),
337                this=seq_get(args, 0),
338            ),
339            "DATETIME_ADD": parse_date_delta_with_interval(exp.DatetimeAdd),
340            "DATETIME_SUB": parse_date_delta_with_interval(exp.DatetimeSub),
341            "DIV": binary_from_function(exp.IntDiv),
342            "FORMAT_DATE": lambda args: exp.TimeToStr(
343                this=exp.TsOrDsToDate(this=seq_get(args, 1)), format=seq_get(args, 0)
344            ),
345            "GENERATE_ARRAY": exp.GenerateSeries.from_arg_list,
346            "JSON_EXTRACT_SCALAR": lambda args: exp.JSONExtractScalar(
347                this=seq_get(args, 0), expression=seq_get(args, 1) or exp.Literal.string("$")
348            ),
349            "MD5": exp.MD5Digest.from_arg_list,
350            "TO_HEX": _parse_to_hex,
351            "PARSE_DATE": lambda args: format_time_lambda(exp.StrToDate, "bigquery")(
352                [seq_get(args, 1), seq_get(args, 0)]
353            ),
354            "PARSE_TIMESTAMP": _parse_timestamp,
355            "REGEXP_CONTAINS": exp.RegexpLike.from_arg_list,
356            "REGEXP_EXTRACT": lambda args: exp.RegexpExtract(
357                this=seq_get(args, 0),
358                expression=seq_get(args, 1),
359                position=seq_get(args, 2),
360                occurrence=seq_get(args, 3),
361                group=exp.Literal.number(1) if re.compile(args[1].name).groups == 1 else None,
362            ),
363            "SHA256": lambda args: exp.SHA2(this=seq_get(args, 0), length=exp.Literal.number(256)),
364            "SHA512": lambda args: exp.SHA2(this=seq_get(args, 0), length=exp.Literal.number(512)),
365            "SPLIT": lambda args: exp.Split(
366                # https://cloud.google.com/bigquery/docs/reference/standard-sql/string_functions#split
367                this=seq_get(args, 0),
368                expression=seq_get(args, 1) or exp.Literal.string(","),
369            ),
370            "TIME": _parse_time,
371            "TIME_ADD": parse_date_delta_with_interval(exp.TimeAdd),
372            "TIME_SUB": parse_date_delta_with_interval(exp.TimeSub),
373            "TIMESTAMP_ADD": parse_date_delta_with_interval(exp.TimestampAdd),
374            "TIMESTAMP_SUB": parse_date_delta_with_interval(exp.TimestampSub),
375            "TIMESTAMP_MICROS": lambda args: exp.UnixToTime(
376                this=seq_get(args, 0), scale=exp.UnixToTime.MICROS
377            ),
378            "TIMESTAMP_MILLIS": lambda args: exp.UnixToTime(
379                this=seq_get(args, 0), scale=exp.UnixToTime.MILLIS
380            ),
381            "TIMESTAMP_SECONDS": lambda args: exp.UnixToTime(
382                this=seq_get(args, 0), scale=exp.UnixToTime.SECONDS
383            ),
384            "TO_JSON_STRING": exp.JSONFormat.from_arg_list,
385        }
386
387        FUNCTION_PARSERS = {
388            **parser.Parser.FUNCTION_PARSERS,
389            "ARRAY": lambda self: self.expression(exp.Array, expressions=[self._parse_statement()]),
390        }
391        FUNCTION_PARSERS.pop("TRIM")
392
393        NO_PAREN_FUNCTIONS = {
394            **parser.Parser.NO_PAREN_FUNCTIONS,
395            TokenType.CURRENT_DATETIME: exp.CurrentDatetime,
396        }
397
398        NESTED_TYPE_TOKENS = {
399            *parser.Parser.NESTED_TYPE_TOKENS,
400            TokenType.TABLE,
401        }
402
403        ID_VAR_TOKENS = {
404            *parser.Parser.ID_VAR_TOKENS,
405            TokenType.VALUES,
406        }
407
408        PROPERTY_PARSERS = {
409            **parser.Parser.PROPERTY_PARSERS,
410            "NOT DETERMINISTIC": lambda self: self.expression(
411                exp.StabilityProperty, this=exp.Literal.string("VOLATILE")
412            ),
413            "OPTIONS": lambda self: self._parse_with_property(),
414        }
415
416        CONSTRAINT_PARSERS = {
417            **parser.Parser.CONSTRAINT_PARSERS,
418            "OPTIONS": lambda self: exp.Properties(expressions=self._parse_with_property()),
419        }
420
421        RANGE_PARSERS = parser.Parser.RANGE_PARSERS.copy()
422        RANGE_PARSERS.pop(TokenType.OVERLAPS, None)
423
424        NULL_TOKENS = {TokenType.NULL, TokenType.UNKNOWN}
425
426        STATEMENT_PARSERS = {
427            **parser.Parser.STATEMENT_PARSERS,
428            TokenType.END: lambda self: self._parse_as_command(self._prev),
429            TokenType.FOR: lambda self: self._parse_for_in(),
430        }
431
432        BRACKET_OFFSETS = {
433            "OFFSET": (0, False),
434            "ORDINAL": (1, False),
435            "SAFE_OFFSET": (0, True),
436            "SAFE_ORDINAL": (1, True),
437        }
438
439        def _parse_for_in(self) -> exp.ForIn:
440            this = self._parse_range()
441            self._match_text_seq("DO")
442            return self.expression(exp.ForIn, this=this, expression=self._parse_statement())
443
444        def _parse_table_part(self, schema: bool = False) -> t.Optional[exp.Expression]:
445            this = super()._parse_table_part(schema=schema) or self._parse_number()
446
447            # https://cloud.google.com/bigquery/docs/reference/standard-sql/lexical#table_names
448            if isinstance(this, exp.Identifier):
449                table_name = this.name
450                while self._match(TokenType.DASH, advance=False) and self._next:
451                    self._advance(2)
452                    table_name += f"-{self._prev.text}"
453
454                this = exp.Identifier(this=table_name, quoted=this.args.get("quoted"))
455            elif isinstance(this, exp.Literal):
456                table_name = this.name
457
458                if self._is_connected() and self._parse_var(any_token=True):
459                    table_name += self._prev.text
460
461                this = exp.Identifier(this=table_name, quoted=True)
462
463            return this
464
465        def _parse_table_parts(self, schema: bool = False) -> exp.Table:
466            table = super()._parse_table_parts(schema=schema)
467            if isinstance(table.this, exp.Identifier) and "." in table.name:
468                catalog, db, this, *rest = (
469                    t.cast(t.Optional[exp.Expression], exp.to_identifier(x))
470                    for x in split_num_words(table.name, ".", 3)
471                )
472
473                if rest and this:
474                    this = exp.Dot.build(t.cast(t.List[exp.Expression], [this, *rest]))
475
476                table = exp.Table(this=this, db=db, catalog=catalog)
477
478            return table
479
480        def _parse_json_object(self) -> exp.JSONObject:
481            json_object = super()._parse_json_object()
482            array_kv_pair = seq_get(json_object.expressions, 0)
483
484            # Converts BQ's "signature 2" of JSON_OBJECT into SQLGlot's canonical representation
485            # https://cloud.google.com/bigquery/docs/reference/standard-sql/json_functions#json_object_signature2
486            if (
487                array_kv_pair
488                and isinstance(array_kv_pair.this, exp.Array)
489                and isinstance(array_kv_pair.expression, exp.Array)
490            ):
491                keys = array_kv_pair.this.expressions
492                values = array_kv_pair.expression.expressions
493
494                json_object.set(
495                    "expressions",
496                    [exp.JSONKeyValue(this=k, expression=v) for k, v in zip(keys, values)],
497                )
498
499            return json_object
500
501        def _parse_bracket(self, this: t.Optional[exp.Expression]) -> t.Optional[exp.Expression]:
502            bracket = super()._parse_bracket(this)
503
504            if this is bracket:
505                return bracket
506
507            if isinstance(bracket, exp.Bracket):
508                for expression in bracket.expressions:
509                    name = expression.name.upper()
510
511                    if name not in self.BRACKET_OFFSETS:
512                        break
513
514                    offset, safe = self.BRACKET_OFFSETS[name]
515                    bracket.set("offset", offset)
516                    bracket.set("safe", safe)
517                    expression.replace(expression.expressions[0])
518
519            return bracket
520
521    class Generator(generator.Generator):
522        EXPLICIT_UNION = True
523        INTERVAL_ALLOWS_PLURAL_FORM = False
524        JOIN_HINTS = False
525        QUERY_HINTS = False
526        TABLE_HINTS = False
527        LIMIT_FETCH = "LIMIT"
528        RENAME_TABLE_WITH_DB = False
529        NVL2_SUPPORTED = False
530        UNNEST_WITH_ORDINALITY = False
531        COLLATE_IS_FUNC = True
532        LIMIT_ONLY_LITERALS = True
533        SUPPORTS_TABLE_ALIAS_COLUMNS = False
534        UNPIVOT_ALIASES_ARE_IDENTIFIERS = False
535
536        TRANSFORMS = {
537            **generator.Generator.TRANSFORMS,
538            exp.ApproxDistinct: rename_func("APPROX_COUNT_DISTINCT"),
539            exp.ArgMax: arg_max_or_min_no_count("MAX_BY"),
540            exp.ArgMin: arg_max_or_min_no_count("MIN_BY"),
541            exp.ArrayContains: _array_contains_sql,
542            exp.ArraySize: rename_func("ARRAY_LENGTH"),
543            exp.Cast: transforms.preprocess([transforms.remove_precision_parameterized_types]),
544            exp.CollateProperty: lambda self, e: f"DEFAULT COLLATE {self.sql(e, 'this')}"
545            if e.args.get("default")
546            else f"COLLATE {self.sql(e, 'this')}",
547            exp.CountIf: rename_func("COUNTIF"),
548            exp.Create: _create_sql,
549            exp.CTE: transforms.preprocess([_pushdown_cte_column_names]),
550            exp.DateAdd: date_add_interval_sql("DATE", "ADD"),
551            exp.DateDiff: lambda self, e: f"DATE_DIFF({self.sql(e, 'this')}, {self.sql(e, 'expression')}, {self.sql(e.args.get('unit', 'DAY'))})",
552            exp.DateFromParts: rename_func("DATE"),
553            exp.DateStrToDate: datestrtodate_sql,
554            exp.DateSub: date_add_interval_sql("DATE", "SUB"),
555            exp.DatetimeAdd: date_add_interval_sql("DATETIME", "ADD"),
556            exp.DatetimeSub: date_add_interval_sql("DATETIME", "SUB"),
557            exp.DateTrunc: lambda self, e: self.func("DATE_TRUNC", e.this, e.text("unit")),
558            exp.GenerateSeries: rename_func("GENERATE_ARRAY"),
559            exp.GetPath: path_to_jsonpath(),
560            exp.GroupConcat: rename_func("STRING_AGG"),
561            exp.Hex: rename_func("TO_HEX"),
562            exp.If: if_sql(false_value="NULL"),
563            exp.ILike: no_ilike_sql,
564            exp.IntDiv: rename_func("DIV"),
565            exp.JSONFormat: rename_func("TO_JSON_STRING"),
566            exp.JSONKeyValue: json_keyvalue_comma_sql,
567            exp.Max: max_or_greatest,
568            exp.MD5: lambda self, e: self.func("TO_HEX", self.func("MD5", e.this)),
569            exp.MD5Digest: rename_func("MD5"),
570            exp.Min: min_or_least,
571            exp.PartitionedByProperty: lambda self, e: f"PARTITION BY {self.sql(e, 'this')}",
572            exp.RegexpExtract: lambda self, e: self.func(
573                "REGEXP_EXTRACT",
574                e.this,
575                e.expression,
576                e.args.get("position"),
577                e.args.get("occurrence"),
578            ),
579            exp.RegexpReplace: regexp_replace_sql,
580            exp.RegexpLike: rename_func("REGEXP_CONTAINS"),
581            exp.ReturnsProperty: _returnsproperty_sql,
582            exp.Select: transforms.preprocess(
583                [
584                    transforms.explode_to_unnest(),
585                    _unqualify_unnest,
586                    transforms.eliminate_distinct_on,
587                    _alias_ordered_group,
588                    transforms.eliminate_semi_and_anti_joins,
589                ]
590            ),
591            exp.SHA2: lambda self, e: self.func(
592                f"SHA256" if e.text("length") == "256" else "SHA512", e.this
593            ),
594            exp.StabilityProperty: lambda self, e: f"DETERMINISTIC"
595            if e.name == "IMMUTABLE"
596            else "NOT DETERMINISTIC",
597            exp.StrToDate: lambda self, e: f"PARSE_DATE({self.format_time(e)}, {self.sql(e, 'this')})",
598            exp.StrToTime: lambda self, e: self.func(
599                "PARSE_TIMESTAMP", self.format_time(e), e.this, e.args.get("zone")
600            ),
601            exp.TimeAdd: date_add_interval_sql("TIME", "ADD"),
602            exp.TimeFromParts: rename_func("TIME"),
603            exp.TimeSub: date_add_interval_sql("TIME", "SUB"),
604            exp.TimestampAdd: date_add_interval_sql("TIMESTAMP", "ADD"),
605            exp.TimestampSub: date_add_interval_sql("TIMESTAMP", "SUB"),
606            exp.TimeStrToTime: timestrtotime_sql,
607            exp.Trim: lambda self, e: self.func(f"TRIM", e.this, e.expression),
608            exp.TsOrDsAdd: _ts_or_ds_add_sql,
609            exp.TsOrDsDiff: _ts_or_ds_diff_sql,
610            exp.TsOrDsToTime: rename_func("TIME"),
611            exp.Unhex: rename_func("FROM_HEX"),
612            exp.UnixDate: rename_func("UNIX_DATE"),
613            exp.UnixToTime: _unix_to_time_sql,
614            exp.Values: _derived_table_values_to_unnest,
615            exp.VariancePop: rename_func("VAR_POP"),
616        }
617
618        TYPE_MAPPING = {
619            **generator.Generator.TYPE_MAPPING,
620            exp.DataType.Type.BIGDECIMAL: "BIGNUMERIC",
621            exp.DataType.Type.BIGINT: "INT64",
622            exp.DataType.Type.BINARY: "BYTES",
623            exp.DataType.Type.BOOLEAN: "BOOL",
624            exp.DataType.Type.CHAR: "STRING",
625            exp.DataType.Type.DECIMAL: "NUMERIC",
626            exp.DataType.Type.DOUBLE: "FLOAT64",
627            exp.DataType.Type.FLOAT: "FLOAT64",
628            exp.DataType.Type.INT: "INT64",
629            exp.DataType.Type.NCHAR: "STRING",
630            exp.DataType.Type.NVARCHAR: "STRING",
631            exp.DataType.Type.SMALLINT: "INT64",
632            exp.DataType.Type.TEXT: "STRING",
633            exp.DataType.Type.TIMESTAMP: "DATETIME",
634            exp.DataType.Type.TIMESTAMPTZ: "TIMESTAMP",
635            exp.DataType.Type.TIMESTAMPLTZ: "TIMESTAMP",
636            exp.DataType.Type.TINYINT: "INT64",
637            exp.DataType.Type.VARBINARY: "BYTES",
638            exp.DataType.Type.VARCHAR: "STRING",
639            exp.DataType.Type.VARIANT: "ANY TYPE",
640        }
641
642        PROPERTIES_LOCATION = {
643            **generator.Generator.PROPERTIES_LOCATION,
644            exp.PartitionedByProperty: exp.Properties.Location.POST_SCHEMA,
645            exp.VolatileProperty: exp.Properties.Location.UNSUPPORTED,
646        }
647
648        # from: https://cloud.google.com/bigquery/docs/reference/standard-sql/lexical#reserved_keywords
649        RESERVED_KEYWORDS = {
650            *generator.Generator.RESERVED_KEYWORDS,
651            "all",
652            "and",
653            "any",
654            "array",
655            "as",
656            "asc",
657            "assert_rows_modified",
658            "at",
659            "between",
660            "by",
661            "case",
662            "cast",
663            "collate",
664            "contains",
665            "create",
666            "cross",
667            "cube",
668            "current",
669            "default",
670            "define",
671            "desc",
672            "distinct",
673            "else",
674            "end",
675            "enum",
676            "escape",
677            "except",
678            "exclude",
679            "exists",
680            "extract",
681            "false",
682            "fetch",
683            "following",
684            "for",
685            "from",
686            "full",
687            "group",
688            "grouping",
689            "groups",
690            "hash",
691            "having",
692            "if",
693            "ignore",
694            "in",
695            "inner",
696            "intersect",
697            "interval",
698            "into",
699            "is",
700            "join",
701            "lateral",
702            "left",
703            "like",
704            "limit",
705            "lookup",
706            "merge",
707            "natural",
708            "new",
709            "no",
710            "not",
711            "null",
712            "nulls",
713            "of",
714            "on",
715            "or",
716            "order",
717            "outer",
718            "over",
719            "partition",
720            "preceding",
721            "proto",
722            "qualify",
723            "range",
724            "recursive",
725            "respect",
726            "right",
727            "rollup",
728            "rows",
729            "select",
730            "set",
731            "some",
732            "struct",
733            "tablesample",
734            "then",
735            "to",
736            "treat",
737            "true",
738            "unbounded",
739            "union",
740            "unnest",
741            "using",
742            "when",
743            "where",
744            "window",
745            "with",
746            "within",
747        }
748
749        def timetostr_sql(self, expression: exp.TimeToStr) -> str:
750            if isinstance(expression.this, exp.TsOrDsToDate):
751                this: exp.Expression = expression.this
752            else:
753                this = expression
754
755            return f"FORMAT_DATE({self.format_time(expression)}, {self.sql(this, 'this')})"
756
757        def struct_sql(self, expression: exp.Struct) -> str:
758            args = []
759            for expr in expression.expressions:
760                if isinstance(expr, self.KEY_VALUE_DEFINITIONS):
761                    arg = f"{self.sql(expr, 'expression')} AS {expr.this.name}"
762                else:
763                    arg = self.sql(expr)
764
765                args.append(arg)
766
767            return self.func("STRUCT", *args)
768
769        def eq_sql(self, expression: exp.EQ) -> str:
770            # Operands of = cannot be NULL in BigQuery
771            if isinstance(expression.left, exp.Null) or isinstance(expression.right, exp.Null):
772                if not isinstance(expression.parent, exp.Update):
773                    return "NULL"
774
775            return self.binary(expression, "=")
776
777        def attimezone_sql(self, expression: exp.AtTimeZone) -> str:
778            parent = expression.parent
779
780            # BigQuery allows CAST(.. AS {STRING|TIMESTAMP} [FORMAT <fmt> [AT TIME ZONE <tz>]]).
781            # Only the TIMESTAMP one should use the below conversion, when AT TIME ZONE is included.
782            if not isinstance(parent, exp.Cast) or not parent.to.is_type("text"):
783                return self.func(
784                    "TIMESTAMP", self.func("DATETIME", expression.this, expression.args.get("zone"))
785                )
786
787            return super().attimezone_sql(expression)
788
789        def trycast_sql(self, expression: exp.TryCast) -> str:
790            return self.cast_sql(expression, safe_prefix="SAFE_")
791
792        def cte_sql(self, expression: exp.CTE) -> str:
793            if expression.alias_column_names:
794                self.unsupported("Column names in CTE definition are not supported.")
795            return super().cte_sql(expression)
796
797        def array_sql(self, expression: exp.Array) -> str:
798            first_arg = seq_get(expression.expressions, 0)
799            if isinstance(first_arg, exp.Subqueryable):
800                return f"ARRAY{self.wrap(self.sql(first_arg))}"
801
802            return inline_array_sql(self, expression)
803
804        def bracket_sql(self, expression: exp.Bracket) -> str:
805            this = self.sql(expression, "this")
806            expressions = expression.expressions
807
808            if len(expressions) == 1:
809                arg = expressions[0]
810                if arg.type is None:
811                    from sqlglot.optimizer.annotate_types import annotate_types
812
813                    arg = annotate_types(arg)
814
815                if arg.type and arg.type.this in exp.DataType.TEXT_TYPES:
816                    # BQ doesn't support bracket syntax with string values
817                    return f"{this}.{arg.name}"
818
819            expressions_sql = ", ".join(self.sql(e) for e in expressions)
820            offset = expression.args.get("offset")
821
822            if offset == 0:
823                expressions_sql = f"OFFSET({expressions_sql})"
824            elif offset == 1:
825                expressions_sql = f"ORDINAL({expressions_sql})"
826            elif offset is not None:
827                self.unsupported(f"Unsupported array offset: {offset}")
828
829            if expression.args.get("safe"):
830                expressions_sql = f"SAFE_{expressions_sql}"
831
832            return f"{this}[{expressions_sql}]"
833
834        def transaction_sql(self, *_) -> str:
835            return "BEGIN TRANSACTION"
836
837        def commit_sql(self, *_) -> str:
838            return "COMMIT TRANSACTION"
839
840        def rollback_sql(self, *_) -> str:
841            return "ROLLBACK TRANSACTION"
842
843        def in_unnest_op(self, expression: exp.Unnest) -> str:
844            return self.sql(expression)
845
846        def except_op(self, expression: exp.Except) -> str:
847            if not expression.args.get("distinct", False):
848                self.unsupported("EXCEPT without DISTINCT is not supported in BigQuery")
849            return f"EXCEPT{' DISTINCT' if expression.args.get('distinct') else ' ALL'}"
850
851        def intersect_op(self, expression: exp.Intersect) -> str:
852            if not expression.args.get("distinct", False):
853                self.unsupported("INTERSECT without DISTINCT is not supported in BigQuery")
854            return f"INTERSECT{' DISTINCT' if expression.args.get('distinct') else ' ALL'}"
855
856        def with_properties(self, properties: exp.Properties) -> str:
857            return self.properties(properties, prefix=self.seg("OPTIONS"))
858
859        def version_sql(self, expression: exp.Version) -> str:
860            if expression.name == "TIMESTAMP":
861                expression.set("this", "SYSTEM_TIME")
862            return super().version_sql(expression)
WEEK_OFFSET = -1

Determines the day of week of DATE_TRUNC(week). Defaults to 0 (Monday). -1 would be Sunday.

UNNEST_COLUMN_ONLY = True

Determines whether or not UNNEST table aliases are treated as column aliases.

SUPPORTS_USER_DEFINED_TYPES = False

Determines whether or not user-defined data types are supported.

SUPPORTS_SEMI_ANTI_JOIN = False

Determines whether or not SEMI or ANTI joins are supported.

LOG_BASE_FIRST = False

Determines whether the base comes first in the LOG function.

NORMALIZATION_STRATEGY = <NormalizationStrategy.CASE_INSENSITIVE: 'CASE_INSENSITIVE'>

Specifies the strategy according to which identifiers should be normalized.

NORMALIZE_FUNCTIONS: bool | str = False

Determines how function names are going to be normalized.

TIME_MAPPING: Dict[str, str] = {'%D': '%m/%d/%y'}

Associates this dialect's time formats with their equivalent Python strftime format.

ESCAPE_SEQUENCES: Dict[str, str] = {'\\a': '\x07', '\\b': '\x08', '\\f': '\x0c', '\\n': '\n', '\\r': '\r', '\\t': '\t', '\\v': '\x0b'}

Mapping of an unescaped escape sequence to the corresponding character.

FORMAT_MAPPING: Dict[str, str] = {'DD': '%d', 'MM': '%m', 'MON': '%b', 'MONTH': '%B', 'YYYY': '%Y', 'YY': '%y', 'HH': '%I', 'HH12': '%I', 'HH24': '%H', 'MI': '%M', 'SS': '%S', 'SSSSS': '%f', 'TZH': '%z'}

Helper which is used for parsing the special syntax CAST(x AS DATE FORMAT 'yyyy'). If empty, the corresponding trie will be constructed off of TIME_MAPPING.

PSEUDOCOLUMNS: Set[str] = {'_PARTITIONDATE', '_PARTITIONTIME'}

Columns that are auto-generated by the engine corresponding to this dialect. For example, such columns may be excluded from SELECT * queries.

def normalize_identifier(self, expression: ~E) -> ~E:
274    def normalize_identifier(self, expression: E) -> E:
275        if isinstance(expression, exp.Identifier):
276            parent = expression.parent
277            while isinstance(parent, exp.Dot):
278                parent = parent.parent
279
280            # In BigQuery, CTEs aren't case-sensitive, but table names are (by default, at least).
281            # The following check is essentially a heuristic to detect tables based on whether or
282            # not they're qualified. It also avoids normalizing UDFs, because they're case-sensitive.
283            if (
284                not isinstance(parent, exp.UserDefinedFunction)
285                and not (isinstance(parent, exp.Table) and parent.db)
286                and not expression.meta.get("is_table")
287            ):
288                expression.set("this", expression.this.lower())
289
290        return expression

Transforms an identifier in a way that resembles how it'd be resolved by this dialect.

For example, an identifier like FoO would be resolved as foo in Postgres, because it lowercases all unquoted identifiers. On the other hand, Snowflake uppercases them, so it would resolve it as FOO. If it was quoted, it'd need to be treated as case-sensitive, and so any normalization would be prohibited in order to avoid "breaking" the identifier.

There are also dialects like Spark, which are case-insensitive even when quotes are present, and dialects like MySQL, whose resolution rules match those employed by the underlying operating system, for example they may always be case-sensitive in Linux.

Finally, the normalization behavior of some engines can even be controlled through flags, like in Redshift's case, where users can explicitly set enable_case_sensitive_identifier.

SQLGlot aims to understand and handle all of these different behaviors gracefully, so that it can analyze queries in the optimizer and successfully capture their semantics.

tokenizer_class = <class 'BigQuery.Tokenizer'>
parser_class = <class 'BigQuery.Parser'>
generator_class = <class 'BigQuery.Generator'>
TIME_TRIE: Dict = {'%': {'D': {0: True}}}
FORMAT_TRIE: Dict = {'D': {'D': {0: True}}, 'M': {'M': {0: True}, 'O': {'N': {0: True, 'T': {'H': {0: True}}}}, 'I': {0: True}}, 'Y': {'Y': {'Y': {'Y': {0: True}}, 0: True}}, 'H': {'H': {0: True, '1': {'2': {0: True}}, '2': {'4': {0: True}}}}, 'S': {'S': {0: True, 'S': {'S': {'S': {0: True}}}}}, 'T': {'Z': {'H': {0: True}}}}
INVERSE_TIME_MAPPING: Dict[str, str] = {'%m/%d/%y': '%D'}
INVERSE_TIME_TRIE: Dict = {'%': {'m': {'/': {'%': {'d': {'/': {'%': {'y': {0: True}}}}}}}}}
INVERSE_ESCAPE_SEQUENCES: Dict[str, str] = {'\x07': '\\a', '\x08': '\\b', '\x0c': '\\f', '\n': '\\n', '\r': '\\r', '\t': '\\t', '\x0b': '\\v'}
QUOTE_START = "'"
QUOTE_END = "'"
IDENTIFIER_START = '`'
IDENTIFIER_END = '`'
BIT_START: Optional[str] = None
BIT_END: Optional[str] = None
HEX_START: Optional[str] = '0x'
HEX_END: Optional[str] = ''
BYTE_START: Optional[str] = "b'"
BYTE_END: Optional[str] = "'"
UNICODE_START: Optional[str] = None
UNICODE_END: Optional[str] = None
class BigQuery.Tokenizer(sqlglot.tokens.Tokenizer):
292    class Tokenizer(tokens.Tokenizer):
293        QUOTES = ["'", '"', '"""', "'''"]
294        COMMENTS = ["--", "#", ("/*", "*/")]
295        IDENTIFIERS = ["`"]
296        STRING_ESCAPES = ["\\"]
297
298        HEX_STRINGS = [("0x", ""), ("0X", "")]
299
300        BYTE_STRINGS = [
301            (prefix + q, q) for q in t.cast(t.List[str], QUOTES) for prefix in ("b", "B")
302        ]
303
304        RAW_STRINGS = [
305            (prefix + q, q) for q in t.cast(t.List[str], QUOTES) for prefix in ("r", "R")
306        ]
307
308        KEYWORDS = {
309            **tokens.Tokenizer.KEYWORDS,
310            "ANY TYPE": TokenType.VARIANT,
311            "BEGIN": TokenType.COMMAND,
312            "BEGIN TRANSACTION": TokenType.BEGIN,
313            "BYTES": TokenType.BINARY,
314            "CURRENT_DATETIME": TokenType.CURRENT_DATETIME,
315            "DECLARE": TokenType.COMMAND,
316            "FLOAT64": TokenType.DOUBLE,
317            "FOR SYSTEM_TIME": TokenType.TIMESTAMP_SNAPSHOT,
318            "MODEL": TokenType.MODEL,
319            "NOT DETERMINISTIC": TokenType.VOLATILE,
320            "RECORD": TokenType.STRUCT,
321            "TIMESTAMP": TokenType.TIMESTAMPTZ,
322        }
323        KEYWORDS.pop("DIV")
QUOTES = ["'", '"', '"""', "'''"]
COMMENTS = ['--', '#', ('/*', '*/')]
IDENTIFIERS = ['`']
STRING_ESCAPES = ['\\']
HEX_STRINGS = [('0x', ''), ('0X', '')]
BYTE_STRINGS = [("b'", "'"), ("B'", "'"), ('b"', '"'), ('B"', '"'), ('b"""', '"""'), ('B"""', '"""'), ("b'''", "'''"), ("B'''", "'''")]
RAW_STRINGS = [("r'", "'"), ("R'", "'"), ('r"', '"'), ('R"', '"'), ('r"""', '"""'), ('R"""', '"""'), ("r'''", "'''"), ("R'''", "'''")]
KEYWORDS = {'{%': <TokenType.BLOCK_START: 'BLOCK_START'>, '{%+': <TokenType.BLOCK_START: 'BLOCK_START'>, '{%-': <TokenType.BLOCK_START: 'BLOCK_START'>, '%}': <TokenType.BLOCK_END: 'BLOCK_END'>, '+%}': <TokenType.BLOCK_END: 'BLOCK_END'>, '-%}': <TokenType.BLOCK_END: 'BLOCK_END'>, '{{+': <TokenType.BLOCK_START: 'BLOCK_START'>, '{{-': <TokenType.BLOCK_START: 'BLOCK_START'>, '+}}': <TokenType.BLOCK_END: 'BLOCK_END'>, '-}}': <TokenType.BLOCK_END: 'BLOCK_END'>, '/*+': <TokenType.HINT: 'HINT'>, '==': <TokenType.EQ: 'EQ'>, '::': <TokenType.DCOLON: 'DCOLON'>, '||': <TokenType.DPIPE: 'DPIPE'>, '>=': <TokenType.GTE: 'GTE'>, '<=': <TokenType.LTE: 'LTE'>, '<>': <TokenType.NEQ: 'NEQ'>, '!=': <TokenType.NEQ: 'NEQ'>, ':=': <TokenType.COLON_EQ: 'COLON_EQ'>, '<=>': <TokenType.NULLSAFE_EQ: 'NULLSAFE_EQ'>, '->': <TokenType.ARROW: 'ARROW'>, '->>': <TokenType.DARROW: 'DARROW'>, '=>': <TokenType.FARROW: 'FARROW'>, '#>': <TokenType.HASH_ARROW: 'HASH_ARROW'>, '#>>': <TokenType.DHASH_ARROW: 'DHASH_ARROW'>, '<->': <TokenType.LR_ARROW: 'LR_ARROW'>, '&&': <TokenType.DAMP: 'DAMP'>, '??': <TokenType.DQMARK: 'DQMARK'>, 'ALL': <TokenType.ALL: 'ALL'>, 'ALWAYS': <TokenType.ALWAYS: 'ALWAYS'>, 'AND': <TokenType.AND: 'AND'>, 'ANTI': <TokenType.ANTI: 'ANTI'>, 'ANY': <TokenType.ANY: 'ANY'>, 'ASC': <TokenType.ASC: 'ASC'>, 'AS': <TokenType.ALIAS: 'ALIAS'>, 'ASOF': <TokenType.ASOF: 'ASOF'>, 'AUTOINCREMENT': <TokenType.AUTO_INCREMENT: 'AUTO_INCREMENT'>, 'AUTO_INCREMENT': <TokenType.AUTO_INCREMENT: 'AUTO_INCREMENT'>, 'BEGIN': <TokenType.COMMAND: 'COMMAND'>, 'BETWEEN': <TokenType.BETWEEN: 'BETWEEN'>, 'CACHE': <TokenType.CACHE: 'CACHE'>, 'UNCACHE': <TokenType.UNCACHE: 'UNCACHE'>, 'CASE': <TokenType.CASE: 'CASE'>, 'CHARACTER SET': <TokenType.CHARACTER_SET: 'CHARACTER_SET'>, 'CLUSTER BY': <TokenType.CLUSTER_BY: 'CLUSTER_BY'>, 'COLLATE': <TokenType.COLLATE: 'COLLATE'>, 'COLUMN': <TokenType.COLUMN: 'COLUMN'>, 'COMMIT': <TokenType.COMMIT: 'COMMIT'>, 'CONNECT BY': <TokenType.CONNECT_BY: 'CONNECT_BY'>, 'CONSTRAINT': <TokenType.CONSTRAINT: 'CONSTRAINT'>, 'CREATE': <TokenType.CREATE: 'CREATE'>, 'CROSS': <TokenType.CROSS: 'CROSS'>, 'CUBE': <TokenType.CUBE: 'CUBE'>, 'CURRENT_DATE': <TokenType.CURRENT_DATE: 'CURRENT_DATE'>, 'CURRENT_TIME': <TokenType.CURRENT_TIME: 'CURRENT_TIME'>, 'CURRENT_TIMESTAMP': <TokenType.CURRENT_TIMESTAMP: 'CURRENT_TIMESTAMP'>, 'CURRENT_USER': <TokenType.CURRENT_USER: 'CURRENT_USER'>, 'DATABASE': <TokenType.DATABASE: 'DATABASE'>, 'DEFAULT': <TokenType.DEFAULT: 'DEFAULT'>, 'DELETE': <TokenType.DELETE: 'DELETE'>, 'DESC': <TokenType.DESC: 'DESC'>, 'DESCRIBE': <TokenType.DESCRIBE: 'DESCRIBE'>, 'DISTINCT': <TokenType.DISTINCT: 'DISTINCT'>, 'DISTRIBUTE BY': <TokenType.DISTRIBUTE_BY: 'DISTRIBUTE_BY'>, 'DROP': <TokenType.DROP: 'DROP'>, 'ELSE': <TokenType.ELSE: 'ELSE'>, 'END': <TokenType.END: 'END'>, 'ESCAPE': <TokenType.ESCAPE: 'ESCAPE'>, 'EXCEPT': <TokenType.EXCEPT: 'EXCEPT'>, 'EXECUTE': <TokenType.EXECUTE: 'EXECUTE'>, 'EXISTS': <TokenType.EXISTS: 'EXISTS'>, 'FALSE': <TokenType.FALSE: 'FALSE'>, 'FETCH': <TokenType.FETCH: 'FETCH'>, 'FILTER': <TokenType.FILTER: 'FILTER'>, 'FIRST': <TokenType.FIRST: 'FIRST'>, 'FULL': <TokenType.FULL: 'FULL'>, 'FUNCTION': <TokenType.FUNCTION: 'FUNCTION'>, 'FOR': <TokenType.FOR: 'FOR'>, 'FOREIGN KEY': <TokenType.FOREIGN_KEY: 'FOREIGN_KEY'>, 'FORMAT': <TokenType.FORMAT: 'FORMAT'>, 'FROM': <TokenType.FROM: 'FROM'>, 'GEOGRAPHY': <TokenType.GEOGRAPHY: 'GEOGRAPHY'>, 'GEOMETRY': <TokenType.GEOMETRY: 'GEOMETRY'>, 'GLOB': <TokenType.GLOB: 'GLOB'>, 'GROUP BY': <TokenType.GROUP_BY: 'GROUP_BY'>, 'GROUPING SETS': <TokenType.GROUPING_SETS: 'GROUPING_SETS'>, 'HAVING': <TokenType.HAVING: 'HAVING'>, 'ILIKE': <TokenType.ILIKE: 'ILIKE'>, 'IN': <TokenType.IN: 'IN'>, 'INDEX': <TokenType.INDEX: 'INDEX'>, 'INET': <TokenType.INET: 'INET'>, 'INNER': <TokenType.INNER: 'INNER'>, 'INSERT': <TokenType.INSERT: 'INSERT'>, 'INTERVAL': <TokenType.INTERVAL: 'INTERVAL'>, 'INTERSECT': <TokenType.INTERSECT: 'INTERSECT'>, 'INTO': <TokenType.INTO: 'INTO'>, 'IS': <TokenType.IS: 'IS'>, 'ISNULL': <TokenType.ISNULL: 'ISNULL'>, 'JOIN': <TokenType.JOIN: 'JOIN'>, 'KEEP': <TokenType.KEEP: 'KEEP'>, 'KILL': <TokenType.KILL: 'KILL'>, 'LATERAL': <TokenType.LATERAL: 'LATERAL'>, 'LEFT': <TokenType.LEFT: 'LEFT'>, 'LIKE': <TokenType.LIKE: 'LIKE'>, 'LIMIT': <TokenType.LIMIT: 'LIMIT'>, 'LOAD': <TokenType.LOAD: 'LOAD'>, 'LOCK': <TokenType.LOCK: 'LOCK'>, 'MERGE': <TokenType.MERGE: 'MERGE'>, 'NATURAL': <TokenType.NATURAL: 'NATURAL'>, 'NEXT': <TokenType.NEXT: 'NEXT'>, 'NOT': <TokenType.NOT: 'NOT'>, 'NOTNULL': <TokenType.NOTNULL: 'NOTNULL'>, 'NULL': <TokenType.NULL: 'NULL'>, 'OBJECT': <TokenType.OBJECT: 'OBJECT'>, 'OFFSET': <TokenType.OFFSET: 'OFFSET'>, 'ON': <TokenType.ON: 'ON'>, 'OR': <TokenType.OR: 'OR'>, 'XOR': <TokenType.XOR: 'XOR'>, 'ORDER BY': <TokenType.ORDER_BY: 'ORDER_BY'>, 'ORDINALITY': <TokenType.ORDINALITY: 'ORDINALITY'>, 'OUTER': <TokenType.OUTER: 'OUTER'>, 'OVER': <TokenType.OVER: 'OVER'>, 'OVERLAPS': <TokenType.OVERLAPS: 'OVERLAPS'>, 'OVERWRITE': <TokenType.OVERWRITE: 'OVERWRITE'>, 'PARTITION': <TokenType.PARTITION: 'PARTITION'>, 'PARTITION BY': <TokenType.PARTITION_BY: 'PARTITION_BY'>, 'PARTITIONED BY': <TokenType.PARTITION_BY: 'PARTITION_BY'>, 'PARTITIONED_BY': <TokenType.PARTITION_BY: 'PARTITION_BY'>, 'PERCENT': <TokenType.PERCENT: 'PERCENT'>, 'PIVOT': <TokenType.PIVOT: 'PIVOT'>, 'PRAGMA': <TokenType.PRAGMA: 'PRAGMA'>, 'PRIMARY KEY': <TokenType.PRIMARY_KEY: 'PRIMARY_KEY'>, 'PROCEDURE': <TokenType.PROCEDURE: 'PROCEDURE'>, 'QUALIFY': <TokenType.QUALIFY: 'QUALIFY'>, 'RANGE': <TokenType.RANGE: 'RANGE'>, 'RECURSIVE': <TokenType.RECURSIVE: 'RECURSIVE'>, 'REGEXP': <TokenType.RLIKE: 'RLIKE'>, 'REPLACE': <TokenType.REPLACE: 'REPLACE'>, 'RETURNING': <TokenType.RETURNING: 'RETURNING'>, 'REFERENCES': <TokenType.REFERENCES: 'REFERENCES'>, 'RIGHT': <TokenType.RIGHT: 'RIGHT'>, 'RLIKE': <TokenType.RLIKE: 'RLIKE'>, 'ROLLBACK': <TokenType.ROLLBACK: 'ROLLBACK'>, 'ROLLUP': <TokenType.ROLLUP: 'ROLLUP'>, 'ROW': <TokenType.ROW: 'ROW'>, 'ROWS': <TokenType.ROWS: 'ROWS'>, 'SCHEMA': <TokenType.SCHEMA: 'SCHEMA'>, 'SELECT': <TokenType.SELECT: 'SELECT'>, 'SEMI': <TokenType.SEMI: 'SEMI'>, 'SET': <TokenType.SET: 'SET'>, 'SETTINGS': <TokenType.SETTINGS: 'SETTINGS'>, 'SHOW': <TokenType.SHOW: 'SHOW'>, 'SIMILAR TO': <TokenType.SIMILAR_TO: 'SIMILAR_TO'>, 'SOME': <TokenType.SOME: 'SOME'>, 'SORT BY': <TokenType.SORT_BY: 'SORT_BY'>, 'START WITH': <TokenType.START_WITH: 'START_WITH'>, 'TABLE': <TokenType.TABLE: 'TABLE'>, 'TABLESAMPLE': <TokenType.TABLE_SAMPLE: 'TABLE_SAMPLE'>, 'TEMP': <TokenType.TEMPORARY: 'TEMPORARY'>, 'TEMPORARY': <TokenType.TEMPORARY: 'TEMPORARY'>, 'THEN': <TokenType.THEN: 'THEN'>, 'TRUE': <TokenType.TRUE: 'TRUE'>, 'UNION': <TokenType.UNION: 'UNION'>, 'UNKNOWN': <TokenType.UNKNOWN: 'UNKNOWN'>, 'UNNEST': <TokenType.UNNEST: 'UNNEST'>, 'UNPIVOT': <TokenType.UNPIVOT: 'UNPIVOT'>, 'UPDATE': <TokenType.UPDATE: 'UPDATE'>, 'USE': <TokenType.USE: 'USE'>, 'USING': <TokenType.USING: 'USING'>, 'UUID': <TokenType.UUID: 'UUID'>, 'VALUES': <TokenType.VALUES: 'VALUES'>, 'VIEW': <TokenType.VIEW: 'VIEW'>, 'VOLATILE': <TokenType.VOLATILE: 'VOLATILE'>, 'WHEN': <TokenType.WHEN: 'WHEN'>, 'WHERE': <TokenType.WHERE: 'WHERE'>, 'WINDOW': <TokenType.WINDOW: 'WINDOW'>, 'WITH': <TokenType.WITH: 'WITH'>, 'APPLY': <TokenType.APPLY: 'APPLY'>, 'ARRAY': <TokenType.ARRAY: 'ARRAY'>, 'BIT': <TokenType.BIT: 'BIT'>, 'BOOL': <TokenType.BOOLEAN: 'BOOLEAN'>, 'BOOLEAN': <TokenType.BOOLEAN: 'BOOLEAN'>, 'BYTE': <TokenType.TINYINT: 'TINYINT'>, 'MEDIUMINT': <TokenType.MEDIUMINT: 'MEDIUMINT'>, 'INT1': <TokenType.TINYINT: 'TINYINT'>, 'TINYINT': <TokenType.TINYINT: 'TINYINT'>, 'INT16': <TokenType.SMALLINT: 'SMALLINT'>, 'SHORT': <TokenType.SMALLINT: 'SMALLINT'>, 'SMALLINT': <TokenType.SMALLINT: 'SMALLINT'>, 'INT128': <TokenType.INT128: 'INT128'>, 'HUGEINT': <TokenType.INT128: 'INT128'>, 'INT2': <TokenType.SMALLINT: 'SMALLINT'>, 'INTEGER': <TokenType.INT: 'INT'>, 'INT': <TokenType.INT: 'INT'>, 'INT4': <TokenType.INT: 'INT'>, 'INT32': <TokenType.INT: 'INT'>, 'INT64': <TokenType.BIGINT: 'BIGINT'>, 'LONG': <TokenType.BIGINT: 'BIGINT'>, 'BIGINT': <TokenType.BIGINT: 'BIGINT'>, 'INT8': <TokenType.TINYINT: 'TINYINT'>, 'DEC': <TokenType.DECIMAL: 'DECIMAL'>, 'DECIMAL': <TokenType.DECIMAL: 'DECIMAL'>, 'BIGDECIMAL': <TokenType.BIGDECIMAL: 'BIGDECIMAL'>, 'BIGNUMERIC': <TokenType.BIGDECIMAL: 'BIGDECIMAL'>, 'MAP': <TokenType.MAP: 'MAP'>, 'NULLABLE': <TokenType.NULLABLE: 'NULLABLE'>, 'NUMBER': <TokenType.DECIMAL: 'DECIMAL'>, 'NUMERIC': <TokenType.DECIMAL: 'DECIMAL'>, 'FIXED': <TokenType.DECIMAL: 'DECIMAL'>, 'REAL': <TokenType.FLOAT: 'FLOAT'>, 'FLOAT': <TokenType.FLOAT: 'FLOAT'>, 'FLOAT4': <TokenType.FLOAT: 'FLOAT'>, 'FLOAT8': <TokenType.DOUBLE: 'DOUBLE'>, 'DOUBLE': <TokenType.DOUBLE: 'DOUBLE'>, 'DOUBLE PRECISION': <TokenType.DOUBLE: 'DOUBLE'>, 'JSON': <TokenType.JSON: 'JSON'>, 'CHAR': <TokenType.CHAR: 'CHAR'>, 'CHARACTER': <TokenType.CHAR: 'CHAR'>, 'NCHAR': <TokenType.NCHAR: 'NCHAR'>, 'VARCHAR': <TokenType.VARCHAR: 'VARCHAR'>, 'VARCHAR2': <TokenType.VARCHAR: 'VARCHAR'>, 'NVARCHAR': <TokenType.NVARCHAR: 'NVARCHAR'>, 'NVARCHAR2': <TokenType.NVARCHAR: 'NVARCHAR'>, 'STR': <TokenType.TEXT: 'TEXT'>, 'STRING': <TokenType.TEXT: 'TEXT'>, 'TEXT': <TokenType.TEXT: 'TEXT'>, 'LONGTEXT': <TokenType.LONGTEXT: 'LONGTEXT'>, 'MEDIUMTEXT': <TokenType.MEDIUMTEXT: 'MEDIUMTEXT'>, 'TINYTEXT': <TokenType.TINYTEXT: 'TINYTEXT'>, 'CLOB': <TokenType.TEXT: 'TEXT'>, 'LONGVARCHAR': <TokenType.TEXT: 'TEXT'>, 'BINARY': <TokenType.BINARY: 'BINARY'>, 'BLOB': <TokenType.VARBINARY: 'VARBINARY'>, 'LONGBLOB': <TokenType.LONGBLOB: 'LONGBLOB'>, 'MEDIUMBLOB': <TokenType.MEDIUMBLOB: 'MEDIUMBLOB'>, 'TINYBLOB': <TokenType.TINYBLOB: 'TINYBLOB'>, 'BYTEA': <TokenType.VARBINARY: 'VARBINARY'>, 'VARBINARY': <TokenType.VARBINARY: 'VARBINARY'>, 'TIME': <TokenType.TIME: 'TIME'>, 'TIMETZ': <TokenType.TIMETZ: 'TIMETZ'>, 'TIMESTAMP': <TokenType.TIMESTAMPTZ: 'TIMESTAMPTZ'>, 'TIMESTAMPTZ': <TokenType.TIMESTAMPTZ: 'TIMESTAMPTZ'>, 'TIMESTAMPLTZ': <TokenType.TIMESTAMPLTZ: 'TIMESTAMPLTZ'>, 'DATE': <TokenType.DATE: 'DATE'>, 'DATETIME': <TokenType.DATETIME: 'DATETIME'>, 'INT4RANGE': <TokenType.INT4RANGE: 'INT4RANGE'>, 'INT4MULTIRANGE': <TokenType.INT4MULTIRANGE: 'INT4MULTIRANGE'>, 'INT8RANGE': <TokenType.INT8RANGE: 'INT8RANGE'>, 'INT8MULTIRANGE': <TokenType.INT8MULTIRANGE: 'INT8MULTIRANGE'>, 'NUMRANGE': <TokenType.NUMRANGE: 'NUMRANGE'>, 'NUMMULTIRANGE': <TokenType.NUMMULTIRANGE: 'NUMMULTIRANGE'>, 'TSRANGE': <TokenType.TSRANGE: 'TSRANGE'>, 'TSMULTIRANGE': <TokenType.TSMULTIRANGE: 'TSMULTIRANGE'>, 'TSTZRANGE': <TokenType.TSTZRANGE: 'TSTZRANGE'>, 'TSTZMULTIRANGE': <TokenType.TSTZMULTIRANGE: 'TSTZMULTIRANGE'>, 'DATERANGE': <TokenType.DATERANGE: 'DATERANGE'>, 'DATEMULTIRANGE': <TokenType.DATEMULTIRANGE: 'DATEMULTIRANGE'>, 'UNIQUE': <TokenType.UNIQUE: 'UNIQUE'>, 'STRUCT': <TokenType.STRUCT: 'STRUCT'>, 'VARIANT': <TokenType.VARIANT: 'VARIANT'>, 'ALTER': <TokenType.ALTER: 'ALTER'>, 'ANALYZE': <TokenType.COMMAND: 'COMMAND'>, 'CALL': <TokenType.COMMAND: 'COMMAND'>, 'COMMENT': <TokenType.COMMENT: 'COMMENT'>, 'COPY': <TokenType.COMMAND: 'COMMAND'>, 'EXPLAIN': <TokenType.COMMAND: 'COMMAND'>, 'GRANT': <TokenType.COMMAND: 'COMMAND'>, 'OPTIMIZE': <TokenType.COMMAND: 'COMMAND'>, 'PREPARE': <TokenType.COMMAND: 'COMMAND'>, 'TRUNCATE': <TokenType.COMMAND: 'COMMAND'>, 'VACUUM': <TokenType.COMMAND: 'COMMAND'>, 'USER-DEFINED': <TokenType.USERDEFINED: 'USERDEFINED'>, 'FOR VERSION': <TokenType.VERSION_SNAPSHOT: 'VERSION_SNAPSHOT'>, 'FOR TIMESTAMP': <TokenType.TIMESTAMP_SNAPSHOT: 'TIMESTAMP_SNAPSHOT'>, 'ANY TYPE': <TokenType.VARIANT: 'VARIANT'>, 'BEGIN TRANSACTION': <TokenType.BEGIN: 'BEGIN'>, 'BYTES': <TokenType.BINARY: 'BINARY'>, 'CURRENT_DATETIME': <TokenType.CURRENT_DATETIME: 'CURRENT_DATETIME'>, 'DECLARE': <TokenType.COMMAND: 'COMMAND'>, 'FLOAT64': <TokenType.DOUBLE: 'DOUBLE'>, 'FOR SYSTEM_TIME': <TokenType.TIMESTAMP_SNAPSHOT: 'TIMESTAMP_SNAPSHOT'>, 'MODEL': <TokenType.MODEL: 'MODEL'>, 'NOT DETERMINISTIC': <TokenType.VOLATILE: 'VOLATILE'>, 'RECORD': <TokenType.STRUCT: 'STRUCT'>}
class BigQuery.Parser(sqlglot.parser.Parser):
325    class Parser(parser.Parser):
326        PREFIXED_PIVOT_COLUMNS = True
327
328        LOG_DEFAULTS_TO_LN = True
329
330        FUNCTIONS = {
331            **parser.Parser.FUNCTIONS,
332            "DATE": _parse_date,
333            "DATE_ADD": parse_date_delta_with_interval(exp.DateAdd),
334            "DATE_SUB": parse_date_delta_with_interval(exp.DateSub),
335            "DATE_TRUNC": lambda args: exp.DateTrunc(
336                unit=exp.Literal.string(str(seq_get(args, 1))),
337                this=seq_get(args, 0),
338            ),
339            "DATETIME_ADD": parse_date_delta_with_interval(exp.DatetimeAdd),
340            "DATETIME_SUB": parse_date_delta_with_interval(exp.DatetimeSub),
341            "DIV": binary_from_function(exp.IntDiv),
342            "FORMAT_DATE": lambda args: exp.TimeToStr(
343                this=exp.TsOrDsToDate(this=seq_get(args, 1)), format=seq_get(args, 0)
344            ),
345            "GENERATE_ARRAY": exp.GenerateSeries.from_arg_list,
346            "JSON_EXTRACT_SCALAR": lambda args: exp.JSONExtractScalar(
347                this=seq_get(args, 0), expression=seq_get(args, 1) or exp.Literal.string("$")
348            ),
349            "MD5": exp.MD5Digest.from_arg_list,
350            "TO_HEX": _parse_to_hex,
351            "PARSE_DATE": lambda args: format_time_lambda(exp.StrToDate, "bigquery")(
352                [seq_get(args, 1), seq_get(args, 0)]
353            ),
354            "PARSE_TIMESTAMP": _parse_timestamp,
355            "REGEXP_CONTAINS": exp.RegexpLike.from_arg_list,
356            "REGEXP_EXTRACT": lambda args: exp.RegexpExtract(
357                this=seq_get(args, 0),
358                expression=seq_get(args, 1),
359                position=seq_get(args, 2),
360                occurrence=seq_get(args, 3),
361                group=exp.Literal.number(1) if re.compile(args[1].name).groups == 1 else None,
362            ),
363            "SHA256": lambda args: exp.SHA2(this=seq_get(args, 0), length=exp.Literal.number(256)),
364            "SHA512": lambda args: exp.SHA2(this=seq_get(args, 0), length=exp.Literal.number(512)),
365            "SPLIT": lambda args: exp.Split(
366                # https://cloud.google.com/bigquery/docs/reference/standard-sql/string_functions#split
367                this=seq_get(args, 0),
368                expression=seq_get(args, 1) or exp.Literal.string(","),
369            ),
370            "TIME": _parse_time,
371            "TIME_ADD": parse_date_delta_with_interval(exp.TimeAdd),
372            "TIME_SUB": parse_date_delta_with_interval(exp.TimeSub),
373            "TIMESTAMP_ADD": parse_date_delta_with_interval(exp.TimestampAdd),
374            "TIMESTAMP_SUB": parse_date_delta_with_interval(exp.TimestampSub),
375            "TIMESTAMP_MICROS": lambda args: exp.UnixToTime(
376                this=seq_get(args, 0), scale=exp.UnixToTime.MICROS
377            ),
378            "TIMESTAMP_MILLIS": lambda args: exp.UnixToTime(
379                this=seq_get(args, 0), scale=exp.UnixToTime.MILLIS
380            ),
381            "TIMESTAMP_SECONDS": lambda args: exp.UnixToTime(
382                this=seq_get(args, 0), scale=exp.UnixToTime.SECONDS
383            ),
384            "TO_JSON_STRING": exp.JSONFormat.from_arg_list,
385        }
386
387        FUNCTION_PARSERS = {
388            **parser.Parser.FUNCTION_PARSERS,
389            "ARRAY": lambda self: self.expression(exp.Array, expressions=[self._parse_statement()]),
390        }
391        FUNCTION_PARSERS.pop("TRIM")
392
393        NO_PAREN_FUNCTIONS = {
394            **parser.Parser.NO_PAREN_FUNCTIONS,
395            TokenType.CURRENT_DATETIME: exp.CurrentDatetime,
396        }
397
398        NESTED_TYPE_TOKENS = {
399            *parser.Parser.NESTED_TYPE_TOKENS,
400            TokenType.TABLE,
401        }
402
403        ID_VAR_TOKENS = {
404            *parser.Parser.ID_VAR_TOKENS,
405            TokenType.VALUES,
406        }
407
408        PROPERTY_PARSERS = {
409            **parser.Parser.PROPERTY_PARSERS,
410            "NOT DETERMINISTIC": lambda self: self.expression(
411                exp.StabilityProperty, this=exp.Literal.string("VOLATILE")
412            ),
413            "OPTIONS": lambda self: self._parse_with_property(),
414        }
415
416        CONSTRAINT_PARSERS = {
417            **parser.Parser.CONSTRAINT_PARSERS,
418            "OPTIONS": lambda self: exp.Properties(expressions=self._parse_with_property()),
419        }
420
421        RANGE_PARSERS = parser.Parser.RANGE_PARSERS.copy()
422        RANGE_PARSERS.pop(TokenType.OVERLAPS, None)
423
424        NULL_TOKENS = {TokenType.NULL, TokenType.UNKNOWN}
425
426        STATEMENT_PARSERS = {
427            **parser.Parser.STATEMENT_PARSERS,
428            TokenType.END: lambda self: self._parse_as_command(self._prev),
429            TokenType.FOR: lambda self: self._parse_for_in(),
430        }
431
432        BRACKET_OFFSETS = {
433            "OFFSET": (0, False),
434            "ORDINAL": (1, False),
435            "SAFE_OFFSET": (0, True),
436            "SAFE_ORDINAL": (1, True),
437        }
438
439        def _parse_for_in(self) -> exp.ForIn:
440            this = self._parse_range()
441            self._match_text_seq("DO")
442            return self.expression(exp.ForIn, this=this, expression=self._parse_statement())
443
444        def _parse_table_part(self, schema: bool = False) -> t.Optional[exp.Expression]:
445            this = super()._parse_table_part(schema=schema) or self._parse_number()
446
447            # https://cloud.google.com/bigquery/docs/reference/standard-sql/lexical#table_names
448            if isinstance(this, exp.Identifier):
449                table_name = this.name
450                while self._match(TokenType.DASH, advance=False) and self._next:
451                    self._advance(2)
452                    table_name += f"-{self._prev.text}"
453
454                this = exp.Identifier(this=table_name, quoted=this.args.get("quoted"))
455            elif isinstance(this, exp.Literal):
456                table_name = this.name
457
458                if self._is_connected() and self._parse_var(any_token=True):
459                    table_name += self._prev.text
460
461                this = exp.Identifier(this=table_name, quoted=True)
462
463            return this
464
465        def _parse_table_parts(self, schema: bool = False) -> exp.Table:
466            table = super()._parse_table_parts(schema=schema)
467            if isinstance(table.this, exp.Identifier) and "." in table.name:
468                catalog, db, this, *rest = (
469                    t.cast(t.Optional[exp.Expression], exp.to_identifier(x))
470                    for x in split_num_words(table.name, ".", 3)
471                )
472
473                if rest and this:
474                    this = exp.Dot.build(t.cast(t.List[exp.Expression], [this, *rest]))
475
476                table = exp.Table(this=this, db=db, catalog=catalog)
477
478            return table
479
480        def _parse_json_object(self) -> exp.JSONObject:
481            json_object = super()._parse_json_object()
482            array_kv_pair = seq_get(json_object.expressions, 0)
483
484            # Converts BQ's "signature 2" of JSON_OBJECT into SQLGlot's canonical representation
485            # https://cloud.google.com/bigquery/docs/reference/standard-sql/json_functions#json_object_signature2
486            if (
487                array_kv_pair
488                and isinstance(array_kv_pair.this, exp.Array)
489                and isinstance(array_kv_pair.expression, exp.Array)
490            ):
491                keys = array_kv_pair.this.expressions
492                values = array_kv_pair.expression.expressions
493
494                json_object.set(
495                    "expressions",
496                    [exp.JSONKeyValue(this=k, expression=v) for k, v in zip(keys, values)],
497                )
498
499            return json_object
500
501        def _parse_bracket(self, this: t.Optional[exp.Expression]) -> t.Optional[exp.Expression]:
502            bracket = super()._parse_bracket(this)
503
504            if this is bracket:
505                return bracket
506
507            if isinstance(bracket, exp.Bracket):
508                for expression in bracket.expressions:
509                    name = expression.name.upper()
510
511                    if name not in self.BRACKET_OFFSETS:
512                        break
513
514                    offset, safe = self.BRACKET_OFFSETS[name]
515                    bracket.set("offset", offset)
516                    bracket.set("safe", safe)
517                    expression.replace(expression.expressions[0])
518
519            return bracket

Parser consumes a list of tokens produced by the Tokenizer and produces a parsed syntax tree.

Arguments:
  • error_level: The desired error level. Default: ErrorLevel.IMMEDIATE
  • error_message_context: Determines the amount of context to capture from a query string when displaying the error message (in number of characters). Default: 100
  • max_errors: Maximum number of error messages to include in a raised ParseError. This is only relevant if error_level is ErrorLevel.RAISE. Default: 3
PREFIXED_PIVOT_COLUMNS = True
LOG_DEFAULTS_TO_LN = True
FUNCTIONS = {'ABS': <bound method Func.from_arg_list of <class 'sqlglot.expressions.Abs'>>, 'ANONYMOUS_AGG_FUNC': <bound method Func.from_arg_list of <class 'sqlglot.expressions.AnonymousAggFunc'>>, 'ANY_VALUE': <bound method Func.from_arg_list of <class 'sqlglot.expressions.AnyValue'>>, 'APPROX_DISTINCT': <bound method Func.from_arg_list of <class 'sqlglot.expressions.ApproxDistinct'>>, 'APPROX_COUNT_DISTINCT': <bound method Func.from_arg_list of <class 'sqlglot.expressions.ApproxDistinct'>>, 'APPROX_QUANTILE': <bound method Func.from_arg_list of <class 'sqlglot.expressions.ApproxQuantile'>>, 'APPROX_TOP_K': <bound method Func.from_arg_list of <class 'sqlglot.expressions.ApproxTopK'>>, 'ARG_MAX': <bound method Func.from_arg_list of <class 'sqlglot.expressions.ArgMax'>>, 'ARGMAX': <bound method Func.from_arg_list of <class 'sqlglot.expressions.ArgMax'>>, 'MAX_BY': <bound method Func.from_arg_list of <class 'sqlglot.expressions.ArgMax'>>, 'ARG_MIN': <bound method Func.from_arg_list of <class 'sqlglot.expressions.ArgMin'>>, 'ARGMIN': <bound method Func.from_arg_list of <class 'sqlglot.expressions.ArgMin'>>, 'MIN_BY': <bound method Func.from_arg_list of <class 'sqlglot.expressions.ArgMin'>>, 'ARRAY': <bound method Func.from_arg_list of <class 'sqlglot.expressions.Array'>>, 'ARRAY_AGG': <bound method Func.from_arg_list of <class 'sqlglot.expressions.ArrayAgg'>>, 'ARRAY_ALL': <bound method Func.from_arg_list of <class 'sqlglot.expressions.ArrayAll'>>, 'ARRAY_ANY': <bound method Func.from_arg_list of <class 'sqlglot.expressions.ArrayAny'>>, 'ARRAY_CONCAT': <bound method Func.from_arg_list of <class 'sqlglot.expressions.ArrayConcat'>>, 'ARRAY_CAT': <bound method Func.from_arg_list of <class 'sqlglot.expressions.ArrayConcat'>>, 'ARRAY_CONTAINS': <bound method Func.from_arg_list of <class 'sqlglot.expressions.ArrayContains'>>, 'FILTER': <bound method Func.from_arg_list of <class 'sqlglot.expressions.ArrayFilter'>>, 'ARRAY_FILTER': <bound method Func.from_arg_list of <class 'sqlglot.expressions.ArrayFilter'>>, 'ARRAY_JOIN': <bound method Func.from_arg_list of <class 'sqlglot.expressions.ArrayJoin'>>, 'ARRAY_SIZE': <bound method Func.from_arg_list of <class 'sqlglot.expressions.ArraySize'>>, 'ARRAY_SORT': <bound method Func.from_arg_list of <class 'sqlglot.expressions.ArraySort'>>, 'ARRAY_SUM': <bound method Func.from_arg_list of <class 'sqlglot.expressions.ArraySum'>>, 'ARRAY_UNION_AGG': <bound method Func.from_arg_list of <class 'sqlglot.expressions.ArrayUnionAgg'>>, 'ARRAY_UNIQUE_AGG': <bound method Func.from_arg_list of <class 'sqlglot.expressions.ArrayUniqueAgg'>>, 'AVG': <bound method Func.from_arg_list of <class 'sqlglot.expressions.Avg'>>, 'CASE': <bound method Func.from_arg_list of <class 'sqlglot.expressions.Case'>>, 'CAST': <bound method Func.from_arg_list of <class 'sqlglot.expressions.Cast'>>, 'CAST_TO_STR_TYPE': <bound method Func.from_arg_list of <class 'sqlglot.expressions.CastToStrType'>>, 'CEIL': <bound method Func.from_arg_list of <class 'sqlglot.expressions.Ceil'>>, 'CEILING': <bound method Func.from_arg_list of <class 'sqlglot.expressions.Ceil'>>, 'CHR': <bound method Func.from_arg_list of <class 'sqlglot.expressions.Chr'>>, 'CHAR': <bound method Func.from_arg_list of <class 'sqlglot.expressions.Chr'>>, 'COALESCE': <bound method Func.from_arg_list of <class 'sqlglot.expressions.Coalesce'>>, 'IFNULL': <bound method Func.from_arg_list of <class 'sqlglot.expressions.Coalesce'>>, 'NVL': <bound method Func.from_arg_list of <class 'sqlglot.expressions.Coalesce'>>, 'COLLATE': <bound method Func.from_arg_list of <class 'sqlglot.expressions.Collate'>>, 'COMBINED_AGG_FUNC': <bound method Func.from_arg_list of <class 'sqlglot.expressions.CombinedAggFunc'>>, 'COMBINED_PARAMETERIZED_AGG': <bound method Func.from_arg_list of <class 'sqlglot.expressions.CombinedParameterizedAgg'>>, 'CONCAT': <function Parser.<lambda>>, 'CONCAT_WS': <function Parser.<lambda>>, 'COUNT': <bound method Func.from_arg_list of <class 'sqlglot.expressions.Count'>>, 'COUNT_IF': <bound method Func.from_arg_list of <class 'sqlglot.expressions.CountIf'>>, 'COUNTIF': <bound method Func.from_arg_list of <class 'sqlglot.expressions.CountIf'>>, 'CURRENT_DATE': <bound method Func.from_arg_list of <class 'sqlglot.expressions.CurrentDate'>>, 'CURRENT_DATETIME': <bound method Func.from_arg_list of <class 'sqlglot.expressions.CurrentDatetime'>>, 'CURRENT_TIME': <bound method Func.from_arg_list of <class 'sqlglot.expressions.CurrentTime'>>, 'CURRENT_TIMESTAMP': <bound method Func.from_arg_list of <class 'sqlglot.expressions.CurrentTimestamp'>>, 'CURRENT_USER': <bound method Func.from_arg_list of <class 'sqlglot.expressions.CurrentUser'>>, 'DATE': <function _parse_date>, 'DATE_ADD': <function parse_date_delta_with_interval.<locals>.func>, 'DATEDIFF': <bound method Func.from_arg_list of <class 'sqlglot.expressions.DateDiff'>>, 'DATE_DIFF': <bound method Func.from_arg_list of <class 'sqlglot.expressions.DateDiff'>>, 'DATE_FROM_PARTS': <bound method Func.from_arg_list of <class 'sqlglot.expressions.DateFromParts'>>, 'DATEFROMPARTS': <bound method Func.from_arg_list of <class 'sqlglot.expressions.DateFromParts'>>, 'DATE_STR_TO_DATE': <bound method Func.from_arg_list of <class 'sqlglot.expressions.DateStrToDate'>>, 'DATE_SUB': <function parse_date_delta_with_interval.<locals>.func>, 'DATE_TO_DATE_STR': <function Parser.<lambda>>, 'DATE_TO_DI': <bound method Func.from_arg_list of <class 'sqlglot.expressions.DateToDi'>>, 'DATE_TRUNC': <function BigQuery.Parser.<lambda>>, 'DATETIME_ADD': <function parse_date_delta_with_interval.<locals>.func>, 'DATETIME_DIFF': <bound method Func.from_arg_list of <class 'sqlglot.expressions.DatetimeDiff'>>, 'DATETIME_SUB': <function parse_date_delta_with_interval.<locals>.func>, 'DATETIME_TRUNC': <bound method Func.from_arg_list of <class 'sqlglot.expressions.DatetimeTrunc'>>, 'DAY': <bound method Func.from_arg_list of <class 'sqlglot.expressions.Day'>>, 'DAY_OF_MONTH': <bound method Func.from_arg_list of <class 'sqlglot.expressions.DayOfMonth'>>, 'DAYOFMONTH': <bound method Func.from_arg_list of <class 'sqlglot.expressions.DayOfMonth'>>, 'DAY_OF_WEEK': <bound method Func.from_arg_list of <class 'sqlglot.expressions.DayOfWeek'>>, 'DAYOFWEEK': <bound method Func.from_arg_list of <class 'sqlglot.expressions.DayOfWeek'>>, 'DAY_OF_YEAR': <bound method Func.from_arg_list of <class 'sqlglot.expressions.DayOfYear'>>, 'DAYOFYEAR': <bound method Func.from_arg_list of <class 'sqlglot.expressions.DayOfYear'>>, 'DECODE': <bound method Func.from_arg_list of <class 'sqlglot.expressions.Decode'>>, 'DI_TO_DATE': <bound method Func.from_arg_list of <class 'sqlglot.expressions.DiToDate'>>, 'ENCODE': <bound method Func.from_arg_list of <class 'sqlglot.expressions.Encode'>>, 'EXP': <bound method Func.from_arg_list of <class 'sqlglot.expressions.Exp'>>, 'EXPLODE': <bound method Func.from_arg_list of <class 'sqlglot.expressions.Explode'>>, 'EXPLODE_OUTER': <bound method Func.from_arg_list of <class 'sqlglot.expressions.ExplodeOuter'>>, 'EXTRACT': <bound method Func.from_arg_list of <class 'sqlglot.expressions.Extract'>>, 'FIRST': <bound method Func.from_arg_list of <class 'sqlglot.expressions.First'>>, 'FLATTEN': <bound method Func.from_arg_list of <class 'sqlglot.expressions.Flatten'>>, 'FLOOR': <bound method Func.from_arg_list of <class 'sqlglot.expressions.Floor'>>, 'FROM_BASE': <bound method Func.from_arg_list of <class 'sqlglot.expressions.FromBase'>>, 'FROM_BASE64': <bound method Func.from_arg_list of <class 'sqlglot.expressions.FromBase64'>>, 'GENERATE_SERIES': <bound method Func.from_arg_list of <class 'sqlglot.expressions.GenerateSeries'>>, 'GET_PATH': <bound method Func.from_arg_list of <class 'sqlglot.expressions.GetPath'>>, 'GREATEST': <bound method Func.from_arg_list of <class 'sqlglot.expressions.Greatest'>>, 'GROUP_CONCAT': <bound method Func.from_arg_list of <class 'sqlglot.expressions.GroupConcat'>>, 'HEX': <bound method Func.from_arg_list of <class 'sqlglot.expressions.Hex'>>, 'HLL': <bound method Func.from_arg_list of <class 'sqlglot.expressions.Hll'>>, 'IF': <bound method Func.from_arg_list of <class 'sqlglot.expressions.If'>>, 'INITCAP': <bound method Func.from_arg_list of <class 'sqlglot.expressions.Initcap'>>, 'IS_INF': <bound method Func.from_arg_list of <class 'sqlglot.expressions.IsInf'>>, 'ISINF': <bound method Func.from_arg_list of <class 'sqlglot.expressions.IsInf'>>, 'IS_NAN': <bound method Func.from_arg_list of <class 'sqlglot.expressions.IsNan'>>, 'ISNAN': <bound method Func.from_arg_list of <class 'sqlglot.expressions.IsNan'>>, 'J_S_O_N_ARRAY': <bound method Func.from_arg_list of <class 'sqlglot.expressions.JSONArray'>>, 'J_S_O_N_ARRAY_AGG': <bound method Func.from_arg_list of <class 'sqlglot.expressions.JSONArrayAgg'>>, 'JSON_ARRAY_CONTAINS': <bound method Func.from_arg_list of <class 'sqlglot.expressions.JSONArrayContains'>>, 'JSONB_EXTRACT': <bound method Func.from_arg_list of <class 'sqlglot.expressions.JSONBExtract'>>, 'JSONB_EXTRACT_SCALAR': <bound method Func.from_arg_list of <class 'sqlglot.expressions.JSONBExtractScalar'>>, 'JSON_EXTRACT': <bound method Func.from_arg_list of <class 'sqlglot.expressions.JSONExtract'>>, 'JSON_EXTRACT_SCALAR': <function BigQuery.Parser.<lambda>>, 'JSON_FORMAT': <bound method Func.from_arg_list of <class 'sqlglot.expressions.JSONFormat'>>, 'J_S_O_N_OBJECT': <bound method Func.from_arg_list of <class 'sqlglot.expressions.JSONObject'>>, 'J_S_O_N_TABLE': <bound method Func.from_arg_list of <class 'sqlglot.expressions.JSONTable'>>, 'LAST': <bound method Func.from_arg_list of <class 'sqlglot.expressions.Last'>>, 'LAST_DAY': <bound method Func.from_arg_list of <class 'sqlglot.expressions.LastDay'>>, 'LAST_DAY_OF_MONTH': <bound method Func.from_arg_list of <class 'sqlglot.expressions.LastDay'>>, 'LEAST': <bound method Func.from_arg_list of <class 'sqlglot.expressions.Least'>>, 'LEFT': <bound method Func.from_arg_list of <class 'sqlglot.expressions.Left'>>, 'LENGTH': <bound method Func.from_arg_list of <class 'sqlglot.expressions.Length'>>, 'LEN': <bound method Func.from_arg_list of <class 'sqlglot.expressions.Length'>>, 'LEVENSHTEIN': <bound method Func.from_arg_list of <class 'sqlglot.expressions.Levenshtein'>>, 'LN': <bound method Func.from_arg_list of <class 'sqlglot.expressions.Ln'>>, 'LOG': <function parse_logarithm>, 'LOG10': <bound method Func.from_arg_list of <class 'sqlglot.expressions.Log10'>>, 'LOG2': <bound method Func.from_arg_list of <class 'sqlglot.expressions.Log2'>>, 'LOGICAL_AND': <bound method Func.from_arg_list of <class 'sqlglot.expressions.LogicalAnd'>>, 'BOOL_AND': <bound method Func.from_arg_list of <class 'sqlglot.expressions.LogicalAnd'>>, 'BOOLAND_AGG': <bound method Func.from_arg_list of <class 'sqlglot.expressions.LogicalAnd'>>, 'LOGICAL_OR': <bound method Func.from_arg_list of <class 'sqlglot.expressions.LogicalOr'>>, 'BOOL_OR': <bound method Func.from_arg_list of <class 'sqlglot.expressions.LogicalOr'>>, 'BOOLOR_AGG': <bound method Func.from_arg_list of <class 'sqlglot.expressions.LogicalOr'>>, 'LOWER': <bound method Func.from_arg_list of <class 'sqlglot.expressions.Lower'>>, 'LCASE': <bound method Func.from_arg_list of <class 'sqlglot.expressions.Lower'>>, 'MD5': <bound method Func.from_arg_list of <class 'sqlglot.expressions.MD5Digest'>>, 'MD5_DIGEST': <bound method Func.from_arg_list of <class 'sqlglot.expressions.MD5Digest'>>, 'MAP': <bound method Func.from_arg_list of <class 'sqlglot.expressions.Map'>>, 'MAP_FROM_ENTRIES': <bound method Func.from_arg_list of <class 'sqlglot.expressions.MapFromEntries'>>, 'MATCH_AGAINST': <bound method Func.from_arg_list of <class 'sqlglot.expressions.MatchAgainst'>>, 'MAX': <bound method Func.from_arg_list of <class 'sqlglot.expressions.Max'>>, 'MIN': <bound method Func.from_arg_list of <class 'sqlglot.expressions.Min'>>, 'MONTH': <bound method Func.from_arg_list of <class 'sqlglot.expressions.Month'>>, 'MONTHS_BETWEEN': <bound method Func.from_arg_list of <class 'sqlglot.expressions.MonthsBetween'>>, 'NEXT_VALUE_FOR': <bound method Func.from_arg_list of <class 'sqlglot.expressions.NextValueFor'>>, 'NULLIF': <bound method Func.from_arg_list of <class 'sqlglot.expressions.Nullif'>>, 'NUMBER_TO_STR': <bound method Func.from_arg_list of <class 'sqlglot.expressions.NumberToStr'>>, 'NVL2': <bound method Func.from_arg_list of <class 'sqlglot.expressions.Nvl2'>>, 'OPEN_J_S_O_N': <bound method Func.from_arg_list of <class 'sqlglot.expressions.OpenJSON'>>, 'PARAMETERIZED_AGG': <bound method Func.from_arg_list of <class 'sqlglot.expressions.ParameterizedAgg'>>, 'PARSE_JSON': <bound method Func.from_arg_list of <class 'sqlglot.expressions.ParseJSON'>>, 'JSON_PARSE': <bound method Func.from_arg_list of <class 'sqlglot.expressions.ParseJSON'>>, 'PERCENTILE_CONT': <bound method Func.from_arg_list of <class 'sqlglot.expressions.PercentileCont'>>, 'PERCENTILE_DISC': <bound method Func.from_arg_list of <class 'sqlglot.expressions.PercentileDisc'>>, 'POSEXPLODE': <bound method Func.from_arg_list of <class 'sqlglot.expressions.Posexplode'>>, 'POSEXPLODE_OUTER': <bound method Func.from_arg_list of <class 'sqlglot.expressions.PosexplodeOuter'>>, 'POWER': <bound method Func.from_arg_list of <class 'sqlglot.expressions.Pow'>>, 'POW': <bound method Func.from_arg_list of <class 'sqlglot.expressions.Pow'>>, 'PREDICT': <bound method Func.from_arg_list of <class 'sqlglot.expressions.Predict'>>, 'QUANTILE': <bound method Func.from_arg_list of <class 'sqlglot.expressions.Quantile'>>, 'RAND': <bound method Func.from_arg_list of <class 'sqlglot.expressions.Rand'>>, 'RANDOM': <bound method Func.from_arg_list of <class 'sqlglot.expressions.Rand'>>, 'RANDN': <bound method Func.from_arg_list of <class 'sqlglot.expressions.Randn'>>, 'RANGE_N': <bound method Func.from_arg_list of <class 'sqlglot.expressions.RangeN'>>, 'READ_CSV': <bound method Func.from_arg_list of <class 'sqlglot.expressions.ReadCSV'>>, 'REDUCE': <bound method Func.from_arg_list of <class 'sqlglot.expressions.Reduce'>>, 'REGEXP_EXTRACT': <function BigQuery.Parser.<lambda>>, 'REGEXP_I_LIKE': <bound method Func.from_arg_list of <class 'sqlglot.expressions.RegexpILike'>>, 'REGEXP_LIKE': <bound method Func.from_arg_list of <class 'sqlglot.expressions.RegexpLike'>>, 'REGEXP_REPLACE': <bound method Func.from_arg_list of <class 'sqlglot.expressions.RegexpReplace'>>, 'REGEXP_SPLIT': <bound method Func.from_arg_list of <class 'sqlglot.expressions.RegexpSplit'>>, 'REPEAT': <bound method Func.from_arg_list of <class 'sqlglot.expressions.Repeat'>>, 'RIGHT': <bound method Func.from_arg_list of <class 'sqlglot.expressions.Right'>>, 'ROUND': <bound method Func.from_arg_list of <class 'sqlglot.expressions.Round'>>, 'ROW_NUMBER': <bound method Func.from_arg_list of <class 'sqlglot.expressions.RowNumber'>>, 'SHA': <bound method Func.from_arg_list of <class 'sqlglot.expressions.SHA'>>, 'SHA1': <bound method Func.from_arg_list of <class 'sqlglot.expressions.SHA'>>, 'SHA2': <bound method Func.from_arg_list of <class 'sqlglot.expressions.SHA2'>>, 'SAFE_DIVIDE': <bound method Func.from_arg_list of <class 'sqlglot.expressions.SafeDivide'>>, 'SORT_ARRAY': <bound method Func.from_arg_list of <class 'sqlglot.expressions.SortArray'>>, 'SPLIT': <function BigQuery.Parser.<lambda>>, 'SQRT': <bound method Func.from_arg_list of <class 'sqlglot.expressions.Sqrt'>>, 'STANDARD_HASH': <bound method Func.from_arg_list of <class 'sqlglot.expressions.StandardHash'>>, 'STAR_MAP': <bound method Func.from_arg_list of <class 'sqlglot.expressions.StarMap'>>, 'STARTS_WITH': <bound method Func.from_arg_list of <class 'sqlglot.expressions.StartsWith'>>, 'STARTSWITH': <bound method Func.from_arg_list of <class 'sqlglot.expressions.StartsWith'>>, 'STDDEV': <bound method Func.from_arg_list of <class 'sqlglot.expressions.Stddev'>>, 'STDDEV_POP': <bound method Func.from_arg_list of <class 'sqlglot.expressions.StddevPop'>>, 'STDDEV_SAMP': <bound method Func.from_arg_list of <class 'sqlglot.expressions.StddevSamp'>>, 'STR_POSITION': <bound method Func.from_arg_list of <class 'sqlglot.expressions.StrPosition'>>, 'STR_TO_DATE': <bound method Func.from_arg_list of <class 'sqlglot.expressions.StrToDate'>>, 'STR_TO_MAP': <bound method Func.from_arg_list of <class 'sqlglot.expressions.StrToMap'>>, 'STR_TO_TIME': <bound method Func.from_arg_list of <class 'sqlglot.expressions.StrToTime'>>, 'STR_TO_UNIX': <bound method Func.from_arg_list of <class 'sqlglot.expressions.StrToUnix'>>, 'STRUCT': <bound method Func.from_arg_list of <class 'sqlglot.expressions.Struct'>>, 'STRUCT_EXTRACT': <bound method Func.from_arg_list of <class 'sqlglot.expressions.StructExtract'>>, 'STUFF': <bound method Func.from_arg_list of <class 'sqlglot.expressions.Stuff'>>, 'INSERT': <bound method Func.from_arg_list of <class 'sqlglot.expressions.Stuff'>>, 'SUBSTRING': <bound method Func.from_arg_list of <class 'sqlglot.expressions.Substring'>>, 'SUM': <bound method Func.from_arg_list of <class 'sqlglot.expressions.Sum'>>, 'TIME_ADD': <function parse_date_delta_with_interval.<locals>.func>, 'TIME_DIFF': <bound method Func.from_arg_list of <class 'sqlglot.expressions.TimeDiff'>>, 'TIME_FROM_PARTS': <bound method Func.from_arg_list of <class 'sqlglot.expressions.TimeFromParts'>>, 'TIMEFROMPARTS': <bound method Func.from_arg_list of <class 'sqlglot.expressions.TimeFromParts'>>, 'TIME_STR_TO_DATE': <bound method Func.from_arg_list of <class 'sqlglot.expressions.TimeStrToDate'>>, 'TIME_STR_TO_TIME': <bound method Func.from_arg_list of <class 'sqlglot.expressions.TimeStrToTime'>>, 'TIME_STR_TO_UNIX': <bound method Func.from_arg_list of <class 'sqlglot.expressions.TimeStrToUnix'>>, 'TIME_SUB': <function parse_date_delta_with_interval.<locals>.func>, 'TIME_TO_STR': <bound method Func.from_arg_list of <class 'sqlglot.expressions.TimeToStr'>>, 'TIME_TO_TIME_STR': <function Parser.<lambda>>, 'TIME_TO_UNIX': <bound method Func.from_arg_list of <class 'sqlglot.expressions.TimeToUnix'>>, 'TIME_TRUNC': <bound method Func.from_arg_list of <class 'sqlglot.expressions.TimeTrunc'>>, 'TIMESTAMP': <bound method Func.from_arg_list of <class 'sqlglot.expressions.Timestamp'>>, 'TIMESTAMP_ADD': <function parse_date_delta_with_interval.<locals>.func>, 'TIMESTAMP_DIFF': <bound method Func.from_arg_list of <class 'sqlglot.expressions.TimestampDiff'>>, 'TIMESTAMP_FROM_PARTS': <bound method Func.from_arg_list of <class 'sqlglot.expressions.TimestampFromParts'>>, 'TIMESTAMPFROMPARTS': <bound method Func.from_arg_list of <class 'sqlglot.expressions.TimestampFromParts'>>, 'TIMESTAMP_SUB': <function parse_date_delta_with_interval.<locals>.func>, 'TIMESTAMP_TRUNC': <bound method Func.from_arg_list of <class 'sqlglot.expressions.TimestampTrunc'>>, 'TO_ARRAY': <bound method Func.from_arg_list of <class 'sqlglot.expressions.ToArray'>>, 'TO_BASE64': <bound method Func.from_arg_list of <class 'sqlglot.expressions.ToBase64'>>, 'TO_CHAR': <bound method Func.from_arg_list of <class 'sqlglot.expressions.ToChar'>>, 'TO_DAYS': <bound method Func.from_arg_list of <class 'sqlglot.expressions.ToDays'>>, 'TRANSFORM': <bound method Func.from_arg_list of <class 'sqlglot.expressions.Transform'>>, 'TRIM': <bound method Func.from_arg_list of <class 'sqlglot.expressions.Trim'>>, 'TRY_CAST': <bound method Func.from_arg_list of <class 'sqlglot.expressions.TryCast'>>, 'TS_OR_DI_TO_DI': <bound method Func.from_arg_list of <class 'sqlglot.expressions.TsOrDiToDi'>>, 'TS_OR_DS_ADD': <bound method Func.from_arg_list of <class 'sqlglot.expressions.TsOrDsAdd'>>, 'TS_OR_DS_DIFF': <bound method Func.from_arg_list of <class 'sqlglot.expressions.TsOrDsDiff'>>, 'TS_OR_DS_TO_DATE': <bound method Func.from_arg_list of <class 'sqlglot.expressions.TsOrDsToDate'>>, 'TS_OR_DS_TO_DATE_STR': <function Parser.<lambda>>, 'TS_OR_DS_TO_TIME': <bound method Func.from_arg_list of <class 'sqlglot.expressions.TsOrDsToTime'>>, 'UNHEX': <bound method Func.from_arg_list of <class 'sqlglot.expressions.Unhex'>>, 'UNIX_DATE': <bound method Func.from_arg_list of <class 'sqlglot.expressions.UnixDate'>>, 'UNIX_TO_STR': <bound method Func.from_arg_list of <class 'sqlglot.expressions.UnixToStr'>>, 'UNIX_TO_TIME': <bound method Func.from_arg_list of <class 'sqlglot.expressions.UnixToTime'>>, 'UNIX_TO_TIME_STR': <bound method Func.from_arg_list of <class 'sqlglot.expressions.UnixToTimeStr'>>, 'UPPER': <bound method Func.from_arg_list of <class 'sqlglot.expressions.Upper'>>, 'UCASE': <bound method Func.from_arg_list of <class 'sqlglot.expressions.Upper'>>, 'VAR_MAP': <function parse_var_map>, 'VARIANCE': <bound method Func.from_arg_list of <class 'sqlglot.expressions.Variance'>>, 'VARIANCE_SAMP': <bound method Func.from_arg_list of <class 'sqlglot.expressions.Variance'>>, 'VAR_SAMP': <bound method Func.from_arg_list of <class 'sqlglot.expressions.Variance'>>, 'VARIANCE_POP': <bound method Func.from_arg_list of <class 'sqlglot.expressions.VariancePop'>>, 'VAR_POP': <bound method Func.from_arg_list of <class 'sqlglot.expressions.VariancePop'>>, 'WEEK': <bound method Func.from_arg_list of <class 'sqlglot.expressions.Week'>>, 'WEEK_OF_YEAR': <bound method Func.from_arg_list of <class 'sqlglot.expressions.WeekOfYear'>>, 'WEEKOFYEAR': <bound method Func.from_arg_list of <class 'sqlglot.expressions.WeekOfYear'>>, 'WHEN': <bound method Func.from_arg_list of <class 'sqlglot.expressions.When'>>, 'X_M_L_TABLE': <bound method Func.from_arg_list of <class 'sqlglot.expressions.XMLTable'>>, 'XOR': <bound method Func.from_arg_list of <class 'sqlglot.expressions.Xor'>>, 'YEAR': <bound method Func.from_arg_list of <class 'sqlglot.expressions.Year'>>, 'GLOB': <function Parser.<lambda>>, 'LIKE': <function parse_like>, 'DIV': <function binary_from_function.<locals>.<lambda>>, 'FORMAT_DATE': <function BigQuery.Parser.<lambda>>, 'GENERATE_ARRAY': <bound method Func.from_arg_list of <class 'sqlglot.expressions.GenerateSeries'>>, 'TO_HEX': <function _parse_to_hex>, 'PARSE_DATE': <function BigQuery.Parser.<lambda>>, 'PARSE_TIMESTAMP': <function _parse_timestamp>, 'REGEXP_CONTAINS': <bound method Func.from_arg_list of <class 'sqlglot.expressions.RegexpLike'>>, 'SHA256': <function BigQuery.Parser.<lambda>>, 'SHA512': <function BigQuery.Parser.<lambda>>, 'TIME': <function _parse_time>, 'TIMESTAMP_MICROS': <function BigQuery.Parser.<lambda>>, 'TIMESTAMP_MILLIS': <function BigQuery.Parser.<lambda>>, 'TIMESTAMP_SECONDS': <function BigQuery.Parser.<lambda>>, 'TO_JSON_STRING': <bound method Func.from_arg_list of <class 'sqlglot.expressions.JSONFormat'>>}
FUNCTION_PARSERS = {'ANY_VALUE': <function Parser.<lambda>>, 'CAST': <function Parser.<lambda>>, 'CONVERT': <function Parser.<lambda>>, 'DECODE': <function Parser.<lambda>>, 'EXTRACT': <function Parser.<lambda>>, 'JSON_OBJECT': <function Parser.<lambda>>, 'JSON_TABLE': <function Parser.<lambda>>, 'MATCH': <function Parser.<lambda>>, 'OPENJSON': <function Parser.<lambda>>, 'POSITION': <function Parser.<lambda>>, 'PREDICT': <function Parser.<lambda>>, 'SAFE_CAST': <function Parser.<lambda>>, 'STRING_AGG': <function Parser.<lambda>>, 'SUBSTRING': <function Parser.<lambda>>, 'TRY_CAST': <function Parser.<lambda>>, 'TRY_CONVERT': <function Parser.<lambda>>, 'ARRAY': <function BigQuery.Parser.<lambda>>}
NO_PAREN_FUNCTIONS = {<TokenType.CURRENT_DATE: 'CURRENT_DATE'>: <class 'sqlglot.expressions.CurrentDate'>, <TokenType.CURRENT_DATETIME: 'CURRENT_DATETIME'>: <class 'sqlglot.expressions.CurrentDatetime'>, <TokenType.CURRENT_TIME: 'CURRENT_TIME'>: <class 'sqlglot.expressions.CurrentTime'>, <TokenType.CURRENT_TIMESTAMP: 'CURRENT_TIMESTAMP'>: <class 'sqlglot.expressions.CurrentTimestamp'>, <TokenType.CURRENT_USER: 'CURRENT_USER'>: <class 'sqlglot.expressions.CurrentUser'>}
NESTED_TYPE_TOKENS = {<TokenType.ARRAY: 'ARRAY'>, <TokenType.NULLABLE: 'NULLABLE'>, <TokenType.MAP: 'MAP'>, <TokenType.STRUCT: 'STRUCT'>, <TokenType.TABLE: 'TABLE'>, <TokenType.NESTED: 'NESTED'>, <TokenType.LOWCARDINALITY: 'LOWCARDINALITY'>}
ID_VAR_TOKENS = {<TokenType.OBJECT_IDENTIFIER: 'OBJECT_IDENTIFIER'>, <TokenType.INT256: 'INT256'>, <TokenType.TEXT: 'TEXT'>, <TokenType.INT8RANGE: 'INT8RANGE'>, <TokenType.CONSTRAINT: 'CONSTRAINT'>, <TokenType.SHOW: 'SHOW'>, <TokenType.MODEL: 'MODEL'>, <TokenType.CURRENT_DATETIME: 'CURRENT_DATETIME'>, <TokenType.SCHEMA: 'SCHEMA'>, <TokenType.TSTZMULTIRANGE: 'TSTZMULTIRANGE'>, <TokenType.ROWS: 'ROWS'>, <TokenType.MEDIUMBLOB: 'MEDIUMBLOB'>, <TokenType.ALL: 'ALL'>, <TokenType.BEGIN: 'BEGIN'>, <TokenType.BINARY: 'BINARY'>, <TokenType.TIMESTAMP_NS: 'TIMESTAMP_NS'>, <TokenType.DECIMAL: 'DECIMAL'>, <TokenType.DELETE: 'DELETE'>, <TokenType.CURRENT_TIMESTAMP: 'CURRENT_TIMESTAMP'>, <TokenType.UINT128: 'UINT128'>, <TokenType.OPERATOR: 'OPERATOR'>, <TokenType.AUTO_INCREMENT: 'AUTO_INCREMENT'>, <TokenType.LONGTEXT: 'LONGTEXT'>, <TokenType.YEAR: 'YEAR'>, <TokenType.COLLATE: 'COLLATE'>, <TokenType.NESTED: 'NESTED'>, <TokenType.BIT: 'BIT'>, <TokenType.FIRST: 'FIRST'>, <TokenType.ANY: 'ANY'>, <TokenType.DATETIME: 'DATETIME'>, <TokenType.NEXT: 'NEXT'>, <TokenType.INT8MULTIRANGE: 'INT8MULTIRANGE'>, <TokenType.LOAD: 'LOAD'>, <TokenType.UNIQUE: 'UNIQUE'>, <TokenType.BIGSERIAL: 'BIGSERIAL'>, <TokenType.UTINYINT: 'UTINYINT'>, <TokenType.UPDATE: 'UPDATE'>, <TokenType.NULLABLE: 'NULLABLE'>, <TokenType.RIGHT: 'RIGHT'>, <TokenType.JSONB: 'JSONB'>, <TokenType.OFFSET: 'OFFSET'>, <TokenType.SMALLSERIAL: 'SMALLSERIAL'>, <TokenType.ISNULL: 'ISNULL'>, <TokenType.TSRANGE: 'TSRANGE'>, <TokenType.FIXEDSTRING: 'FIXEDSTRING'>, <TokenType.UDECIMAL: 'UDECIMAL'>, <TokenType.VARCHAR: 'VARCHAR'>, <TokenType.DOUBLE: 'DOUBLE'>, <TokenType.CASE: 'CASE'>, <TokenType.FOREIGN_KEY: 'FOREIGN_KEY'>, <TokenType.CURRENT_TIME: 'CURRENT_TIME'>, <TokenType.FALSE: 'FALSE'>, <TokenType.KILL: 'KILL'>, <TokenType.VARBINARY: 'VARBINARY'>, <TokenType.UUID: 'UUID'>, <TokenType.ORDINALITY: 'ORDINALITY'>, <TokenType.TIMESTAMP_MS: 'TIMESTAMP_MS'>, <TokenType.HLLSKETCH: 'HLLSKETCH'>, <TokenType.TIMETZ: 'TIMETZ'>, <TokenType.SET: 'SET'>, <TokenType.DATABASE: 'DATABASE'>, <TokenType.INET: 'INET'>, <TokenType.PSEUDO_TYPE: 'PSEUDO_TYPE'>, <TokenType.INT: 'INT'>, <TokenType.ROW: 'ROW'>, <TokenType.DATE: 'DATE'>, <TokenType.EXECUTE: 'EXECUTE'>, <TokenType.COMMENT: 'COMMENT'>, <TokenType.COMMIT: 'COMMIT'>, <TokenType.MEDIUMTEXT: 'MEDIUMTEXT'>, <TokenType.DIV: 'DIV'>, <TokenType.ROWVERSION: 'ROWVERSION'>, <TokenType.UINT: 'UINT'>, <TokenType.BIGDECIMAL: 'BIGDECIMAL'>, <TokenType.MERGE: 'MERGE'>, <TokenType.JSON: 'JSON'>, <TokenType.KEEP: 'KEEP'>, <TokenType.PARTITION: 'PARTITION'>, <TokenType.TINYINT: 'TINYINT'>, <TokenType.IPADDRESS: 'IPADDRESS'>, <TokenType.CURRENT_DATE: 'CURRENT_DATE'>, <TokenType.USMALLINT: 'USMALLINT'>, <TokenType.USERDEFINED: 'USERDEFINED'>, <TokenType.CACHE: 'CACHE'>, <TokenType.VOLATILE: 'VOLATILE'>, <TokenType.FINAL: 'FINAL'>, <TokenType.XML: 'XML'>, <TokenType.VARIANT: 'VARIANT'>, <TokenType.OVERLAPS: 'OVERLAPS'>, <TokenType.ENUM: 'ENUM'>, <TokenType.PROCEDURE: 'PROCEDURE'>, <TokenType.OVERWRITE: 'OVERWRITE'>, <TokenType.VAR: 'VAR'>, <TokenType.NUMMULTIRANGE: 'NUMMULTIRANGE'>, <TokenType.EXISTS: 'EXISTS'>, <TokenType.UNKNOWN: 'UNKNOWN'>, <TokenType.ANTI: 'ANTI'>, <TokenType.UBIGINT: 'UBIGINT'>, <TokenType.USE: 'USE'>, <TokenType.HSTORE: 'HSTORE'>, <TokenType.IPPREFIX: 'IPPREFIX'>, <TokenType.MEDIUMINT: 'MEDIUMINT'>, <TokenType.DATERANGE: 'DATERANGE'>, <TokenType.TIMESTAMP_S: 'TIMESTAMP_S'>, <TokenType.SOME: 'SOME'>, <TokenType.LOWCARDINALITY: 'LOWCARDINALITY'>, <TokenType.TIMESTAMPLTZ: 'TIMESTAMPLTZ'>, <TokenType.SEMI: 'SEMI'>, <TokenType.SMALLINT: 'SMALLINT'>, <TokenType.NATURAL: 'NATURAL'>, <TokenType.TABLE: 'TABLE'>, <TokenType.FULL: 'FULL'>, <TokenType.REFERENCES: 'REFERENCES'>, <TokenType.FILTER: 'FILTER'>, <TokenType.PRAGMA: 'PRAGMA'>, <TokenType.DICTIONARY: 'DICTIONARY'>, <TokenType.SERIAL: 'SERIAL'>, <TokenType.ASC: 'ASC'>, <TokenType.FUNCTION: 'FUNCTION'>, <TokenType.VALUES: 'VALUES'>, <TokenType.APPLY: 'APPLY'>, <TokenType.MONEY: 'MONEY'>, <TokenType.INT128: 'INT128'>, <TokenType.UINT256: 'UINT256'>, <TokenType.PERCENT: 'PERCENT'>, <TokenType.IS: 'IS'>, <TokenType.PIVOT: 'PIVOT'>, <TokenType.GEOMETRY: 'GEOMETRY'>, <TokenType.ARRAY: 'ARRAY'>, <TokenType.END: 'END'>, <TokenType.TINYBLOB: 'TINYBLOB'>, <TokenType.LONGBLOB: 'LONGBLOB'>, <TokenType.NULL: 'NULL'>, <TokenType.MAP: 'MAP'>, <TokenType.TOP: 'TOP'>, <TokenType.VIEW: 'VIEW'>, <TokenType.COMMAND: 'COMMAND'>, <TokenType.LEFT: 'LEFT'>, <TokenType.DESCRIBE: 'DESCRIBE'>, <TokenType.INT4RANGE: 'INT4RANGE'>, <TokenType.NVARCHAR: 'NVARCHAR'>, <TokenType.IMAGE: 'IMAGE'>, <TokenType.DEFAULT: 'DEFAULT'>, <TokenType.RANGE: 'RANGE'>, <TokenType.CURRENT_USER: 'CURRENT_USER'>, <TokenType.UNIQUEIDENTIFIER: 'UNIQUEIDENTIFIER'>, <TokenType.TIMESTAMP: 'TIMESTAMP'>, <TokenType.TRUE: 'TRUE'>, <TokenType.OBJECT: 'OBJECT'>, <TokenType.RECURSIVE: 'RECURSIVE'>, <TokenType.WINDOW: 'WINDOW'>, <TokenType.SMALLMONEY: 'SMALLMONEY'>, <TokenType.ENUM16: 'ENUM16'>, <TokenType.REPLACE: 'REPLACE'>, <TokenType.INT4MULTIRANGE: 'INT4MULTIRANGE'>, <TokenType.BIGINT: 'BIGINT'>, <TokenType.TSMULTIRANGE: 'TSMULTIRANGE'>, <TokenType.TINYTEXT: 'TINYTEXT'>, <TokenType.UMEDIUMINT: 'UMEDIUMINT'>, <TokenType.TSTZRANGE: 'TSTZRANGE'>, <TokenType.UNPIVOT: 'UNPIVOT'>, <TokenType.TEMPORARY: 'TEMPORARY'>, <TokenType.COLUMN: 'COLUMN'>, <TokenType.TIMESTAMPTZ: 'TIMESTAMPTZ'>, <TokenType.DESC: 'DESC'>, <TokenType.ESCAPE: 'ESCAPE'>, <TokenType.INDEX: 'INDEX'>, <TokenType.INTERVAL: 'INTERVAL'>, <TokenType.DATEMULTIRANGE: 'DATEMULTIRANGE'>, <TokenType.NUMRANGE: 'NUMRANGE'>, <TokenType.REFRESH: 'REFRESH'>, <TokenType.TIME: 'TIME'>, <TokenType.CHAR: 'CHAR'>, <TokenType.SETTINGS: 'SETTINGS'>, <TokenType.ENUM8: 'ENUM8'>, <TokenType.SUPER: 'SUPER'>, <TokenType.FLOAT: 'FLOAT'>, <TokenType.FORMAT: 'FORMAT'>, <TokenType.NCHAR: 'NCHAR'>, <TokenType.DATETIME64: 'DATETIME64'>, <TokenType.BOOLEAN: 'BOOLEAN'>, <TokenType.STRUCT: 'STRUCT'>, <TokenType.GEOGRAPHY: 'GEOGRAPHY'>}
PROPERTY_PARSERS = {'ALGORITHM': <function Parser.<lambda>>, 'AUTO': <function Parser.<lambda>>, 'AUTO_INCREMENT': <function Parser.<lambda>>, 'BLOCKCOMPRESSION': <function Parser.<lambda>>, 'CHARSET': <function Parser.<lambda>>, 'CHARACTER SET': <function Parser.<lambda>>, 'CHECKSUM': <function Parser.<lambda>>, 'CLUSTER BY': <function Parser.<lambda>>, 'CLUSTERED': <function Parser.<lambda>>, 'COLLATE': <function Parser.<lambda>>, 'COMMENT': <function Parser.<lambda>>, 'CONTAINS': <function Parser.<lambda>>, 'COPY': <function Parser.<lambda>>, 'DATABLOCKSIZE': <function Parser.<lambda>>, 'DEFINER': <function Parser.<lambda>>, 'DETERMINISTIC': <function Parser.<lambda>>, 'DISTKEY': <function Parser.<lambda>>, 'DISTSTYLE': <function Parser.<lambda>>, 'ENGINE': <function Parser.<lambda>>, 'EXECUTE': <function Parser.<lambda>>, 'EXTERNAL': <function Parser.<lambda>>, 'FALLBACK': <function Parser.<lambda>>, 'FORMAT': <function Parser.<lambda>>, 'FREESPACE': <function Parser.<lambda>>, 'HEAP': <function Parser.<lambda>>, 'IMMUTABLE': <function Parser.<lambda>>, 'INPUT': <function Parser.<lambda>>, 'JOURNAL': <function Parser.<lambda>>, 'LANGUAGE': <function Parser.<lambda>>, 'LAYOUT': <function Parser.<lambda>>, 'LIFETIME': <function Parser.<lambda>>, 'LIKE': <function Parser.<lambda>>, 'LOCATION': <function Parser.<lambda>>, 'LOCK': <function Parser.<lambda>>, 'LOCKING': <function Parser.<lambda>>, 'LOG': <function Parser.<lambda>>, 'MATERIALIZED': <function Parser.<lambda>>, 'MERGEBLOCKRATIO': <function Parser.<lambda>>, 'MODIFIES': <function Parser.<lambda>>, 'MULTISET': <function Parser.<lambda>>, 'NO': <function Parser.<lambda>>, 'ON': <function Parser.<lambda>>, 'ORDER BY': <function Parser.<lambda>>, 'OUTPUT': <function Parser.<lambda>>, 'PARTITION': <function Parser.<lambda>>, 'PARTITION BY': <function Parser.<lambda>>, 'PARTITIONED BY': <function Parser.<lambda>>, 'PARTITIONED_BY': <function Parser.<lambda>>, 'PRIMARY KEY': <function Parser.<lambda>>, 'RANGE': <function Parser.<lambda>>, 'READS': <function Parser.<lambda>>, 'REMOTE': <function Parser.<lambda>>, 'RETURNS': <function Parser.<lambda>>, 'ROW': <function Parser.<lambda>>, 'ROW_FORMAT': <function Parser.<lambda>>, 'SAMPLE': <function Parser.<lambda>>, 'SET': <function Parser.<lambda>>, 'SETTINGS': <function Parser.<lambda>>, 'SORTKEY': <function Parser.<lambda>>, 'SOURCE': <function Parser.<lambda>>, 'STABLE': <function Parser.<lambda>>, 'STORED': <function Parser.<lambda>>, 'SYSTEM_VERSIONING': <function Parser.<lambda>>, 'TBLPROPERTIES': <function Parser.<lambda>>, 'TEMP': <function Parser.<lambda>>, 'TEMPORARY': <function Parser.<lambda>>, 'TO': <function Parser.<lambda>>, 'TRANSIENT': <function Parser.<lambda>>, 'TRANSFORM': <function Parser.<lambda>>, 'TTL': <function Parser.<lambda>>, 'USING': <function Parser.<lambda>>, 'VOLATILE': <function Parser.<lambda>>, 'WITH': <function Parser.<lambda>>, 'NOT DETERMINISTIC': <function BigQuery.Parser.<lambda>>, 'OPTIONS': <function BigQuery.Parser.<lambda>>}
CONSTRAINT_PARSERS = {'AUTOINCREMENT': <function Parser.<lambda>>, 'AUTO_INCREMENT': <function Parser.<lambda>>, 'CASESPECIFIC': <function Parser.<lambda>>, 'CHARACTER SET': <function Parser.<lambda>>, 'CHECK': <function Parser.<lambda>>, 'COLLATE': <function Parser.<lambda>>, 'COMMENT': <function Parser.<lambda>>, 'COMPRESS': <function Parser.<lambda>>, 'CLUSTERED': <function Parser.<lambda>>, 'NONCLUSTERED': <function Parser.<lambda>>, 'DEFAULT': <function Parser.<lambda>>, 'ENCODE': <function Parser.<lambda>>, 'FOREIGN KEY': <function Parser.<lambda>>, 'FORMAT': <function Parser.<lambda>>, 'GENERATED': <function Parser.<lambda>>, 'IDENTITY': <function Parser.<lambda>>, 'INLINE': <function Parser.<lambda>>, 'LIKE': <function Parser.<lambda>>, 'NOT': <function Parser.<lambda>>, 'NULL': <function Parser.<lambda>>, 'ON': <function Parser.<lambda>>, 'PATH': <function Parser.<lambda>>, 'PERIOD': <function Parser.<lambda>>, 'PRIMARY KEY': <function Parser.<lambda>>, 'REFERENCES': <function Parser.<lambda>>, 'TITLE': <function Parser.<lambda>>, 'TTL': <function Parser.<lambda>>, 'UNIQUE': <function Parser.<lambda>>, 'UPPERCASE': <function Parser.<lambda>>, 'WITH': <function Parser.<lambda>>, 'OPTIONS': <function BigQuery.Parser.<lambda>>}
RANGE_PARSERS = {<TokenType.BETWEEN: 'BETWEEN'>: <function Parser.<lambda>>, <TokenType.GLOB: 'GLOB'>: <function binary_range_parser.<locals>.<lambda>>, <TokenType.ILIKE: 'ILIKE'>: <function binary_range_parser.<locals>.<lambda>>, <TokenType.IN: 'IN'>: <function Parser.<lambda>>, <TokenType.IRLIKE: 'IRLIKE'>: <function binary_range_parser.<locals>.<lambda>>, <TokenType.IS: 'IS'>: <function Parser.<lambda>>, <TokenType.LIKE: 'LIKE'>: <function binary_range_parser.<locals>.<lambda>>, <TokenType.RLIKE: 'RLIKE'>: <function binary_range_parser.<locals>.<lambda>>, <TokenType.SIMILAR_TO: 'SIMILAR_TO'>: <function binary_range_parser.<locals>.<lambda>>, <TokenType.FOR: 'FOR'>: <function Parser.<lambda>>}
NULL_TOKENS = {<TokenType.NULL: 'NULL'>, <TokenType.UNKNOWN: 'UNKNOWN'>}
STATEMENT_PARSERS = {<TokenType.ALTER: 'ALTER'>: <function Parser.<lambda>>, <TokenType.BEGIN: 'BEGIN'>: <function Parser.<lambda>>, <TokenType.CACHE: 'CACHE'>: <function Parser.<lambda>>, <TokenType.COMMIT: 'COMMIT'>: <function Parser.<lambda>>, <TokenType.COMMENT: 'COMMENT'>: <function Parser.<lambda>>, <TokenType.CREATE: 'CREATE'>: <function Parser.<lambda>>, <TokenType.DELETE: 'DELETE'>: <function Parser.<lambda>>, <TokenType.DESC: 'DESC'>: <function Parser.<lambda>>, <TokenType.DESCRIBE: 'DESCRIBE'>: <function Parser.<lambda>>, <TokenType.DROP: 'DROP'>: <function Parser.<lambda>>, <TokenType.INSERT: 'INSERT'>: <function Parser.<lambda>>, <TokenType.KILL: 'KILL'>: <function Parser.<lambda>>, <TokenType.LOAD: 'LOAD'>: <function Parser.<lambda>>, <TokenType.MERGE: 'MERGE'>: <function Parser.<lambda>>, <TokenType.PIVOT: 'PIVOT'>: <function Parser.<lambda>>, <TokenType.PRAGMA: 'PRAGMA'>: <function Parser.<lambda>>, <TokenType.REFRESH: 'REFRESH'>: <function Parser.<lambda>>, <TokenType.ROLLBACK: 'ROLLBACK'>: <function Parser.<lambda>>, <TokenType.SET: 'SET'>: <function Parser.<lambda>>, <TokenType.UNCACHE: 'UNCACHE'>: <function Parser.<lambda>>, <TokenType.UPDATE: 'UPDATE'>: <function Parser.<lambda>>, <TokenType.USE: 'USE'>: <function Parser.<lambda>>, <TokenType.END: 'END'>: <function BigQuery.Parser.<lambda>>, <TokenType.FOR: 'FOR'>: <function BigQuery.Parser.<lambda>>}
BRACKET_OFFSETS = {'OFFSET': (0, False), 'ORDINAL': (1, False), 'SAFE_OFFSET': (0, True), 'SAFE_ORDINAL': (1, True)}
TABLE_ALIAS_TOKENS = {<TokenType.OBJECT_IDENTIFIER: 'OBJECT_IDENTIFIER'>, <TokenType.INT256: 'INT256'>, <TokenType.TEXT: 'TEXT'>, <TokenType.INT8RANGE: 'INT8RANGE'>, <TokenType.CONSTRAINT: 'CONSTRAINT'>, <TokenType.SHOW: 'SHOW'>, <TokenType.MODEL: 'MODEL'>, <TokenType.CURRENT_DATETIME: 'CURRENT_DATETIME'>, <TokenType.SCHEMA: 'SCHEMA'>, <TokenType.TSTZMULTIRANGE: 'TSTZMULTIRANGE'>, <TokenType.ROWS: 'ROWS'>, <TokenType.MEDIUMBLOB: 'MEDIUMBLOB'>, <TokenType.ALL: 'ALL'>, <TokenType.BEGIN: 'BEGIN'>, <TokenType.BINARY: 'BINARY'>, <TokenType.TIMESTAMP_NS: 'TIMESTAMP_NS'>, <TokenType.DECIMAL: 'DECIMAL'>, <TokenType.DELETE: 'DELETE'>, <TokenType.CURRENT_TIMESTAMP: 'CURRENT_TIMESTAMP'>, <TokenType.UINT128: 'UINT128'>, <TokenType.OPERATOR: 'OPERATOR'>, <TokenType.AUTO_INCREMENT: 'AUTO_INCREMENT'>, <TokenType.LONGTEXT: 'LONGTEXT'>, <TokenType.YEAR: 'YEAR'>, <TokenType.COLLATE: 'COLLATE'>, <TokenType.NESTED: 'NESTED'>, <TokenType.BIT: 'BIT'>, <TokenType.FIRST: 'FIRST'>, <TokenType.ANY: 'ANY'>, <TokenType.DATETIME: 'DATETIME'>, <TokenType.NEXT: 'NEXT'>, <TokenType.INT8MULTIRANGE: 'INT8MULTIRANGE'>, <TokenType.LOAD: 'LOAD'>, <TokenType.UNIQUE: 'UNIQUE'>, <TokenType.BIGSERIAL: 'BIGSERIAL'>, <TokenType.UTINYINT: 'UTINYINT'>, <TokenType.UPDATE: 'UPDATE'>, <TokenType.NULLABLE: 'NULLABLE'>, <TokenType.JSONB: 'JSONB'>, <TokenType.SMALLSERIAL: 'SMALLSERIAL'>, <TokenType.ISNULL: 'ISNULL'>, <TokenType.TSRANGE: 'TSRANGE'>, <TokenType.FIXEDSTRING: 'FIXEDSTRING'>, <TokenType.UDECIMAL: 'UDECIMAL'>, <TokenType.VARCHAR: 'VARCHAR'>, <TokenType.DOUBLE: 'DOUBLE'>, <TokenType.CASE: 'CASE'>, <TokenType.FOREIGN_KEY: 'FOREIGN_KEY'>, <TokenType.CURRENT_TIME: 'CURRENT_TIME'>, <TokenType.FALSE: 'FALSE'>, <TokenType.KILL: 'KILL'>, <TokenType.VARBINARY: 'VARBINARY'>, <TokenType.UUID: 'UUID'>, <TokenType.ORDINALITY: 'ORDINALITY'>, <TokenType.TIMESTAMP_MS: 'TIMESTAMP_MS'>, <TokenType.HLLSKETCH: 'HLLSKETCH'>, <TokenType.TIMETZ: 'TIMETZ'>, <TokenType.SET: 'SET'>, <TokenType.DATABASE: 'DATABASE'>, <TokenType.INET: 'INET'>, <TokenType.PSEUDO_TYPE: 'PSEUDO_TYPE'>, <TokenType.INT: 'INT'>, <TokenType.ROW: 'ROW'>, <TokenType.DATE: 'DATE'>, <TokenType.EXECUTE: 'EXECUTE'>, <TokenType.COMMENT: 'COMMENT'>, <TokenType.COMMIT: 'COMMIT'>, <TokenType.MEDIUMTEXT: 'MEDIUMTEXT'>, <TokenType.DIV: 'DIV'>, <TokenType.ROWVERSION: 'ROWVERSION'>, <TokenType.UINT: 'UINT'>, <TokenType.BIGDECIMAL: 'BIGDECIMAL'>, <TokenType.MERGE: 'MERGE'>, <TokenType.JSON: 'JSON'>, <TokenType.KEEP: 'KEEP'>, <TokenType.PARTITION: 'PARTITION'>, <TokenType.TINYINT: 'TINYINT'>, <TokenType.IPADDRESS: 'IPADDRESS'>, <TokenType.CURRENT_DATE: 'CURRENT_DATE'>, <TokenType.USMALLINT: 'USMALLINT'>, <TokenType.USERDEFINED: 'USERDEFINED'>, <TokenType.CACHE: 'CACHE'>, <TokenType.VOLATILE: 'VOLATILE'>, <TokenType.FINAL: 'FINAL'>, <TokenType.XML: 'XML'>, <TokenType.VARIANT: 'VARIANT'>, <TokenType.OVERLAPS: 'OVERLAPS'>, <TokenType.ENUM: 'ENUM'>, <TokenType.PROCEDURE: 'PROCEDURE'>, <TokenType.OVERWRITE: 'OVERWRITE'>, <TokenType.VAR: 'VAR'>, <TokenType.NUMMULTIRANGE: 'NUMMULTIRANGE'>, <TokenType.EXISTS: 'EXISTS'>, <TokenType.UNKNOWN: 'UNKNOWN'>, <TokenType.ANTI: 'ANTI'>, <TokenType.UBIGINT: 'UBIGINT'>, <TokenType.USE: 'USE'>, <TokenType.HSTORE: 'HSTORE'>, <TokenType.IPPREFIX: 'IPPREFIX'>, <TokenType.MEDIUMINT: 'MEDIUMINT'>, <TokenType.DATERANGE: 'DATERANGE'>, <TokenType.TIMESTAMP_S: 'TIMESTAMP_S'>, <TokenType.SOME: 'SOME'>, <TokenType.LOWCARDINALITY: 'LOWCARDINALITY'>, <TokenType.TIMESTAMPLTZ: 'TIMESTAMPLTZ'>, <TokenType.SEMI: 'SEMI'>, <TokenType.SMALLINT: 'SMALLINT'>, <TokenType.TABLE: 'TABLE'>, <TokenType.REFERENCES: 'REFERENCES'>, <TokenType.FILTER: 'FILTER'>, <TokenType.PRAGMA: 'PRAGMA'>, <TokenType.DICTIONARY: 'DICTIONARY'>, <TokenType.SERIAL: 'SERIAL'>, <TokenType.ASC: 'ASC'>, <TokenType.FUNCTION: 'FUNCTION'>, <TokenType.MONEY: 'MONEY'>, <TokenType.INT128: 'INT128'>, <TokenType.UINT256: 'UINT256'>, <TokenType.PERCENT: 'PERCENT'>, <TokenType.IS: 'IS'>, <TokenType.PIVOT: 'PIVOT'>, <TokenType.GEOMETRY: 'GEOMETRY'>, <TokenType.ARRAY: 'ARRAY'>, <TokenType.END: 'END'>, <TokenType.TINYBLOB: 'TINYBLOB'>, <TokenType.LONGBLOB: 'LONGBLOB'>, <TokenType.NULL: 'NULL'>, <TokenType.MAP: 'MAP'>, <TokenType.TOP: 'TOP'>, <TokenType.VIEW: 'VIEW'>, <TokenType.COMMAND: 'COMMAND'>, <TokenType.DESCRIBE: 'DESCRIBE'>, <TokenType.INT4RANGE: 'INT4RANGE'>, <TokenType.NVARCHAR: 'NVARCHAR'>, <TokenType.IMAGE: 'IMAGE'>, <TokenType.DEFAULT: 'DEFAULT'>, <TokenType.RANGE: 'RANGE'>, <TokenType.CURRENT_USER: 'CURRENT_USER'>, <TokenType.UNIQUEIDENTIFIER: 'UNIQUEIDENTIFIER'>, <TokenType.TIMESTAMP: 'TIMESTAMP'>, <TokenType.TRUE: 'TRUE'>, <TokenType.OBJECT: 'OBJECT'>, <TokenType.RECURSIVE: 'RECURSIVE'>, <TokenType.SMALLMONEY: 'SMALLMONEY'>, <TokenType.ENUM16: 'ENUM16'>, <TokenType.REPLACE: 'REPLACE'>, <TokenType.INT4MULTIRANGE: 'INT4MULTIRANGE'>, <TokenType.BIGINT: 'BIGINT'>, <TokenType.TSMULTIRANGE: 'TSMULTIRANGE'>, <TokenType.TINYTEXT: 'TINYTEXT'>, <TokenType.UMEDIUMINT: 'UMEDIUMINT'>, <TokenType.TSTZRANGE: 'TSTZRANGE'>, <TokenType.UNPIVOT: 'UNPIVOT'>, <TokenType.TEMPORARY: 'TEMPORARY'>, <TokenType.COLUMN: 'COLUMN'>, <TokenType.TIMESTAMPTZ: 'TIMESTAMPTZ'>, <TokenType.DESC: 'DESC'>, <TokenType.ESCAPE: 'ESCAPE'>, <TokenType.INDEX: 'INDEX'>, <TokenType.INTERVAL: 'INTERVAL'>, <TokenType.DATEMULTIRANGE: 'DATEMULTIRANGE'>, <TokenType.NUMRANGE: 'NUMRANGE'>, <TokenType.REFRESH: 'REFRESH'>, <TokenType.TIME: 'TIME'>, <TokenType.CHAR: 'CHAR'>, <TokenType.SETTINGS: 'SETTINGS'>, <TokenType.ENUM8: 'ENUM8'>, <TokenType.SUPER: 'SUPER'>, <TokenType.FLOAT: 'FLOAT'>, <TokenType.FORMAT: 'FORMAT'>, <TokenType.NCHAR: 'NCHAR'>, <TokenType.DATETIME64: 'DATETIME64'>, <TokenType.BOOLEAN: 'BOOLEAN'>, <TokenType.STRUCT: 'STRUCT'>, <TokenType.GEOGRAPHY: 'GEOGRAPHY'>}
SHOW_TRIE: Dict = {}
SET_TRIE: Dict = {'GLOBAL': {0: True}, 'LOCAL': {0: True}, 'SESSION': {0: True}, 'TRANSACTION': {0: True}}
class BigQuery.Generator(sqlglot.generator.Generator):
521    class Generator(generator.Generator):
522        EXPLICIT_UNION = True
523        INTERVAL_ALLOWS_PLURAL_FORM = False
524        JOIN_HINTS = False
525        QUERY_HINTS = False
526        TABLE_HINTS = False
527        LIMIT_FETCH = "LIMIT"
528        RENAME_TABLE_WITH_DB = False
529        NVL2_SUPPORTED = False
530        UNNEST_WITH_ORDINALITY = False
531        COLLATE_IS_FUNC = True
532        LIMIT_ONLY_LITERALS = True
533        SUPPORTS_TABLE_ALIAS_COLUMNS = False
534        UNPIVOT_ALIASES_ARE_IDENTIFIERS = False
535
536        TRANSFORMS = {
537            **generator.Generator.TRANSFORMS,
538            exp.ApproxDistinct: rename_func("APPROX_COUNT_DISTINCT"),
539            exp.ArgMax: arg_max_or_min_no_count("MAX_BY"),
540            exp.ArgMin: arg_max_or_min_no_count("MIN_BY"),
541            exp.ArrayContains: _array_contains_sql,
542            exp.ArraySize: rename_func("ARRAY_LENGTH"),
543            exp.Cast: transforms.preprocess([transforms.remove_precision_parameterized_types]),
544            exp.CollateProperty: lambda self, e: f"DEFAULT COLLATE {self.sql(e, 'this')}"
545            if e.args.get("default")
546            else f"COLLATE {self.sql(e, 'this')}",
547            exp.CountIf: rename_func("COUNTIF"),
548            exp.Create: _create_sql,
549            exp.CTE: transforms.preprocess([_pushdown_cte_column_names]),
550            exp.DateAdd: date_add_interval_sql("DATE", "ADD"),
551            exp.DateDiff: lambda self, e: f"DATE_DIFF({self.sql(e, 'this')}, {self.sql(e, 'expression')}, {self.sql(e.args.get('unit', 'DAY'))})",
552            exp.DateFromParts: rename_func("DATE"),
553            exp.DateStrToDate: datestrtodate_sql,
554            exp.DateSub: date_add_interval_sql("DATE", "SUB"),
555            exp.DatetimeAdd: date_add_interval_sql("DATETIME", "ADD"),
556            exp.DatetimeSub: date_add_interval_sql("DATETIME", "SUB"),
557            exp.DateTrunc: lambda self, e: self.func("DATE_TRUNC", e.this, e.text("unit")),
558            exp.GenerateSeries: rename_func("GENERATE_ARRAY"),
559            exp.GetPath: path_to_jsonpath(),
560            exp.GroupConcat: rename_func("STRING_AGG"),
561            exp.Hex: rename_func("TO_HEX"),
562            exp.If: if_sql(false_value="NULL"),
563            exp.ILike: no_ilike_sql,
564            exp.IntDiv: rename_func("DIV"),
565            exp.JSONFormat: rename_func("TO_JSON_STRING"),
566            exp.JSONKeyValue: json_keyvalue_comma_sql,
567            exp.Max: max_or_greatest,
568            exp.MD5: lambda self, e: self.func("TO_HEX", self.func("MD5", e.this)),
569            exp.MD5Digest: rename_func("MD5"),
570            exp.Min: min_or_least,
571            exp.PartitionedByProperty: lambda self, e: f"PARTITION BY {self.sql(e, 'this')}",
572            exp.RegexpExtract: lambda self, e: self.func(
573                "REGEXP_EXTRACT",
574                e.this,
575                e.expression,
576                e.args.get("position"),
577                e.args.get("occurrence"),
578            ),
579            exp.RegexpReplace: regexp_replace_sql,
580            exp.RegexpLike: rename_func("REGEXP_CONTAINS"),
581            exp.ReturnsProperty: _returnsproperty_sql,
582            exp.Select: transforms.preprocess(
583                [
584                    transforms.explode_to_unnest(),
585                    _unqualify_unnest,
586                    transforms.eliminate_distinct_on,
587                    _alias_ordered_group,
588                    transforms.eliminate_semi_and_anti_joins,
589                ]
590            ),
591            exp.SHA2: lambda self, e: self.func(
592                f"SHA256" if e.text("length") == "256" else "SHA512", e.this
593            ),
594            exp.StabilityProperty: lambda self, e: f"DETERMINISTIC"
595            if e.name == "IMMUTABLE"
596            else "NOT DETERMINISTIC",
597            exp.StrToDate: lambda self, e: f"PARSE_DATE({self.format_time(e)}, {self.sql(e, 'this')})",
598            exp.StrToTime: lambda self, e: self.func(
599                "PARSE_TIMESTAMP", self.format_time(e), e.this, e.args.get("zone")
600            ),
601            exp.TimeAdd: date_add_interval_sql("TIME", "ADD"),
602            exp.TimeFromParts: rename_func("TIME"),
603            exp.TimeSub: date_add_interval_sql("TIME", "SUB"),
604            exp.TimestampAdd: date_add_interval_sql("TIMESTAMP", "ADD"),
605            exp.TimestampSub: date_add_interval_sql("TIMESTAMP", "SUB"),
606            exp.TimeStrToTime: timestrtotime_sql,
607            exp.Trim: lambda self, e: self.func(f"TRIM", e.this, e.expression),
608            exp.TsOrDsAdd: _ts_or_ds_add_sql,
609            exp.TsOrDsDiff: _ts_or_ds_diff_sql,
610            exp.TsOrDsToTime: rename_func("TIME"),
611            exp.Unhex: rename_func("FROM_HEX"),
612            exp.UnixDate: rename_func("UNIX_DATE"),
613            exp.UnixToTime: _unix_to_time_sql,
614            exp.Values: _derived_table_values_to_unnest,
615            exp.VariancePop: rename_func("VAR_POP"),
616        }
617
618        TYPE_MAPPING = {
619            **generator.Generator.TYPE_MAPPING,
620            exp.DataType.Type.BIGDECIMAL: "BIGNUMERIC",
621            exp.DataType.Type.BIGINT: "INT64",
622            exp.DataType.Type.BINARY: "BYTES",
623            exp.DataType.Type.BOOLEAN: "BOOL",
624            exp.DataType.Type.CHAR: "STRING",
625            exp.DataType.Type.DECIMAL: "NUMERIC",
626            exp.DataType.Type.DOUBLE: "FLOAT64",
627            exp.DataType.Type.FLOAT: "FLOAT64",
628            exp.DataType.Type.INT: "INT64",
629            exp.DataType.Type.NCHAR: "STRING",
630            exp.DataType.Type.NVARCHAR: "STRING",
631            exp.DataType.Type.SMALLINT: "INT64",
632            exp.DataType.Type.TEXT: "STRING",
633            exp.DataType.Type.TIMESTAMP: "DATETIME",
634            exp.DataType.Type.TIMESTAMPTZ: "TIMESTAMP",
635            exp.DataType.Type.TIMESTAMPLTZ: "TIMESTAMP",
636            exp.DataType.Type.TINYINT: "INT64",
637            exp.DataType.Type.VARBINARY: "BYTES",
638            exp.DataType.Type.VARCHAR: "STRING",
639            exp.DataType.Type.VARIANT: "ANY TYPE",
640        }
641
642        PROPERTIES_LOCATION = {
643            **generator.Generator.PROPERTIES_LOCATION,
644            exp.PartitionedByProperty: exp.Properties.Location.POST_SCHEMA,
645            exp.VolatileProperty: exp.Properties.Location.UNSUPPORTED,
646        }
647
648        # from: https://cloud.google.com/bigquery/docs/reference/standard-sql/lexical#reserved_keywords
649        RESERVED_KEYWORDS = {
650            *generator.Generator.RESERVED_KEYWORDS,
651            "all",
652            "and",
653            "any",
654            "array",
655            "as",
656            "asc",
657            "assert_rows_modified",
658            "at",
659            "between",
660            "by",
661            "case",
662            "cast",
663            "collate",
664            "contains",
665            "create",
666            "cross",
667            "cube",
668            "current",
669            "default",
670            "define",
671            "desc",
672            "distinct",
673            "else",
674            "end",
675            "enum",
676            "escape",
677            "except",
678            "exclude",
679            "exists",
680            "extract",
681            "false",
682            "fetch",
683            "following",
684            "for",
685            "from",
686            "full",
687            "group",
688            "grouping",
689            "groups",
690            "hash",
691            "having",
692            "if",
693            "ignore",
694            "in",
695            "inner",
696            "intersect",
697            "interval",
698            "into",
699            "is",
700            "join",
701            "lateral",
702            "left",
703            "like",
704            "limit",
705            "lookup",
706            "merge",
707            "natural",
708            "new",
709            "no",
710            "not",
711            "null",
712            "nulls",
713            "of",
714            "on",
715            "or",
716            "order",
717            "outer",
718            "over",
719            "partition",
720            "preceding",
721            "proto",
722            "qualify",
723            "range",
724            "recursive",
725            "respect",
726            "right",
727            "rollup",
728            "rows",
729            "select",
730            "set",
731            "some",
732            "struct",
733            "tablesample",
734            "then",
735            "to",
736            "treat",
737            "true",
738            "unbounded",
739            "union",
740            "unnest",
741            "using",
742            "when",
743            "where",
744            "window",
745            "with",
746            "within",
747        }
748
749        def timetostr_sql(self, expression: exp.TimeToStr) -> str:
750            if isinstance(expression.this, exp.TsOrDsToDate):
751                this: exp.Expression = expression.this
752            else:
753                this = expression
754
755            return f"FORMAT_DATE({self.format_time(expression)}, {self.sql(this, 'this')})"
756
757        def struct_sql(self, expression: exp.Struct) -> str:
758            args = []
759            for expr in expression.expressions:
760                if isinstance(expr, self.KEY_VALUE_DEFINITIONS):
761                    arg = f"{self.sql(expr, 'expression')} AS {expr.this.name}"
762                else:
763                    arg = self.sql(expr)
764
765                args.append(arg)
766
767            return self.func("STRUCT", *args)
768
769        def eq_sql(self, expression: exp.EQ) -> str:
770            # Operands of = cannot be NULL in BigQuery
771            if isinstance(expression.left, exp.Null) or isinstance(expression.right, exp.Null):
772                if not isinstance(expression.parent, exp.Update):
773                    return "NULL"
774
775            return self.binary(expression, "=")
776
777        def attimezone_sql(self, expression: exp.AtTimeZone) -> str:
778            parent = expression.parent
779
780            # BigQuery allows CAST(.. AS {STRING|TIMESTAMP} [FORMAT <fmt> [AT TIME ZONE <tz>]]).
781            # Only the TIMESTAMP one should use the below conversion, when AT TIME ZONE is included.
782            if not isinstance(parent, exp.Cast) or not parent.to.is_type("text"):
783                return self.func(
784                    "TIMESTAMP", self.func("DATETIME", expression.this, expression.args.get("zone"))
785                )
786
787            return super().attimezone_sql(expression)
788
789        def trycast_sql(self, expression: exp.TryCast) -> str:
790            return self.cast_sql(expression, safe_prefix="SAFE_")
791
792        def cte_sql(self, expression: exp.CTE) -> str:
793            if expression.alias_column_names:
794                self.unsupported("Column names in CTE definition are not supported.")
795            return super().cte_sql(expression)
796
797        def array_sql(self, expression: exp.Array) -> str:
798            first_arg = seq_get(expression.expressions, 0)
799            if isinstance(first_arg, exp.Subqueryable):
800                return f"ARRAY{self.wrap(self.sql(first_arg))}"
801
802            return inline_array_sql(self, expression)
803
804        def bracket_sql(self, expression: exp.Bracket) -> str:
805            this = self.sql(expression, "this")
806            expressions = expression.expressions
807
808            if len(expressions) == 1:
809                arg = expressions[0]
810                if arg.type is None:
811                    from sqlglot.optimizer.annotate_types import annotate_types
812
813                    arg = annotate_types(arg)
814
815                if arg.type and arg.type.this in exp.DataType.TEXT_TYPES:
816                    # BQ doesn't support bracket syntax with string values
817                    return f"{this}.{arg.name}"
818
819            expressions_sql = ", ".join(self.sql(e) for e in expressions)
820            offset = expression.args.get("offset")
821
822            if offset == 0:
823                expressions_sql = f"OFFSET({expressions_sql})"
824            elif offset == 1:
825                expressions_sql = f"ORDINAL({expressions_sql})"
826            elif offset is not None:
827                self.unsupported(f"Unsupported array offset: {offset}")
828
829            if expression.args.get("safe"):
830                expressions_sql = f"SAFE_{expressions_sql}"
831
832            return f"{this}[{expressions_sql}]"
833
834        def transaction_sql(self, *_) -> str:
835            return "BEGIN TRANSACTION"
836
837        def commit_sql(self, *_) -> str:
838            return "COMMIT TRANSACTION"
839
840        def rollback_sql(self, *_) -> str:
841            return "ROLLBACK TRANSACTION"
842
843        def in_unnest_op(self, expression: exp.Unnest) -> str:
844            return self.sql(expression)
845
846        def except_op(self, expression: exp.Except) -> str:
847            if not expression.args.get("distinct", False):
848                self.unsupported("EXCEPT without DISTINCT is not supported in BigQuery")
849            return f"EXCEPT{' DISTINCT' if expression.args.get('distinct') else ' ALL'}"
850
851        def intersect_op(self, expression: exp.Intersect) -> str:
852            if not expression.args.get("distinct", False):
853                self.unsupported("INTERSECT without DISTINCT is not supported in BigQuery")
854            return f"INTERSECT{' DISTINCT' if expression.args.get('distinct') else ' ALL'}"
855
856        def with_properties(self, properties: exp.Properties) -> str:
857            return self.properties(properties, prefix=self.seg("OPTIONS"))
858
859        def version_sql(self, expression: exp.Version) -> str:
860            if expression.name == "TIMESTAMP":
861                expression.set("this", "SYSTEM_TIME")
862            return super().version_sql(expression)

Generator converts a given syntax tree to the corresponding SQL string.

Arguments:
  • pretty: Whether or not to format the produced SQL string. Default: False.
  • identify: Determines when an identifier should be quoted. Possible values are: False (default): Never quote, except in cases where it's mandatory by the dialect. True or 'always': Always quote. 'safe': Only quote identifiers that are case insensitive.
  • normalize: Whether or not to normalize identifiers to lowercase. Default: False.
  • pad: Determines the pad size in a formatted string. Default: 2.
  • indent: Determines the indentation size in a formatted string. Default: 2.
  • normalize_functions: Whether or not to normalize all function names. Possible values are: "upper" or True (default): Convert names to uppercase. "lower": Convert names to lowercase. False: Disables function name normalization.
  • unsupported_level: Determines the generator's behavior when it encounters unsupported expressions. Default ErrorLevel.WARN.
  • max_unsupported: Maximum number of unsupported messages to include in a raised UnsupportedError. This is only relevant if unsupported_level is ErrorLevel.RAISE. Default: 3
  • leading_comma: Determines whether or not the comma is leading or trailing in select expressions. This is only relevant when generating in pretty mode. Default: False
  • max_text_width: The max number of characters in a segment before creating new lines in pretty mode. The default is on the smaller end because the length only represents a segment and not the true line length. Default: 80
  • comments: Whether or not to preserve comments in the output SQL code. Default: True
EXPLICIT_UNION = True
INTERVAL_ALLOWS_PLURAL_FORM = False
JOIN_HINTS = False
QUERY_HINTS = False
TABLE_HINTS = False
LIMIT_FETCH = 'LIMIT'
RENAME_TABLE_WITH_DB = False
NVL2_SUPPORTED = False
UNNEST_WITH_ORDINALITY = False
COLLATE_IS_FUNC = True
LIMIT_ONLY_LITERALS = True
SUPPORTS_TABLE_ALIAS_COLUMNS = False
UNPIVOT_ALIASES_ARE_IDENTIFIERS = False
TRANSFORMS = {<class 'sqlglot.expressions.DateAdd'>: <function date_add_interval_sql.<locals>.func>, <class 'sqlglot.expressions.CaseSpecificColumnConstraint'>: <function Generator.<lambda>>, <class 'sqlglot.expressions.CharacterSetColumnConstraint'>: <function Generator.<lambda>>, <class 'sqlglot.expressions.CharacterSetProperty'>: <function Generator.<lambda>>, <class 'sqlglot.expressions.CheckColumnConstraint'>: <function Generator.<lambda>>, <class 'sqlglot.expressions.ClusteredColumnConstraint'>: <function Generator.<lambda>>, <class 'sqlglot.expressions.CollateColumnConstraint'>: <function Generator.<lambda>>, <class 'sqlglot.expressions.AutoRefreshProperty'>: <function Generator.<lambda>>, <class 'sqlglot.expressions.CopyGrantsProperty'>: <function Generator.<lambda>>, <class 'sqlglot.expressions.CommentColumnConstraint'>: <function Generator.<lambda>>, <class 'sqlglot.expressions.DateFormatColumnConstraint'>: <function Generator.<lambda>>, <class 'sqlglot.expressions.DefaultColumnConstraint'>: <function Generator.<lambda>>, <class 'sqlglot.expressions.EncodeColumnConstraint'>: <function Generator.<lambda>>, <class 'sqlglot.expressions.ExecuteAsProperty'>: <function Generator.<lambda>>, <class 'sqlglot.expressions.ExternalProperty'>: <function Generator.<lambda>>, <class 'sqlglot.expressions.HeapProperty'>: <function Generator.<lambda>>, <class 'sqlglot.expressions.InlineLengthColumnConstraint'>: <function Generator.<lambda>>, <class 'sqlglot.expressions.InputModelProperty'>: <function Generator.<lambda>>, <class 'sqlglot.expressions.IntervalSpan'>: <function Generator.<lambda>>, <class 'sqlglot.expressions.LanguageProperty'>: <function Generator.<lambda>>, <class 'sqlglot.expressions.LocationProperty'>: <function Generator.<lambda>>, <class 'sqlglot.expressions.LogProperty'>: <function Generator.<lambda>>, <class 'sqlglot.expressions.MaterializedProperty'>: <function Generator.<lambda>>, <class 'sqlglot.expressions.NoPrimaryIndexProperty'>: <function Generator.<lambda>>, <class 'sqlglot.expressions.NonClusteredColumnConstraint'>: <function Generator.<lambda>>, <class 'sqlglot.expressions.NotForReplicationColumnConstraint'>: <function Generator.<lambda>>, <class 'sqlglot.expressions.OnCommitProperty'>: <function Generator.<lambda>>, <class 'sqlglot.expressions.OnProperty'>: <function Generator.<lambda>>, <class 'sqlglot.expressions.OnUpdateColumnConstraint'>: <function Generator.<lambda>>, <class 'sqlglot.expressions.OutputModelProperty'>: <function Generator.<lambda>>, <class 'sqlglot.expressions.PathColumnConstraint'>: <function Generator.<lambda>>, <class 'sqlglot.expressions.RemoteWithConnectionModelProperty'>: <function Generator.<lambda>>, <class 'sqlglot.expressions.ReturnsProperty'>: <function _returnsproperty_sql>, <class 'sqlglot.expressions.SampleProperty'>: <function Generator.<lambda>>, <class 'sqlglot.expressions.SetProperty'>: <function Generator.<lambda>>, <class 'sqlglot.expressions.SettingsProperty'>: <function Generator.<lambda>>, <class 'sqlglot.expressions.SqlReadWriteProperty'>: <function Generator.<lambda>>, <class 'sqlglot.expressions.SqlSecurityProperty'>: <function Generator.<lambda>>, <class 'sqlglot.expressions.StabilityProperty'>: <function BigQuery.Generator.<lambda>>, <class 'sqlglot.expressions.TemporaryProperty'>: <function Generator.<lambda>>, <class 'sqlglot.expressions.ToTableProperty'>: <function Generator.<lambda>>, <class 'sqlglot.expressions.TransientProperty'>: <function Generator.<lambda>>, <class 'sqlglot.expressions.TransformModelProperty'>: <function Generator.<lambda>>, <class 'sqlglot.expressions.TitleColumnConstraint'>: <function Generator.<lambda>>, <class 'sqlglot.expressions.UppercaseColumnConstraint'>: <function Generator.<lambda>>, <class 'sqlglot.expressions.VarMap'>: <function Generator.<lambda>>, <class 'sqlglot.expressions.VolatileProperty'>: <function Generator.<lambda>>, <class 'sqlglot.expressions.WithJournalTableProperty'>: <function Generator.<lambda>>, <class 'sqlglot.expressions.ApproxDistinct'>: <function rename_func.<locals>.<lambda>>, <class 'sqlglot.expressions.ArgMax'>: <function arg_max_or_min_no_count.<locals>._arg_max_or_min_sql>, <class 'sqlglot.expressions.ArgMin'>: <function arg_max_or_min_no_count.<locals>._arg_max_or_min_sql>, <class 'sqlglot.expressions.ArrayContains'>: <function _array_contains_sql>, <class 'sqlglot.expressions.ArraySize'>: <function rename_func.<locals>.<lambda>>, <class 'sqlglot.expressions.Cast'>: <function preprocess.<locals>._to_sql>, <class 'sqlglot.expressions.CollateProperty'>: <function BigQuery.Generator.<lambda>>, <class 'sqlglot.expressions.CountIf'>: <function rename_func.<locals>.<lambda>>, <class 'sqlglot.expressions.Create'>: <function _create_sql>, <class 'sqlglot.expressions.CTE'>: <function preprocess.<locals>._to_sql>, <class 'sqlglot.expressions.DateDiff'>: <function BigQuery.Generator.<lambda>>, <class 'sqlglot.expressions.DateFromParts'>: <function rename_func.<locals>.<lambda>>, <class 'sqlglot.expressions.DateStrToDate'>: <function datestrtodate_sql>, <class 'sqlglot.expressions.DateSub'>: <function date_add_interval_sql.<locals>.func>, <class 'sqlglot.expressions.DatetimeAdd'>: <function date_add_interval_sql.<locals>.func>, <class 'sqlglot.expressions.DatetimeSub'>: <function date_add_interval_sql.<locals>.func>, <class 'sqlglot.expressions.DateTrunc'>: <function BigQuery.Generator.<lambda>>, <class 'sqlglot.expressions.GenerateSeries'>: <function rename_func.<locals>.<lambda>>, <class 'sqlglot.expressions.GetPath'>: <function path_to_jsonpath.<locals>._transform>, <class 'sqlglot.expressions.GroupConcat'>: <function rename_func.<locals>.<lambda>>, <class 'sqlglot.expressions.Hex'>: <function rename_func.<locals>.<lambda>>, <class 'sqlglot.expressions.If'>: <function if_sql.<locals>._if_sql>, <class 'sqlglot.expressions.ILike'>: <function no_ilike_sql>, <class 'sqlglot.expressions.IntDiv'>: <function rename_func.<locals>.<lambda>>, <class 'sqlglot.expressions.JSONFormat'>: <function rename_func.<locals>.<lambda>>, <class 'sqlglot.expressions.JSONKeyValue'>: <function json_keyvalue_comma_sql>, <class 'sqlglot.expressions.Max'>: <function max_or_greatest>, <class 'sqlglot.expressions.MD5'>: <function BigQuery.Generator.<lambda>>, <class 'sqlglot.expressions.MD5Digest'>: <function rename_func.<locals>.<lambda>>, <class 'sqlglot.expressions.Min'>: <function min_or_least>, <class 'sqlglot.expressions.PartitionedByProperty'>: <function BigQuery.Generator.<lambda>>, <class 'sqlglot.expressions.RegexpExtract'>: <function BigQuery.Generator.<lambda>>, <class 'sqlglot.expressions.RegexpReplace'>: <function regexp_replace_sql>, <class 'sqlglot.expressions.RegexpLike'>: <function rename_func.<locals>.<lambda>>, <class 'sqlglot.expressions.Select'>: <function preprocess.<locals>._to_sql>, <class 'sqlglot.expressions.SHA2'>: <function BigQuery.Generator.<lambda>>, <class 'sqlglot.expressions.StrToDate'>: <function BigQuery.Generator.<lambda>>, <class 'sqlglot.expressions.StrToTime'>: <function BigQuery.Generator.<lambda>>, <class 'sqlglot.expressions.TimeAdd'>: <function date_add_interval_sql.<locals>.func>, <class 'sqlglot.expressions.TimeFromParts'>: <function rename_func.<locals>.<lambda>>, <class 'sqlglot.expressions.TimeSub'>: <function date_add_interval_sql.<locals>.func>, <class 'sqlglot.expressions.TimestampAdd'>: <function date_add_interval_sql.<locals>.func>, <class 'sqlglot.expressions.TimestampSub'>: <function date_add_interval_sql.<locals>.func>, <class 'sqlglot.expressions.TimeStrToTime'>: <function timestrtotime_sql>, <class 'sqlglot.expressions.Trim'>: <function BigQuery.Generator.<lambda>>, <class 'sqlglot.expressions.TsOrDsAdd'>: <function _ts_or_ds_add_sql>, <class 'sqlglot.expressions.TsOrDsDiff'>: <function _ts_or_ds_diff_sql>, <class 'sqlglot.expressions.TsOrDsToTime'>: <function rename_func.<locals>.<lambda>>, <class 'sqlglot.expressions.Unhex'>: <function rename_func.<locals>.<lambda>>, <class 'sqlglot.expressions.UnixDate'>: <function rename_func.<locals>.<lambda>>, <class 'sqlglot.expressions.UnixToTime'>: <function _unix_to_time_sql>, <class 'sqlglot.expressions.Values'>: <function _derived_table_values_to_unnest>, <class 'sqlglot.expressions.VariancePop'>: <function rename_func.<locals>.<lambda>>}
TYPE_MAPPING = {<Type.NCHAR: 'NCHAR'>: 'STRING', <Type.NVARCHAR: 'NVARCHAR'>: 'STRING', <Type.MEDIUMTEXT: 'MEDIUMTEXT'>: 'TEXT', <Type.LONGTEXT: 'LONGTEXT'>: 'TEXT', <Type.TINYTEXT: 'TINYTEXT'>: 'TEXT', <Type.MEDIUMBLOB: 'MEDIUMBLOB'>: 'BLOB', <Type.LONGBLOB: 'LONGBLOB'>: 'BLOB', <Type.TINYBLOB: 'TINYBLOB'>: 'BLOB', <Type.INET: 'INET'>: 'INET', <Type.BIGDECIMAL: 'BIGDECIMAL'>: 'BIGNUMERIC', <Type.BIGINT: 'BIGINT'>: 'INT64', <Type.BINARY: 'BINARY'>: 'BYTES', <Type.BOOLEAN: 'BOOLEAN'>: 'BOOL', <Type.CHAR: 'CHAR'>: 'STRING', <Type.DECIMAL: 'DECIMAL'>: 'NUMERIC', <Type.DOUBLE: 'DOUBLE'>: 'FLOAT64', <Type.FLOAT: 'FLOAT'>: 'FLOAT64', <Type.INT: 'INT'>: 'INT64', <Type.SMALLINT: 'SMALLINT'>: 'INT64', <Type.TEXT: 'TEXT'>: 'STRING', <Type.TIMESTAMP: 'TIMESTAMP'>: 'DATETIME', <Type.TIMESTAMPTZ: 'TIMESTAMPTZ'>: 'TIMESTAMP', <Type.TIMESTAMPLTZ: 'TIMESTAMPLTZ'>: 'TIMESTAMP', <Type.TINYINT: 'TINYINT'>: 'INT64', <Type.VARBINARY: 'VARBINARY'>: 'BYTES', <Type.VARCHAR: 'VARCHAR'>: 'STRING', <Type.VARIANT: 'VARIANT'>: 'ANY TYPE'}
PROPERTIES_LOCATION = {<class 'sqlglot.expressions.AlgorithmProperty'>: <Location.POST_CREATE: 'POST_CREATE'>, <class 'sqlglot.expressions.AutoIncrementProperty'>: <Location.POST_SCHEMA: 'POST_SCHEMA'>, <class 'sqlglot.expressions.AutoRefreshProperty'>: <Location.POST_SCHEMA: 'POST_SCHEMA'>, <class 'sqlglot.expressions.BlockCompressionProperty'>: <Location.POST_NAME: 'POST_NAME'>, <class 'sqlglot.expressions.CharacterSetProperty'>: <Location.POST_SCHEMA: 'POST_SCHEMA'>, <class 'sqlglot.expressions.ChecksumProperty'>: <Location.POST_NAME: 'POST_NAME'>, <class 'sqlglot.expressions.CollateProperty'>: <Location.POST_SCHEMA: 'POST_SCHEMA'>, <class 'sqlglot.expressions.CopyGrantsProperty'>: <Location.POST_SCHEMA: 'POST_SCHEMA'>, <class 'sqlglot.expressions.Cluster'>: <Location.POST_SCHEMA: 'POST_SCHEMA'>, <class 'sqlglot.expressions.ClusteredByProperty'>: <Location.POST_SCHEMA: 'POST_SCHEMA'>, <class 'sqlglot.expressions.DataBlocksizeProperty'>: <Location.POST_NAME: 'POST_NAME'>, <class 'sqlglot.expressions.DefinerProperty'>: <Location.POST_CREATE: 'POST_CREATE'>, <class 'sqlglot.expressions.DictRange'>: <Location.POST_SCHEMA: 'POST_SCHEMA'>, <class 'sqlglot.expressions.DictProperty'>: <Location.POST_SCHEMA: 'POST_SCHEMA'>, <class 'sqlglot.expressions.DistKeyProperty'>: <Location.POST_SCHEMA: 'POST_SCHEMA'>, <class 'sqlglot.expressions.DistStyleProperty'>: <Location.POST_SCHEMA: 'POST_SCHEMA'>, <class 'sqlglot.expressions.EngineProperty'>: <Location.POST_SCHEMA: 'POST_SCHEMA'>, <class 'sqlglot.expressions.ExecuteAsProperty'>: <Location.POST_SCHEMA: 'POST_SCHEMA'>, <class 'sqlglot.expressions.ExternalProperty'>: <Location.POST_CREATE: 'POST_CREATE'>, <class 'sqlglot.expressions.FallbackProperty'>: <Location.POST_NAME: 'POST_NAME'>, <class 'sqlglot.expressions.FileFormatProperty'>: <Location.POST_WITH: 'POST_WITH'>, <class 'sqlglot.expressions.FreespaceProperty'>: <Location.POST_NAME: 'POST_NAME'>, <class 'sqlglot.expressions.HeapProperty'>: <Location.POST_WITH: 'POST_WITH'>, <class 'sqlglot.expressions.InputModelProperty'>: <Location.POST_SCHEMA: 'POST_SCHEMA'>, <class 'sqlglot.expressions.IsolatedLoadingProperty'>: <Location.POST_NAME: 'POST_NAME'>, <class 'sqlglot.expressions.JournalProperty'>: <Location.POST_NAME: 'POST_NAME'>, <class 'sqlglot.expressions.LanguageProperty'>: <Location.POST_SCHEMA: 'POST_SCHEMA'>, <class 'sqlglot.expressions.LikeProperty'>: <Location.POST_SCHEMA: 'POST_SCHEMA'>, <class 'sqlglot.expressions.LocationProperty'>: <Location.POST_SCHEMA: 'POST_SCHEMA'>, <class 'sqlglot.expressions.LockingProperty'>: <Location.POST_ALIAS: 'POST_ALIAS'>, <class 'sqlglot.expressions.LogProperty'>: <Location.POST_NAME: 'POST_NAME'>, <class 'sqlglot.expressions.MaterializedProperty'>: <Location.POST_CREATE: 'POST_CREATE'>, <class 'sqlglot.expressions.MergeBlockRatioProperty'>: <Location.POST_NAME: 'POST_NAME'>, <class 'sqlglot.expressions.NoPrimaryIndexProperty'>: <Location.POST_EXPRESSION: 'POST_EXPRESSION'>, <class 'sqlglot.expressions.OnProperty'>: <Location.POST_SCHEMA: 'POST_SCHEMA'>, <class 'sqlglot.expressions.OnCommitProperty'>: <Location.POST_EXPRESSION: 'POST_EXPRESSION'>, <class 'sqlglot.expressions.Order'>: <Location.POST_SCHEMA: 'POST_SCHEMA'>, <class 'sqlglot.expressions.OutputModelProperty'>: <Location.POST_SCHEMA: 'POST_SCHEMA'>, <class 'sqlglot.expressions.PartitionedByProperty'>: <Location.POST_SCHEMA: 'POST_SCHEMA'>, <class 'sqlglot.expressions.PartitionedOfProperty'>: <Location.POST_SCHEMA: 'POST_SCHEMA'>, <class 'sqlglot.expressions.PrimaryKey'>: <Location.POST_SCHEMA: 'POST_SCHEMA'>, <class 'sqlglot.expressions.Property'>: <Location.POST_WITH: 'POST_WITH'>, <class 'sqlglot.expressions.RemoteWithConnectionModelProperty'>: <Location.POST_SCHEMA: 'POST_SCHEMA'>, <class 'sqlglot.expressions.ReturnsProperty'>: <Location.POST_SCHEMA: 'POST_SCHEMA'>, <class 'sqlglot.expressions.RowFormatProperty'>: <Location.POST_SCHEMA: 'POST_SCHEMA'>, <class 'sqlglot.expressions.RowFormatDelimitedProperty'>: <Location.POST_SCHEMA: 'POST_SCHEMA'>, <class 'sqlglot.expressions.RowFormatSerdeProperty'>: <Location.POST_SCHEMA: 'POST_SCHEMA'>, <class 'sqlglot.expressions.SampleProperty'>: <Location.POST_SCHEMA: 'POST_SCHEMA'>, <class 'sqlglot.expressions.SchemaCommentProperty'>: <Location.POST_SCHEMA: 'POST_SCHEMA'>, <class 'sqlglot.expressions.SerdeProperties'>: <Location.POST_SCHEMA: 'POST_SCHEMA'>, <class 'sqlglot.expressions.Set'>: <Location.POST_SCHEMA: 'POST_SCHEMA'>, <class 'sqlglot.expressions.SettingsProperty'>: <Location.POST_SCHEMA: 'POST_SCHEMA'>, <class 'sqlglot.expressions.SetProperty'>: <Location.POST_CREATE: 'POST_CREATE'>, <class 'sqlglot.expressions.SortKeyProperty'>: <Location.POST_SCHEMA: 'POST_SCHEMA'>, <class 'sqlglot.expressions.SqlReadWriteProperty'>: <Location.POST_SCHEMA: 'POST_SCHEMA'>, <class 'sqlglot.expressions.SqlSecurityProperty'>: <Location.POST_CREATE: 'POST_CREATE'>, <class 'sqlglot.expressions.StabilityProperty'>: <Location.POST_SCHEMA: 'POST_SCHEMA'>, <class 'sqlglot.expressions.TemporaryProperty'>: <Location.POST_CREATE: 'POST_CREATE'>, <class 'sqlglot.expressions.ToTableProperty'>: <Location.POST_SCHEMA: 'POST_SCHEMA'>, <class 'sqlglot.expressions.TransientProperty'>: <Location.POST_CREATE: 'POST_CREATE'>, <class 'sqlglot.expressions.TransformModelProperty'>: <Location.POST_SCHEMA: 'POST_SCHEMA'>, <class 'sqlglot.expressions.MergeTreeTTL'>: <Location.POST_SCHEMA: 'POST_SCHEMA'>, <class 'sqlglot.expressions.VolatileProperty'>: <Location.UNSUPPORTED: 'UNSUPPORTED'>, <class 'sqlglot.expressions.WithDataProperty'>: <Location.POST_EXPRESSION: 'POST_EXPRESSION'>, <class 'sqlglot.expressions.WithJournalTableProperty'>: <Location.POST_NAME: 'POST_NAME'>, <class 'sqlglot.expressions.WithSystemVersioningProperty'>: <Location.POST_SCHEMA: 'POST_SCHEMA'>}
RESERVED_KEYWORDS = {'on', 'then', 'when', 'or', 'not', 'asc', 'to', 'except', 'recursive', 'all', 'inner', 'for', 'cross', 'unbounded', 'of', 'escape', 'unnest', 'using', 'merge', 'where', 'new', 'enum', 'nulls', 'exists', 'create', 'like', 'current', 'any', 'natural', 'partition', 'rows', 'assert_rows_modified', 'select', 'no', 'order', 'rollup', 'lookup', 'ignore', 'left', 'lateral', 'set', 'with', 'range', 'false', 'case', 'desc', 'end', 'interval', 'group', 'union', 'join', 'contains', 'preceding', 'outer', 'by', 'intersect', 'treat', 'within', 'distinct', 'default', 'if', 'qualify', 'cast', 'in', 'as', 'exclude', 'grouping', 'tablesample', 'limit', 'full', 'at', 'fetch', 'cube', 'respect', 'array', 'true', 'having', 'is', 'window', 'groups', 'into', 'else', 'over', 'collate', 'extract', 'following', 'struct', 'right', 'between', 'define', 'and', 'from', 'null', 'some', 'hash', 'proto'}
def timetostr_sql(self, expression: sqlglot.expressions.TimeToStr) -> str:
749        def timetostr_sql(self, expression: exp.TimeToStr) -> str:
750            if isinstance(expression.this, exp.TsOrDsToDate):
751                this: exp.Expression = expression.this
752            else:
753                this = expression
754
755            return f"FORMAT_DATE({self.format_time(expression)}, {self.sql(this, 'this')})"
def struct_sql(self, expression: sqlglot.expressions.Struct) -> str:
757        def struct_sql(self, expression: exp.Struct) -> str:
758            args = []
759            for expr in expression.expressions:
760                if isinstance(expr, self.KEY_VALUE_DEFINITIONS):
761                    arg = f"{self.sql(expr, 'expression')} AS {expr.this.name}"
762                else:
763                    arg = self.sql(expr)
764
765                args.append(arg)
766
767            return self.func("STRUCT", *args)
def eq_sql(self, expression: sqlglot.expressions.EQ) -> str:
769        def eq_sql(self, expression: exp.EQ) -> str:
770            # Operands of = cannot be NULL in BigQuery
771            if isinstance(expression.left, exp.Null) or isinstance(expression.right, exp.Null):
772                if not isinstance(expression.parent, exp.Update):
773                    return "NULL"
774
775            return self.binary(expression, "=")
def attimezone_sql(self, expression: sqlglot.expressions.AtTimeZone) -> str:
777        def attimezone_sql(self, expression: exp.AtTimeZone) -> str:
778            parent = expression.parent
779
780            # BigQuery allows CAST(.. AS {STRING|TIMESTAMP} [FORMAT <fmt> [AT TIME ZONE <tz>]]).
781            # Only the TIMESTAMP one should use the below conversion, when AT TIME ZONE is included.
782            if not isinstance(parent, exp.Cast) or not parent.to.is_type("text"):
783                return self.func(
784                    "TIMESTAMP", self.func("DATETIME", expression.this, expression.args.get("zone"))
785                )
786
787            return super().attimezone_sql(expression)
def trycast_sql(self, expression: sqlglot.expressions.TryCast) -> str:
789        def trycast_sql(self, expression: exp.TryCast) -> str:
790            return self.cast_sql(expression, safe_prefix="SAFE_")
def cte_sql(self, expression: sqlglot.expressions.CTE) -> str:
792        def cte_sql(self, expression: exp.CTE) -> str:
793            if expression.alias_column_names:
794                self.unsupported("Column names in CTE definition are not supported.")
795            return super().cte_sql(expression)
def array_sql(self, expression: sqlglot.expressions.Array) -> str:
797        def array_sql(self, expression: exp.Array) -> str:
798            first_arg = seq_get(expression.expressions, 0)
799            if isinstance(first_arg, exp.Subqueryable):
800                return f"ARRAY{self.wrap(self.sql(first_arg))}"
801
802            return inline_array_sql(self, expression)
def bracket_sql(self, expression: sqlglot.expressions.Bracket) -> str:
804        def bracket_sql(self, expression: exp.Bracket) -> str:
805            this = self.sql(expression, "this")
806            expressions = expression.expressions
807
808            if len(expressions) == 1:
809                arg = expressions[0]
810                if arg.type is None:
811                    from sqlglot.optimizer.annotate_types import annotate_types
812
813                    arg = annotate_types(arg)
814
815                if arg.type and arg.type.this in exp.DataType.TEXT_TYPES:
816                    # BQ doesn't support bracket syntax with string values
817                    return f"{this}.{arg.name}"
818
819            expressions_sql = ", ".join(self.sql(e) for e in expressions)
820            offset = expression.args.get("offset")
821
822            if offset == 0:
823                expressions_sql = f"OFFSET({expressions_sql})"
824            elif offset == 1:
825                expressions_sql = f"ORDINAL({expressions_sql})"
826            elif offset is not None:
827                self.unsupported(f"Unsupported array offset: {offset}")
828
829            if expression.args.get("safe"):
830                expressions_sql = f"SAFE_{expressions_sql}"
831
832            return f"{this}[{expressions_sql}]"
def transaction_sql(self, *_) -> str:
834        def transaction_sql(self, *_) -> str:
835            return "BEGIN TRANSACTION"
def commit_sql(self, *_) -> str:
837        def commit_sql(self, *_) -> str:
838            return "COMMIT TRANSACTION"
def rollback_sql(self, *_) -> str:
840        def rollback_sql(self, *_) -> str:
841            return "ROLLBACK TRANSACTION"
def in_unnest_op(self, expression: sqlglot.expressions.Unnest) -> str:
843        def in_unnest_op(self, expression: exp.Unnest) -> str:
844            return self.sql(expression)
def except_op(self, expression: sqlglot.expressions.Except) -> str:
846        def except_op(self, expression: exp.Except) -> str:
847            if not expression.args.get("distinct", False):
848                self.unsupported("EXCEPT without DISTINCT is not supported in BigQuery")
849            return f"EXCEPT{' DISTINCT' if expression.args.get('distinct') else ' ALL'}"
def intersect_op(self, expression: sqlglot.expressions.Intersect) -> str:
851        def intersect_op(self, expression: exp.Intersect) -> str:
852            if not expression.args.get("distinct", False):
853                self.unsupported("INTERSECT without DISTINCT is not supported in BigQuery")
854            return f"INTERSECT{' DISTINCT' if expression.args.get('distinct') else ' ALL'}"
def with_properties(self, properties: sqlglot.expressions.Properties) -> str:
856        def with_properties(self, properties: exp.Properties) -> str:
857            return self.properties(properties, prefix=self.seg("OPTIONS"))
def version_sql(self, expression: sqlglot.expressions.Version) -> str:
859        def version_sql(self, expression: exp.Version) -> str:
860            if expression.name == "TIMESTAMP":
861                expression.set("this", "SYSTEM_TIME")
862            return super().version_sql(expression)
Inherited Members
sqlglot.generator.Generator
Generator
NULL_ORDERING_SUPPORTED
LOCKING_READS_SUPPORTED
WRAP_DERIVED_VALUES
CREATE_FUNCTION_RETURN_AS
MATCHED_BY_SOURCE
SINGLE_STRING_INTERVAL
GROUPINGS_SEP
INDEX_ON
QUERY_HINT_SEP
IS_BOOL_ALLOWED
DUPLICATE_KEY_UPDATE_WITH_SET
LIMIT_IS_TOP
RETURNING_END
COLUMN_JOIN_MARKS_SUPPORTED
EXTRACT_ALLOWS_QUOTES
TZ_TO_WITH_TIME_ZONE
SELECT_KINDS
VALUES_AS_TABLE
ALTER_TABLE_INCLUDE_COLUMN_KEYWORD
AGGREGATE_FILTER_SUPPORTED
SEMI_ANTI_JOIN_WITH_SIDE
COMPUTED_COLUMN_WITH_TYPE
SUPPORTS_TABLE_COPY
TABLESAMPLE_REQUIRES_PARENS
TABLESAMPLE_SIZE_IS_ROWS
TABLESAMPLE_KEYWORDS
TABLESAMPLE_WITH_METHOD
TABLESAMPLE_SEED_KEYWORD
DATA_TYPE_SPECIFIERS_ALLOWED
ENSURE_BOOLS
CTE_RECURSIVE_KEYWORD_REQUIRED
SUPPORTS_SINGLE_ARG_CONCAT
LAST_DAY_SUPPORTS_DATE_PART
STAR_MAPPING
TIME_PART_SINGULARS
TOKEN_MAPPING
STRUCT_DELIMITER
PARAMETER_TOKEN
WITH_SEPARATED_COMMENTS
EXCLUDE_COMMENTS
UNWRAPPED_INTERVAL_VALUES
EXPRESSIONS_WITHOUT_NESTED_CTES
KEY_VALUE_DEFINITIONS
SENTINEL_LINE_BREAK
pretty
identify
normalize
pad
unsupported_level
max_unsupported
leading_comma
max_text_width
comments
dialect
normalize_functions
unsupported_messages
generate
preprocess
unsupported
sep
seg
pad_comment
maybe_comment
wrap
no_identify
normalize_func
indent
sql
uncache_sql
cache_sql
characterset_sql
column_sql
columnposition_sql
columndef_sql
columnconstraint_sql
computedcolumnconstraint_sql
autoincrementcolumnconstraint_sql
compresscolumnconstraint_sql
generatedasidentitycolumnconstraint_sql
generatedasrowcolumnconstraint_sql
periodforsystemtimeconstraint_sql
notnullcolumnconstraint_sql
transformcolumnconstraint_sql
primarykeycolumnconstraint_sql
uniquecolumnconstraint_sql
createable_sql
create_sql
clone_sql
describe_sql
prepend_ctes
with_sql
tablealias_sql
bitstring_sql
hexstring_sql
bytestring_sql
unicodestring_sql
rawstring_sql
datatypeparam_sql
datatype_sql
directory_sql
delete_sql
drop_sql
except_sql
fetch_sql
filter_sql
hint_sql
index_sql
identifier_sql
inputoutputformat_sql
national_sql
partition_sql
properties_sql
root_properties
properties
locate_properties
property_name
property_sql
likeproperty_sql
fallbackproperty_sql
journalproperty_sql
freespaceproperty_sql
checksumproperty_sql
mergeblockratioproperty_sql
datablocksizeproperty_sql
blockcompressionproperty_sql
isolatedloadingproperty_sql
partitionboundspec_sql
partitionedofproperty_sql
lockingproperty_sql
withdataproperty_sql
withsystemversioningproperty_sql
insert_sql
intersect_sql
introducer_sql
kill_sql
pseudotype_sql
objectidentifier_sql
onconflict_sql
returning_sql
rowformatdelimitedproperty_sql
withtablehint_sql
indextablehint_sql
historicaldata_sql
table_sql
tablesample_sql
pivot_sql
tuple_sql
update_sql
values_sql
var_sql
into_sql
from_sql
group_sql
having_sql
connect_sql
prior_sql
join_sql
lambda_sql
lateral_op
lateral_sql
limit_sql
offset_sql
setitem_sql
set_sql
pragma_sql
lock_sql
literal_sql
escape_str
loaddata_sql
null_sql
boolean_sql
order_sql
withfill_sql
cluster_sql
distribute_sql
sort_sql
ordered_sql
matchrecognize_sql
query_modifiers
offset_limit_modifiers
after_having_modifiers
after_limit_modifiers
select_sql
schema_sql
schema_columns_sql
star_sql
parameter_sql
sessionparameter_sql
placeholder_sql
subquery_sql
qualify_sql
union_sql
union_op
unnest_sql
where_sql
window_sql
partition_by_sql
windowspec_sql
withingroup_sql
between_sql
all_sql
any_sql
exists_sql
case_sql
constraint_sql
nextvaluefor_sql
extract_sql
trim_sql
convert_concat_args
concat_sql
concatws_sql
check_sql
foreignkey_sql
primarykey_sql
if_sql
matchagainst_sql
jsonkeyvalue_sql
formatjson_sql
jsonobject_sql
jsonarray_sql
jsonarrayagg_sql
jsoncolumndef_sql
jsonschema_sql
jsontable_sql
openjsoncolumndef_sql
openjson_sql
in_sql
interval_sql
return_sql
reference_sql
anonymous_sql
paren_sql
neg_sql
not_sql
alias_sql
pivotalias_sql
aliases_sql
atindex_sql
add_sql
and_sql
xor_sql
connector_sql
bitwiseand_sql
bitwiseleftshift_sql
bitwisenot_sql
bitwiseor_sql
bitwiserightshift_sql
bitwisexor_sql
cast_sql
currentdate_sql
collate_sql
command_sql
comment_sql
mergetreettlaction_sql
mergetreettl_sql
altercolumn_sql
renametable_sql
altertable_sql
add_column_sql
droppartition_sql
addconstraint_sql
distinct_sql
ignorenulls_sql
respectnulls_sql
intdiv_sql
dpipe_sql
div_sql
overlaps_sql
distance_sql
dot_sql
propertyeq_sql
escape_sql
glob_sql
gt_sql
gte_sql
ilike_sql
ilikeany_sql
is_sql
like_sql
likeany_sql
similarto_sql
lt_sql
lte_sql
mod_sql
mul_sql
neq_sql
nullsafeeq_sql
nullsafeneq_sql
or_sql
slice_sql
sub_sql
log_sql
use_sql
binary
function_fallback_sql
func
format_args
text_width
format_time
expressions
op_expressions
naked_property
set_operation
tag_sql
token_sql
userdefinedfunction_sql
joinhint_sql
kwarg_sql
when_sql
merge_sql
tochar_sql
dictproperty_sql
dictrange_sql
dictsubproperty_sql
oncluster_sql
clusteredbyproperty_sql
anyvalue_sql
querytransform_sql
indexconstraintoption_sql
indexcolumnconstraint_sql
nvl2_sql
comprehension_sql
columnprefix_sql
opclass_sql
predict_sql
forin_sql
refresh_sql
operator_sql
toarray_sql
tsordstotime_sql
tsordstodate_sql
unixdate_sql
lastday_sql