| # |
| # Licensed to the Apache Software Foundation (ASF) under one or more |
| # contributor license agreements. See the NOTICE file distributed with |
| # this work for additional information regarding copyright ownership. |
| # The ASF licenses this file to You under the Apache License, Version 2.0 |
| # (the "License"); you may not use this file except in compliance with |
| # the License. You may obtain a copy of the License at |
| # |
| # http://www.apache.org/licenses/LICENSE-2.0 |
| # |
| # Unless required by applicable law or agreed to in writing, software |
| # distributed under the License is distributed on an "AS IS" BASIS, |
| # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
| # See the License for the specific language governing permissions and |
| # limitations under the License. |
| # |
| """ |
| @generated by mypy-protobuf. Do not edit manually! |
| isort:skip_file |
| |
| Licensed to the Apache Software Foundation (ASF) under one or more |
| contributor license agreements. See the NOTICE file distributed with |
| this work for additional information regarding copyright ownership. |
| The ASF licenses this file to You under the Apache License, Version 2.0 |
| (the "License"); you may not use this file except in compliance with |
| the License. You may obtain a copy of the License at |
| |
| http://www.apache.org/licenses/LICENSE-2.0 |
| |
| Unless required by applicable law or agreed to in writing, software |
| distributed under the License is distributed on an "AS IS" BASIS, |
| WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
| See the License for the specific language governing permissions and |
| limitations under the License. |
| """ |
| import builtins |
| import collections.abc |
| import google.protobuf.any_pb2 |
| import google.protobuf.descriptor |
| import google.protobuf.internal.containers |
| import google.protobuf.internal.enum_type_wrapper |
| import google.protobuf.message |
| import pyspark.sql.connect.proto.common_pb2 |
| import pyspark.sql.connect.proto.types_pb2 |
| import sys |
| import typing |
| |
| if sys.version_info >= (3, 10): |
| import typing as typing_extensions |
| else: |
| import typing_extensions |
| |
| DESCRIPTOR: google.protobuf.descriptor.FileDescriptor |
| |
| class Expression(google.protobuf.message.Message): |
| """Expression used to refer to fields, functions and similar. This can be used everywhere |
| expressions in SQL appear. |
| """ |
| |
| DESCRIPTOR: google.protobuf.descriptor.Descriptor |
| |
| class Window(google.protobuf.message.Message): |
| """Expression for the OVER clause or WINDOW clause.""" |
| |
| DESCRIPTOR: google.protobuf.descriptor.Descriptor |
| |
| class WindowFrame(google.protobuf.message.Message): |
| """The window frame""" |
| |
| DESCRIPTOR: google.protobuf.descriptor.Descriptor |
| |
| class _FrameType: |
| ValueType = typing.NewType("ValueType", builtins.int) |
| V: typing_extensions.TypeAlias = ValueType |
| |
| class _FrameTypeEnumTypeWrapper( |
| google.protobuf.internal.enum_type_wrapper._EnumTypeWrapper[ |
| Expression.Window.WindowFrame._FrameType.ValueType |
| ], |
| builtins.type, |
| ): # noqa: F821 |
| DESCRIPTOR: google.protobuf.descriptor.EnumDescriptor |
| FRAME_TYPE_UNDEFINED: Expression.Window.WindowFrame._FrameType.ValueType # 0 |
| FRAME_TYPE_ROW: Expression.Window.WindowFrame._FrameType.ValueType # 1 |
| """RowFrame treats rows in a partition individually.""" |
| FRAME_TYPE_RANGE: Expression.Window.WindowFrame._FrameType.ValueType # 2 |
| """RangeFrame treats rows in a partition as groups of peers. |
| All rows having the same 'ORDER BY' ordering are considered as peers. |
| """ |
| |
| class FrameType(_FrameType, metaclass=_FrameTypeEnumTypeWrapper): ... |
| FRAME_TYPE_UNDEFINED: Expression.Window.WindowFrame.FrameType.ValueType # 0 |
| FRAME_TYPE_ROW: Expression.Window.WindowFrame.FrameType.ValueType # 1 |
| """RowFrame treats rows in a partition individually.""" |
| FRAME_TYPE_RANGE: Expression.Window.WindowFrame.FrameType.ValueType # 2 |
| """RangeFrame treats rows in a partition as groups of peers. |
| All rows having the same 'ORDER BY' ordering are considered as peers. |
| """ |
| |
| class FrameBoundary(google.protobuf.message.Message): |
| DESCRIPTOR: google.protobuf.descriptor.Descriptor |
| |
| CURRENT_ROW_FIELD_NUMBER: builtins.int |
| UNBOUNDED_FIELD_NUMBER: builtins.int |
| VALUE_FIELD_NUMBER: builtins.int |
| current_row: builtins.bool |
| """CURRENT ROW boundary""" |
| unbounded: builtins.bool |
| """UNBOUNDED boundary. |
| For lower bound, it will be converted to 'UnboundedPreceding'. |
| for upper bound, it will be converted to 'UnboundedFollowing'. |
| """ |
| @property |
| def value(self) -> global___Expression: |
| """This is an expression for future proofing. We are expecting literals on the server side.""" |
| def __init__( |
| self, |
| *, |
| current_row: builtins.bool = ..., |
| unbounded: builtins.bool = ..., |
| value: global___Expression | None = ..., |
| ) -> None: ... |
| def HasField( |
| self, |
| field_name: typing_extensions.Literal[ |
| "boundary", |
| b"boundary", |
| "current_row", |
| b"current_row", |
| "unbounded", |
| b"unbounded", |
| "value", |
| b"value", |
| ], |
| ) -> builtins.bool: ... |
| def ClearField( |
| self, |
| field_name: typing_extensions.Literal[ |
| "boundary", |
| b"boundary", |
| "current_row", |
| b"current_row", |
| "unbounded", |
| b"unbounded", |
| "value", |
| b"value", |
| ], |
| ) -> None: ... |
| def WhichOneof( |
| self, oneof_group: typing_extensions.Literal["boundary", b"boundary"] |
| ) -> typing_extensions.Literal["current_row", "unbounded", "value"] | None: ... |
| |
| FRAME_TYPE_FIELD_NUMBER: builtins.int |
| LOWER_FIELD_NUMBER: builtins.int |
| UPPER_FIELD_NUMBER: builtins.int |
| frame_type: global___Expression.Window.WindowFrame.FrameType.ValueType |
| """(Required) The type of the frame.""" |
| @property |
| def lower(self) -> global___Expression.Window.WindowFrame.FrameBoundary: |
| """(Required) The lower bound of the frame.""" |
| @property |
| def upper(self) -> global___Expression.Window.WindowFrame.FrameBoundary: |
| """(Required) The upper bound of the frame.""" |
| def __init__( |
| self, |
| *, |
| frame_type: global___Expression.Window.WindowFrame.FrameType.ValueType = ..., |
| lower: global___Expression.Window.WindowFrame.FrameBoundary | None = ..., |
| upper: global___Expression.Window.WindowFrame.FrameBoundary | None = ..., |
| ) -> None: ... |
| def HasField( |
| self, field_name: typing_extensions.Literal["lower", b"lower", "upper", b"upper"] |
| ) -> builtins.bool: ... |
| def ClearField( |
| self, |
| field_name: typing_extensions.Literal[ |
| "frame_type", b"frame_type", "lower", b"lower", "upper", b"upper" |
| ], |
| ) -> None: ... |
| |
| WINDOW_FUNCTION_FIELD_NUMBER: builtins.int |
| PARTITION_SPEC_FIELD_NUMBER: builtins.int |
| ORDER_SPEC_FIELD_NUMBER: builtins.int |
| FRAME_SPEC_FIELD_NUMBER: builtins.int |
| @property |
| def window_function(self) -> global___Expression: |
| """(Required) The window function.""" |
| @property |
| def partition_spec( |
| self, |
| ) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[ |
| global___Expression |
| ]: |
| """(Optional) The way that input rows are partitioned.""" |
| @property |
| def order_spec( |
| self, |
| ) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[ |
| global___Expression.SortOrder |
| ]: |
| """(Optional) Ordering of rows in a partition.""" |
| @property |
| def frame_spec(self) -> global___Expression.Window.WindowFrame: |
| """(Optional) Window frame in a partition. |
| |
| If not set, it will be treated as 'UnspecifiedFrame'. |
| """ |
| def __init__( |
| self, |
| *, |
| window_function: global___Expression | None = ..., |
| partition_spec: collections.abc.Iterable[global___Expression] | None = ..., |
| order_spec: collections.abc.Iterable[global___Expression.SortOrder] | None = ..., |
| frame_spec: global___Expression.Window.WindowFrame | None = ..., |
| ) -> None: ... |
| def HasField( |
| self, |
| field_name: typing_extensions.Literal[ |
| "frame_spec", b"frame_spec", "window_function", b"window_function" |
| ], |
| ) -> builtins.bool: ... |
| def ClearField( |
| self, |
| field_name: typing_extensions.Literal[ |
| "frame_spec", |
| b"frame_spec", |
| "order_spec", |
| b"order_spec", |
| "partition_spec", |
| b"partition_spec", |
| "window_function", |
| b"window_function", |
| ], |
| ) -> None: ... |
| |
| class SortOrder(google.protobuf.message.Message): |
| """SortOrder is used to specify the data ordering, it is normally used in Sort and Window. |
| It is an unevaluable expression and cannot be evaluated, so can not be used in Projection. |
| """ |
| |
| DESCRIPTOR: google.protobuf.descriptor.Descriptor |
| |
| class _SortDirection: |
| ValueType = typing.NewType("ValueType", builtins.int) |
| V: typing_extensions.TypeAlias = ValueType |
| |
| class _SortDirectionEnumTypeWrapper( |
| google.protobuf.internal.enum_type_wrapper._EnumTypeWrapper[ |
| Expression.SortOrder._SortDirection.ValueType |
| ], |
| builtins.type, |
| ): # noqa: F821 |
| DESCRIPTOR: google.protobuf.descriptor.EnumDescriptor |
| SORT_DIRECTION_UNSPECIFIED: Expression.SortOrder._SortDirection.ValueType # 0 |
| SORT_DIRECTION_ASCENDING: Expression.SortOrder._SortDirection.ValueType # 1 |
| SORT_DIRECTION_DESCENDING: Expression.SortOrder._SortDirection.ValueType # 2 |
| |
| class SortDirection(_SortDirection, metaclass=_SortDirectionEnumTypeWrapper): ... |
| SORT_DIRECTION_UNSPECIFIED: Expression.SortOrder.SortDirection.ValueType # 0 |
| SORT_DIRECTION_ASCENDING: Expression.SortOrder.SortDirection.ValueType # 1 |
| SORT_DIRECTION_DESCENDING: Expression.SortOrder.SortDirection.ValueType # 2 |
| |
| class _NullOrdering: |
| ValueType = typing.NewType("ValueType", builtins.int) |
| V: typing_extensions.TypeAlias = ValueType |
| |
| class _NullOrderingEnumTypeWrapper( |
| google.protobuf.internal.enum_type_wrapper._EnumTypeWrapper[ |
| Expression.SortOrder._NullOrdering.ValueType |
| ], |
| builtins.type, |
| ): # noqa: F821 |
| DESCRIPTOR: google.protobuf.descriptor.EnumDescriptor |
| SORT_NULLS_UNSPECIFIED: Expression.SortOrder._NullOrdering.ValueType # 0 |
| SORT_NULLS_FIRST: Expression.SortOrder._NullOrdering.ValueType # 1 |
| SORT_NULLS_LAST: Expression.SortOrder._NullOrdering.ValueType # 2 |
| |
| class NullOrdering(_NullOrdering, metaclass=_NullOrderingEnumTypeWrapper): ... |
| SORT_NULLS_UNSPECIFIED: Expression.SortOrder.NullOrdering.ValueType # 0 |
| SORT_NULLS_FIRST: Expression.SortOrder.NullOrdering.ValueType # 1 |
| SORT_NULLS_LAST: Expression.SortOrder.NullOrdering.ValueType # 2 |
| |
| CHILD_FIELD_NUMBER: builtins.int |
| DIRECTION_FIELD_NUMBER: builtins.int |
| NULL_ORDERING_FIELD_NUMBER: builtins.int |
| @property |
| def child(self) -> global___Expression: |
| """(Required) The expression to be sorted.""" |
| direction: global___Expression.SortOrder.SortDirection.ValueType |
| """(Required) The sort direction, should be ASCENDING or DESCENDING.""" |
| null_ordering: global___Expression.SortOrder.NullOrdering.ValueType |
| """(Required) How to deal with NULLs, should be NULLS_FIRST or NULLS_LAST.""" |
| def __init__( |
| self, |
| *, |
| child: global___Expression | None = ..., |
| direction: global___Expression.SortOrder.SortDirection.ValueType = ..., |
| null_ordering: global___Expression.SortOrder.NullOrdering.ValueType = ..., |
| ) -> None: ... |
| def HasField( |
| self, field_name: typing_extensions.Literal["child", b"child"] |
| ) -> builtins.bool: ... |
| def ClearField( |
| self, |
| field_name: typing_extensions.Literal[ |
| "child", b"child", "direction", b"direction", "null_ordering", b"null_ordering" |
| ], |
| ) -> None: ... |
| |
| class DirectShufflePartitionID(google.protobuf.message.Message): |
| """Expression that takes a partition ID value and passes it through directly for use in |
| shuffle partitioning. This is used with RepartitionByExpression to allow users to |
| directly specify target partition IDs. |
| """ |
| |
| DESCRIPTOR: google.protobuf.descriptor.Descriptor |
| |
| CHILD_FIELD_NUMBER: builtins.int |
| @property |
| def child(self) -> global___Expression: |
| """(Required) The expression that evaluates to the partition ID.""" |
| def __init__( |
| self, |
| *, |
| child: global___Expression | None = ..., |
| ) -> None: ... |
| def HasField( |
| self, field_name: typing_extensions.Literal["child", b"child"] |
| ) -> builtins.bool: ... |
| def ClearField(self, field_name: typing_extensions.Literal["child", b"child"]) -> None: ... |
| |
| class Cast(google.protobuf.message.Message): |
| DESCRIPTOR: google.protobuf.descriptor.Descriptor |
| |
| class _EvalMode: |
| ValueType = typing.NewType("ValueType", builtins.int) |
| V: typing_extensions.TypeAlias = ValueType |
| |
| class _EvalModeEnumTypeWrapper( |
| google.protobuf.internal.enum_type_wrapper._EnumTypeWrapper[ |
| Expression.Cast._EvalMode.ValueType |
| ], |
| builtins.type, |
| ): # noqa: F821 |
| DESCRIPTOR: google.protobuf.descriptor.EnumDescriptor |
| EVAL_MODE_UNSPECIFIED: Expression.Cast._EvalMode.ValueType # 0 |
| EVAL_MODE_LEGACY: Expression.Cast._EvalMode.ValueType # 1 |
| EVAL_MODE_ANSI: Expression.Cast._EvalMode.ValueType # 2 |
| EVAL_MODE_TRY: Expression.Cast._EvalMode.ValueType # 3 |
| |
| class EvalMode(_EvalMode, metaclass=_EvalModeEnumTypeWrapper): ... |
| EVAL_MODE_UNSPECIFIED: Expression.Cast.EvalMode.ValueType # 0 |
| EVAL_MODE_LEGACY: Expression.Cast.EvalMode.ValueType # 1 |
| EVAL_MODE_ANSI: Expression.Cast.EvalMode.ValueType # 2 |
| EVAL_MODE_TRY: Expression.Cast.EvalMode.ValueType # 3 |
| |
| EXPR_FIELD_NUMBER: builtins.int |
| TYPE_FIELD_NUMBER: builtins.int |
| TYPE_STR_FIELD_NUMBER: builtins.int |
| EVAL_MODE_FIELD_NUMBER: builtins.int |
| @property |
| def expr(self) -> global___Expression: |
| """(Required) the expression to be casted.""" |
| @property |
| def type(self) -> pyspark.sql.connect.proto.types_pb2.DataType: ... |
| type_str: builtins.str |
| """If this is set, Server will use Catalyst parser to parse this string to DataType.""" |
| eval_mode: global___Expression.Cast.EvalMode.ValueType |
| """(Optional) The expression evaluation mode.""" |
| def __init__( |
| self, |
| *, |
| expr: global___Expression | None = ..., |
| type: pyspark.sql.connect.proto.types_pb2.DataType | None = ..., |
| type_str: builtins.str = ..., |
| eval_mode: global___Expression.Cast.EvalMode.ValueType = ..., |
| ) -> None: ... |
| def HasField( |
| self, |
| field_name: typing_extensions.Literal[ |
| "cast_to_type", |
| b"cast_to_type", |
| "expr", |
| b"expr", |
| "type", |
| b"type", |
| "type_str", |
| b"type_str", |
| ], |
| ) -> builtins.bool: ... |
| def ClearField( |
| self, |
| field_name: typing_extensions.Literal[ |
| "cast_to_type", |
| b"cast_to_type", |
| "eval_mode", |
| b"eval_mode", |
| "expr", |
| b"expr", |
| "type", |
| b"type", |
| "type_str", |
| b"type_str", |
| ], |
| ) -> None: ... |
| def WhichOneof( |
| self, oneof_group: typing_extensions.Literal["cast_to_type", b"cast_to_type"] |
| ) -> typing_extensions.Literal["type", "type_str"] | None: ... |
| |
| class Literal(google.protobuf.message.Message): |
| DESCRIPTOR: google.protobuf.descriptor.Descriptor |
| |
| class Decimal(google.protobuf.message.Message): |
| DESCRIPTOR: google.protobuf.descriptor.Descriptor |
| |
| VALUE_FIELD_NUMBER: builtins.int |
| PRECISION_FIELD_NUMBER: builtins.int |
| SCALE_FIELD_NUMBER: builtins.int |
| value: builtins.str |
| """the string representation.""" |
| precision: builtins.int |
| """The maximum number of digits allowed in the value. |
| the maximum precision is 38. |
| """ |
| scale: builtins.int |
| """declared scale of decimal literal""" |
| def __init__( |
| self, |
| *, |
| value: builtins.str = ..., |
| precision: builtins.int | None = ..., |
| scale: builtins.int | None = ..., |
| ) -> None: ... |
| def HasField( |
| self, |
| field_name: typing_extensions.Literal[ |
| "_precision", |
| b"_precision", |
| "_scale", |
| b"_scale", |
| "precision", |
| b"precision", |
| "scale", |
| b"scale", |
| ], |
| ) -> builtins.bool: ... |
| def ClearField( |
| self, |
| field_name: typing_extensions.Literal[ |
| "_precision", |
| b"_precision", |
| "_scale", |
| b"_scale", |
| "precision", |
| b"precision", |
| "scale", |
| b"scale", |
| "value", |
| b"value", |
| ], |
| ) -> None: ... |
| @typing.overload |
| def WhichOneof( |
| self, oneof_group: typing_extensions.Literal["_precision", b"_precision"] |
| ) -> typing_extensions.Literal["precision"] | None: ... |
| @typing.overload |
| def WhichOneof( |
| self, oneof_group: typing_extensions.Literal["_scale", b"_scale"] |
| ) -> typing_extensions.Literal["scale"] | None: ... |
| |
| class CalendarInterval(google.protobuf.message.Message): |
| DESCRIPTOR: google.protobuf.descriptor.Descriptor |
| |
| MONTHS_FIELD_NUMBER: builtins.int |
| DAYS_FIELD_NUMBER: builtins.int |
| MICROSECONDS_FIELD_NUMBER: builtins.int |
| months: builtins.int |
| days: builtins.int |
| microseconds: builtins.int |
| def __init__( |
| self, |
| *, |
| months: builtins.int = ..., |
| days: builtins.int = ..., |
| microseconds: builtins.int = ..., |
| ) -> None: ... |
| def ClearField( |
| self, |
| field_name: typing_extensions.Literal[ |
| "days", b"days", "microseconds", b"microseconds", "months", b"months" |
| ], |
| ) -> None: ... |
| |
| class Array(google.protobuf.message.Message): |
| DESCRIPTOR: google.protobuf.descriptor.Descriptor |
| |
| ELEMENT_TYPE_FIELD_NUMBER: builtins.int |
| ELEMENTS_FIELD_NUMBER: builtins.int |
| @property |
| def element_type(self) -> pyspark.sql.connect.proto.types_pb2.DataType: |
| """(Deprecated) The element type of the array. |
| |
| This field is deprecated since Spark 4.1+. Use data_type field instead. |
| """ |
| @property |
| def elements( |
| self, |
| ) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[ |
| global___Expression.Literal |
| ]: |
| """The literal values that make up the array elements.""" |
| def __init__( |
| self, |
| *, |
| element_type: pyspark.sql.connect.proto.types_pb2.DataType | None = ..., |
| elements: collections.abc.Iterable[global___Expression.Literal] | None = ..., |
| ) -> None: ... |
| def HasField( |
| self, field_name: typing_extensions.Literal["element_type", b"element_type"] |
| ) -> builtins.bool: ... |
| def ClearField( |
| self, |
| field_name: typing_extensions.Literal[ |
| "element_type", b"element_type", "elements", b"elements" |
| ], |
| ) -> None: ... |
| |
| class Map(google.protobuf.message.Message): |
| DESCRIPTOR: google.protobuf.descriptor.Descriptor |
| |
| KEY_TYPE_FIELD_NUMBER: builtins.int |
| VALUE_TYPE_FIELD_NUMBER: builtins.int |
| KEYS_FIELD_NUMBER: builtins.int |
| VALUES_FIELD_NUMBER: builtins.int |
| @property |
| def key_type(self) -> pyspark.sql.connect.proto.types_pb2.DataType: |
| """(Deprecated) The key type of the map. |
| |
| This field is deprecated since Spark 4.1+. Use data_type field instead. |
| """ |
| @property |
| def value_type(self) -> pyspark.sql.connect.proto.types_pb2.DataType: |
| """(Deprecated) The value type of the map. |
| |
| This field is deprecated since Spark 4.1+ and should only be set |
| if the data_type field is not set. Use data_type field instead. |
| """ |
| @property |
| def keys( |
| self, |
| ) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[ |
| global___Expression.Literal |
| ]: |
| """The literal keys that make up the map.""" |
| @property |
| def values( |
| self, |
| ) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[ |
| global___Expression.Literal |
| ]: |
| """The literal values that make up the map.""" |
| def __init__( |
| self, |
| *, |
| key_type: pyspark.sql.connect.proto.types_pb2.DataType | None = ..., |
| value_type: pyspark.sql.connect.proto.types_pb2.DataType | None = ..., |
| keys: collections.abc.Iterable[global___Expression.Literal] | None = ..., |
| values: collections.abc.Iterable[global___Expression.Literal] | None = ..., |
| ) -> None: ... |
| def HasField( |
| self, |
| field_name: typing_extensions.Literal[ |
| "key_type", b"key_type", "value_type", b"value_type" |
| ], |
| ) -> builtins.bool: ... |
| def ClearField( |
| self, |
| field_name: typing_extensions.Literal[ |
| "key_type", |
| b"key_type", |
| "keys", |
| b"keys", |
| "value_type", |
| b"value_type", |
| "values", |
| b"values", |
| ], |
| ) -> None: ... |
| |
| class Struct(google.protobuf.message.Message): |
| DESCRIPTOR: google.protobuf.descriptor.Descriptor |
| |
| STRUCT_TYPE_FIELD_NUMBER: builtins.int |
| ELEMENTS_FIELD_NUMBER: builtins.int |
| @property |
| def struct_type(self) -> pyspark.sql.connect.proto.types_pb2.DataType: |
| """(Deprecated) The type of the struct. |
| |
| This field is deprecated since Spark 4.1+ because using DataType as the type of a struct |
| is ambiguous. Use data_type field instead. |
| """ |
| @property |
| def elements( |
| self, |
| ) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[ |
| global___Expression.Literal |
| ]: |
| """The literal values that make up the struct elements.""" |
| def __init__( |
| self, |
| *, |
| struct_type: pyspark.sql.connect.proto.types_pb2.DataType | None = ..., |
| elements: collections.abc.Iterable[global___Expression.Literal] | None = ..., |
| ) -> None: ... |
| def HasField( |
| self, field_name: typing_extensions.Literal["struct_type", b"struct_type"] |
| ) -> builtins.bool: ... |
| def ClearField( |
| self, |
| field_name: typing_extensions.Literal[ |
| "elements", b"elements", "struct_type", b"struct_type" |
| ], |
| ) -> None: ... |
| |
| class SpecializedArray(google.protobuf.message.Message): |
| DESCRIPTOR: google.protobuf.descriptor.Descriptor |
| |
| BOOLS_FIELD_NUMBER: builtins.int |
| INTS_FIELD_NUMBER: builtins.int |
| LONGS_FIELD_NUMBER: builtins.int |
| FLOATS_FIELD_NUMBER: builtins.int |
| DOUBLES_FIELD_NUMBER: builtins.int |
| STRINGS_FIELD_NUMBER: builtins.int |
| @property |
| def bools(self) -> pyspark.sql.connect.proto.common_pb2.Bools: ... |
| @property |
| def ints(self) -> pyspark.sql.connect.proto.common_pb2.Ints: ... |
| @property |
| def longs(self) -> pyspark.sql.connect.proto.common_pb2.Longs: ... |
| @property |
| def floats(self) -> pyspark.sql.connect.proto.common_pb2.Floats: ... |
| @property |
| def doubles(self) -> pyspark.sql.connect.proto.common_pb2.Doubles: ... |
| @property |
| def strings(self) -> pyspark.sql.connect.proto.common_pb2.Strings: ... |
| def __init__( |
| self, |
| *, |
| bools: pyspark.sql.connect.proto.common_pb2.Bools | None = ..., |
| ints: pyspark.sql.connect.proto.common_pb2.Ints | None = ..., |
| longs: pyspark.sql.connect.proto.common_pb2.Longs | None = ..., |
| floats: pyspark.sql.connect.proto.common_pb2.Floats | None = ..., |
| doubles: pyspark.sql.connect.proto.common_pb2.Doubles | None = ..., |
| strings: pyspark.sql.connect.proto.common_pb2.Strings | None = ..., |
| ) -> None: ... |
| def HasField( |
| self, |
| field_name: typing_extensions.Literal[ |
| "bools", |
| b"bools", |
| "doubles", |
| b"doubles", |
| "floats", |
| b"floats", |
| "ints", |
| b"ints", |
| "longs", |
| b"longs", |
| "strings", |
| b"strings", |
| "value_type", |
| b"value_type", |
| ], |
| ) -> builtins.bool: ... |
| def ClearField( |
| self, |
| field_name: typing_extensions.Literal[ |
| "bools", |
| b"bools", |
| "doubles", |
| b"doubles", |
| "floats", |
| b"floats", |
| "ints", |
| b"ints", |
| "longs", |
| b"longs", |
| "strings", |
| b"strings", |
| "value_type", |
| b"value_type", |
| ], |
| ) -> None: ... |
| def WhichOneof( |
| self, oneof_group: typing_extensions.Literal["value_type", b"value_type"] |
| ) -> ( |
| typing_extensions.Literal["bools", "ints", "longs", "floats", "doubles", "strings"] |
| | None |
| ): ... |
| |
| class Time(google.protobuf.message.Message): |
| DESCRIPTOR: google.protobuf.descriptor.Descriptor |
| |
| NANO_FIELD_NUMBER: builtins.int |
| PRECISION_FIELD_NUMBER: builtins.int |
| nano: builtins.int |
| precision: builtins.int |
| """The precision of this time, if omitted, uses the default value of MICROS_PRECISION.""" |
| def __init__( |
| self, |
| *, |
| nano: builtins.int = ..., |
| precision: builtins.int | None = ..., |
| ) -> None: ... |
| def HasField( |
| self, |
| field_name: typing_extensions.Literal[ |
| "_precision", b"_precision", "precision", b"precision" |
| ], |
| ) -> builtins.bool: ... |
| def ClearField( |
| self, |
| field_name: typing_extensions.Literal[ |
| "_precision", b"_precision", "nano", b"nano", "precision", b"precision" |
| ], |
| ) -> None: ... |
| def WhichOneof( |
| self, oneof_group: typing_extensions.Literal["_precision", b"_precision"] |
| ) -> typing_extensions.Literal["precision"] | None: ... |
| |
| NULL_FIELD_NUMBER: builtins.int |
| BINARY_FIELD_NUMBER: builtins.int |
| BOOLEAN_FIELD_NUMBER: builtins.int |
| BYTE_FIELD_NUMBER: builtins.int |
| SHORT_FIELD_NUMBER: builtins.int |
| INTEGER_FIELD_NUMBER: builtins.int |
| LONG_FIELD_NUMBER: builtins.int |
| FLOAT_FIELD_NUMBER: builtins.int |
| DOUBLE_FIELD_NUMBER: builtins.int |
| DECIMAL_FIELD_NUMBER: builtins.int |
| STRING_FIELD_NUMBER: builtins.int |
| DATE_FIELD_NUMBER: builtins.int |
| TIMESTAMP_FIELD_NUMBER: builtins.int |
| TIMESTAMP_NTZ_FIELD_NUMBER: builtins.int |
| CALENDAR_INTERVAL_FIELD_NUMBER: builtins.int |
| YEAR_MONTH_INTERVAL_FIELD_NUMBER: builtins.int |
| DAY_TIME_INTERVAL_FIELD_NUMBER: builtins.int |
| ARRAY_FIELD_NUMBER: builtins.int |
| MAP_FIELD_NUMBER: builtins.int |
| STRUCT_FIELD_NUMBER: builtins.int |
| SPECIALIZED_ARRAY_FIELD_NUMBER: builtins.int |
| TIME_FIELD_NUMBER: builtins.int |
| DATA_TYPE_FIELD_NUMBER: builtins.int |
| @property |
| def null(self) -> pyspark.sql.connect.proto.types_pb2.DataType: ... |
| binary: builtins.bytes |
| boolean: builtins.bool |
| byte: builtins.int |
| short: builtins.int |
| integer: builtins.int |
| long: builtins.int |
| float: builtins.float |
| double: builtins.float |
| @property |
| def decimal(self) -> global___Expression.Literal.Decimal: ... |
| string: builtins.str |
| date: builtins.int |
| """Date in units of days since the UNIX epoch.""" |
| timestamp: builtins.int |
| """Timestamp in units of microseconds since the UNIX epoch.""" |
| timestamp_ntz: builtins.int |
| """Timestamp in units of microseconds since the UNIX epoch (without timezone information).""" |
| @property |
| def calendar_interval(self) -> global___Expression.Literal.CalendarInterval: ... |
| year_month_interval: builtins.int |
| day_time_interval: builtins.int |
| @property |
| def array(self) -> global___Expression.Literal.Array: ... |
| @property |
| def map(self) -> global___Expression.Literal.Map: ... |
| @property |
| def struct(self) -> global___Expression.Literal.Struct: ... |
| @property |
| def specialized_array(self) -> global___Expression.Literal.SpecializedArray: ... |
| @property |
| def time(self) -> global___Expression.Literal.Time: ... |
| @property |
| def data_type(self) -> pyspark.sql.connect.proto.types_pb2.DataType: |
| """Data type information for the literal. |
| This field is required only in the root literal message for null values or |
| for data types (e.g., array, map, or struct) with non-trivial information. |
| If the data_type field is not set at the root level, the data type will be |
| inferred or retrieved from the deprecated data type fields using best efforts. |
| """ |
| def __init__( |
| self, |
| *, |
| null: pyspark.sql.connect.proto.types_pb2.DataType | None = ..., |
| binary: builtins.bytes = ..., |
| boolean: builtins.bool = ..., |
| byte: builtins.int = ..., |
| short: builtins.int = ..., |
| integer: builtins.int = ..., |
| long: builtins.int = ..., |
| float: builtins.float = ..., |
| double: builtins.float = ..., |
| decimal: global___Expression.Literal.Decimal | None = ..., |
| string: builtins.str = ..., |
| date: builtins.int = ..., |
| timestamp: builtins.int = ..., |
| timestamp_ntz: builtins.int = ..., |
| calendar_interval: global___Expression.Literal.CalendarInterval | None = ..., |
| year_month_interval: builtins.int = ..., |
| day_time_interval: builtins.int = ..., |
| array: global___Expression.Literal.Array | None = ..., |
| map: global___Expression.Literal.Map | None = ..., |
| struct: global___Expression.Literal.Struct | None = ..., |
| specialized_array: global___Expression.Literal.SpecializedArray | None = ..., |
| time: global___Expression.Literal.Time | None = ..., |
| data_type: pyspark.sql.connect.proto.types_pb2.DataType | None = ..., |
| ) -> None: ... |
| def HasField( |
| self, |
| field_name: typing_extensions.Literal[ |
| "array", |
| b"array", |
| "binary", |
| b"binary", |
| "boolean", |
| b"boolean", |
| "byte", |
| b"byte", |
| "calendar_interval", |
| b"calendar_interval", |
| "data_type", |
| b"data_type", |
| "date", |
| b"date", |
| "day_time_interval", |
| b"day_time_interval", |
| "decimal", |
| b"decimal", |
| "double", |
| b"double", |
| "float", |
| b"float", |
| "integer", |
| b"integer", |
| "literal_type", |
| b"literal_type", |
| "long", |
| b"long", |
| "map", |
| b"map", |
| "null", |
| b"null", |
| "short", |
| b"short", |
| "specialized_array", |
| b"specialized_array", |
| "string", |
| b"string", |
| "struct", |
| b"struct", |
| "time", |
| b"time", |
| "timestamp", |
| b"timestamp", |
| "timestamp_ntz", |
| b"timestamp_ntz", |
| "year_month_interval", |
| b"year_month_interval", |
| ], |
| ) -> builtins.bool: ... |
| def ClearField( |
| self, |
| field_name: typing_extensions.Literal[ |
| "array", |
| b"array", |
| "binary", |
| b"binary", |
| "boolean", |
| b"boolean", |
| "byte", |
| b"byte", |
| "calendar_interval", |
| b"calendar_interval", |
| "data_type", |
| b"data_type", |
| "date", |
| b"date", |
| "day_time_interval", |
| b"day_time_interval", |
| "decimal", |
| b"decimal", |
| "double", |
| b"double", |
| "float", |
| b"float", |
| "integer", |
| b"integer", |
| "literal_type", |
| b"literal_type", |
| "long", |
| b"long", |
| "map", |
| b"map", |
| "null", |
| b"null", |
| "short", |
| b"short", |
| "specialized_array", |
| b"specialized_array", |
| "string", |
| b"string", |
| "struct", |
| b"struct", |
| "time", |
| b"time", |
| "timestamp", |
| b"timestamp", |
| "timestamp_ntz", |
| b"timestamp_ntz", |
| "year_month_interval", |
| b"year_month_interval", |
| ], |
| ) -> None: ... |
| def WhichOneof( |
| self, oneof_group: typing_extensions.Literal["literal_type", b"literal_type"] |
| ) -> ( |
| typing_extensions.Literal[ |
| "null", |
| "binary", |
| "boolean", |
| "byte", |
| "short", |
| "integer", |
| "long", |
| "float", |
| "double", |
| "decimal", |
| "string", |
| "date", |
| "timestamp", |
| "timestamp_ntz", |
| "calendar_interval", |
| "year_month_interval", |
| "day_time_interval", |
| "array", |
| "map", |
| "struct", |
| "specialized_array", |
| "time", |
| ] |
| | None |
| ): ... |
| |
| class UnresolvedAttribute(google.protobuf.message.Message): |
| """An unresolved attribute that is not explicitly bound to a specific column, but the column |
| is resolved during analysis by name. |
| """ |
| |
| DESCRIPTOR: google.protobuf.descriptor.Descriptor |
| |
| UNPARSED_IDENTIFIER_FIELD_NUMBER: builtins.int |
| PLAN_ID_FIELD_NUMBER: builtins.int |
| IS_METADATA_COLUMN_FIELD_NUMBER: builtins.int |
| unparsed_identifier: builtins.str |
| """(Required) An identifier that will be parsed by Catalyst parser. This should follow the |
| Spark SQL identifier syntax. |
| """ |
| plan_id: builtins.int |
| """(Optional) The id of corresponding connect plan.""" |
| is_metadata_column: builtins.bool |
| """(Optional) The requested column is a metadata column.""" |
| def __init__( |
| self, |
| *, |
| unparsed_identifier: builtins.str = ..., |
| plan_id: builtins.int | None = ..., |
| is_metadata_column: builtins.bool | None = ..., |
| ) -> None: ... |
| def HasField( |
| self, |
| field_name: typing_extensions.Literal[ |
| "_is_metadata_column", |
| b"_is_metadata_column", |
| "_plan_id", |
| b"_plan_id", |
| "is_metadata_column", |
| b"is_metadata_column", |
| "plan_id", |
| b"plan_id", |
| ], |
| ) -> builtins.bool: ... |
| def ClearField( |
| self, |
| field_name: typing_extensions.Literal[ |
| "_is_metadata_column", |
| b"_is_metadata_column", |
| "_plan_id", |
| b"_plan_id", |
| "is_metadata_column", |
| b"is_metadata_column", |
| "plan_id", |
| b"plan_id", |
| "unparsed_identifier", |
| b"unparsed_identifier", |
| ], |
| ) -> None: ... |
| @typing.overload |
| def WhichOneof( |
| self, |
| oneof_group: typing_extensions.Literal["_is_metadata_column", b"_is_metadata_column"], |
| ) -> typing_extensions.Literal["is_metadata_column"] | None: ... |
| @typing.overload |
| def WhichOneof( |
| self, oneof_group: typing_extensions.Literal["_plan_id", b"_plan_id"] |
| ) -> typing_extensions.Literal["plan_id"] | None: ... |
| |
| class UnresolvedFunction(google.protobuf.message.Message): |
| """An unresolved function is not explicitly bound to one explicit function, but the function |
| is resolved during analysis following Sparks name resolution rules. |
| """ |
| |
| DESCRIPTOR: google.protobuf.descriptor.Descriptor |
| |
| FUNCTION_NAME_FIELD_NUMBER: builtins.int |
| ARGUMENTS_FIELD_NUMBER: builtins.int |
| IS_DISTINCT_FIELD_NUMBER: builtins.int |
| IS_USER_DEFINED_FUNCTION_FIELD_NUMBER: builtins.int |
| IS_INTERNAL_FIELD_NUMBER: builtins.int |
| function_name: builtins.str |
| """(Required) name (or unparsed name for user defined function) for the unresolved function.""" |
| @property |
| def arguments( |
| self, |
| ) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[ |
| global___Expression |
| ]: |
| """(Optional) Function arguments. Empty arguments are allowed.""" |
| is_distinct: builtins.bool |
| """(Required) Indicate if this function should be applied on distinct values.""" |
| is_user_defined_function: builtins.bool |
| """(Required) Indicate if this is a user defined function. |
| |
| When it is not a user defined function, Connect will use the function name directly. |
| When it is a user defined function, Connect will parse the function name first. |
| """ |
| is_internal: builtins.bool |
| """(Optional) Indicate if this function is defined in the internal function registry. |
| If not set, the server will try to look up the function in the internal function registry |
| and decide appropriately. |
| """ |
| def __init__( |
| self, |
| *, |
| function_name: builtins.str = ..., |
| arguments: collections.abc.Iterable[global___Expression] | None = ..., |
| is_distinct: builtins.bool = ..., |
| is_user_defined_function: builtins.bool = ..., |
| is_internal: builtins.bool | None = ..., |
| ) -> None: ... |
| def HasField( |
| self, |
| field_name: typing_extensions.Literal[ |
| "_is_internal", b"_is_internal", "is_internal", b"is_internal" |
| ], |
| ) -> builtins.bool: ... |
| def ClearField( |
| self, |
| field_name: typing_extensions.Literal[ |
| "_is_internal", |
| b"_is_internal", |
| "arguments", |
| b"arguments", |
| "function_name", |
| b"function_name", |
| "is_distinct", |
| b"is_distinct", |
| "is_internal", |
| b"is_internal", |
| "is_user_defined_function", |
| b"is_user_defined_function", |
| ], |
| ) -> None: ... |
| def WhichOneof( |
| self, oneof_group: typing_extensions.Literal["_is_internal", b"_is_internal"] |
| ) -> typing_extensions.Literal["is_internal"] | None: ... |
| |
| class ExpressionString(google.protobuf.message.Message): |
| """Expression as string.""" |
| |
| DESCRIPTOR: google.protobuf.descriptor.Descriptor |
| |
| EXPRESSION_FIELD_NUMBER: builtins.int |
| expression: builtins.str |
| """(Required) A SQL expression that will be parsed by Catalyst parser.""" |
| def __init__( |
| self, |
| *, |
| expression: builtins.str = ..., |
| ) -> None: ... |
| def ClearField( |
| self, field_name: typing_extensions.Literal["expression", b"expression"] |
| ) -> None: ... |
| |
| class UnresolvedStar(google.protobuf.message.Message): |
| """UnresolvedStar is used to expand all the fields of a relation or struct.""" |
| |
| DESCRIPTOR: google.protobuf.descriptor.Descriptor |
| |
| UNPARSED_TARGET_FIELD_NUMBER: builtins.int |
| PLAN_ID_FIELD_NUMBER: builtins.int |
| unparsed_target: builtins.str |
| """(Optional) The target of the expansion. |
| |
| If set, it should end with '.*' and will be parsed by 'parseAttributeName' |
| in the server side. |
| """ |
| plan_id: builtins.int |
| """(Optional) The id of corresponding connect plan.""" |
| def __init__( |
| self, |
| *, |
| unparsed_target: builtins.str | None = ..., |
| plan_id: builtins.int | None = ..., |
| ) -> None: ... |
| def HasField( |
| self, |
| field_name: typing_extensions.Literal[ |
| "_plan_id", |
| b"_plan_id", |
| "_unparsed_target", |
| b"_unparsed_target", |
| "plan_id", |
| b"plan_id", |
| "unparsed_target", |
| b"unparsed_target", |
| ], |
| ) -> builtins.bool: ... |
| def ClearField( |
| self, |
| field_name: typing_extensions.Literal[ |
| "_plan_id", |
| b"_plan_id", |
| "_unparsed_target", |
| b"_unparsed_target", |
| "plan_id", |
| b"plan_id", |
| "unparsed_target", |
| b"unparsed_target", |
| ], |
| ) -> None: ... |
| @typing.overload |
| def WhichOneof( |
| self, oneof_group: typing_extensions.Literal["_plan_id", b"_plan_id"] |
| ) -> typing_extensions.Literal["plan_id"] | None: ... |
| @typing.overload |
| def WhichOneof( |
| self, oneof_group: typing_extensions.Literal["_unparsed_target", b"_unparsed_target"] |
| ) -> typing_extensions.Literal["unparsed_target"] | None: ... |
| |
| class UnresolvedRegex(google.protobuf.message.Message): |
| """Represents all of the input attributes to a given relational operator, for example in |
| "SELECT `(id)?+.+` FROM ...". |
| """ |
| |
| DESCRIPTOR: google.protobuf.descriptor.Descriptor |
| |
| COL_NAME_FIELD_NUMBER: builtins.int |
| PLAN_ID_FIELD_NUMBER: builtins.int |
| col_name: builtins.str |
| """(Required) The column name used to extract column with regex.""" |
| plan_id: builtins.int |
| """(Optional) The id of corresponding connect plan.""" |
| def __init__( |
| self, |
| *, |
| col_name: builtins.str = ..., |
| plan_id: builtins.int | None = ..., |
| ) -> None: ... |
| def HasField( |
| self, |
| field_name: typing_extensions.Literal["_plan_id", b"_plan_id", "plan_id", b"plan_id"], |
| ) -> builtins.bool: ... |
| def ClearField( |
| self, |
| field_name: typing_extensions.Literal[ |
| "_plan_id", b"_plan_id", "col_name", b"col_name", "plan_id", b"plan_id" |
| ], |
| ) -> None: ... |
| def WhichOneof( |
| self, oneof_group: typing_extensions.Literal["_plan_id", b"_plan_id"] |
| ) -> typing_extensions.Literal["plan_id"] | None: ... |
| |
| class UnresolvedExtractValue(google.protobuf.message.Message): |
| """Extracts a value or values from an Expression""" |
| |
| DESCRIPTOR: google.protobuf.descriptor.Descriptor |
| |
| CHILD_FIELD_NUMBER: builtins.int |
| EXTRACTION_FIELD_NUMBER: builtins.int |
| @property |
| def child(self) -> global___Expression: |
| """(Required) The expression to extract value from, can be |
| Map, Array, Struct or array of Structs. |
| """ |
| @property |
| def extraction(self) -> global___Expression: |
| """(Required) The expression to describe the extraction, can be |
| key of Map, index of Array, field name of Struct. |
| """ |
| def __init__( |
| self, |
| *, |
| child: global___Expression | None = ..., |
| extraction: global___Expression | None = ..., |
| ) -> None: ... |
| def HasField( |
| self, |
| field_name: typing_extensions.Literal["child", b"child", "extraction", b"extraction"], |
| ) -> builtins.bool: ... |
| def ClearField( |
| self, |
| field_name: typing_extensions.Literal["child", b"child", "extraction", b"extraction"], |
| ) -> None: ... |
| |
| class UpdateFields(google.protobuf.message.Message): |
| """Add, replace or drop a field of `StructType` expression by name.""" |
| |
| DESCRIPTOR: google.protobuf.descriptor.Descriptor |
| |
| STRUCT_EXPRESSION_FIELD_NUMBER: builtins.int |
| FIELD_NAME_FIELD_NUMBER: builtins.int |
| VALUE_EXPRESSION_FIELD_NUMBER: builtins.int |
| @property |
| def struct_expression(self) -> global___Expression: |
| """(Required) The struct expression.""" |
| field_name: builtins.str |
| """(Required) The field name.""" |
| @property |
| def value_expression(self) -> global___Expression: |
| """(Optional) The expression to add or replace. |
| |
| When not set, it means this field will be dropped. |
| """ |
| def __init__( |
| self, |
| *, |
| struct_expression: global___Expression | None = ..., |
| field_name: builtins.str = ..., |
| value_expression: global___Expression | None = ..., |
| ) -> None: ... |
| def HasField( |
| self, |
| field_name: typing_extensions.Literal[ |
| "struct_expression", b"struct_expression", "value_expression", b"value_expression" |
| ], |
| ) -> builtins.bool: ... |
| def ClearField( |
| self, |
| field_name: typing_extensions.Literal[ |
| "field_name", |
| b"field_name", |
| "struct_expression", |
| b"struct_expression", |
| "value_expression", |
| b"value_expression", |
| ], |
| ) -> None: ... |
| |
| class Alias(google.protobuf.message.Message): |
| DESCRIPTOR: google.protobuf.descriptor.Descriptor |
| |
| EXPR_FIELD_NUMBER: builtins.int |
| NAME_FIELD_NUMBER: builtins.int |
| METADATA_FIELD_NUMBER: builtins.int |
| @property |
| def expr(self) -> global___Expression: |
| """(Required) The expression that alias will be added on.""" |
| @property |
| def name( |
| self, |
| ) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[builtins.str]: |
| """(Required) a list of name parts for the alias. |
| |
| Scalar columns only has one name that presents. |
| """ |
| metadata: builtins.str |
| """(Optional) Alias metadata expressed as a JSON map.""" |
| def __init__( |
| self, |
| *, |
| expr: global___Expression | None = ..., |
| name: collections.abc.Iterable[builtins.str] | None = ..., |
| metadata: builtins.str | None = ..., |
| ) -> None: ... |
| def HasField( |
| self, |
| field_name: typing_extensions.Literal[ |
| "_metadata", b"_metadata", "expr", b"expr", "metadata", b"metadata" |
| ], |
| ) -> builtins.bool: ... |
| def ClearField( |
| self, |
| field_name: typing_extensions.Literal[ |
| "_metadata", b"_metadata", "expr", b"expr", "metadata", b"metadata", "name", b"name" |
| ], |
| ) -> None: ... |
| def WhichOneof( |
| self, oneof_group: typing_extensions.Literal["_metadata", b"_metadata"] |
| ) -> typing_extensions.Literal["metadata"] | None: ... |
| |
| class LambdaFunction(google.protobuf.message.Message): |
| DESCRIPTOR: google.protobuf.descriptor.Descriptor |
| |
| FUNCTION_FIELD_NUMBER: builtins.int |
| ARGUMENTS_FIELD_NUMBER: builtins.int |
| @property |
| def function(self) -> global___Expression: |
| """(Required) The lambda function. |
| |
| The function body should use 'UnresolvedAttribute' as arguments, the sever side will |
| replace 'UnresolvedAttribute' with 'UnresolvedNamedLambdaVariable'. |
| """ |
| @property |
| def arguments( |
| self, |
| ) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[ |
| global___Expression.UnresolvedNamedLambdaVariable |
| ]: |
| """(Required) Function variables. Must contains 1 ~ 3 variables.""" |
| def __init__( |
| self, |
| *, |
| function: global___Expression | None = ..., |
| arguments: collections.abc.Iterable[global___Expression.UnresolvedNamedLambdaVariable] |
| | None = ..., |
| ) -> None: ... |
| def HasField( |
| self, field_name: typing_extensions.Literal["function", b"function"] |
| ) -> builtins.bool: ... |
| def ClearField( |
| self, |
| field_name: typing_extensions.Literal[ |
| "arguments", b"arguments", "function", b"function" |
| ], |
| ) -> None: ... |
| |
| class UnresolvedNamedLambdaVariable(google.protobuf.message.Message): |
| DESCRIPTOR: google.protobuf.descriptor.Descriptor |
| |
| NAME_PARTS_FIELD_NUMBER: builtins.int |
| @property |
| def name_parts( |
| self, |
| ) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[builtins.str]: |
| """(Required) a list of name parts for the variable. Must not be empty.""" |
| def __init__( |
| self, |
| *, |
| name_parts: collections.abc.Iterable[builtins.str] | None = ..., |
| ) -> None: ... |
| def ClearField( |
| self, field_name: typing_extensions.Literal["name_parts", b"name_parts"] |
| ) -> None: ... |
| |
| COMMON_FIELD_NUMBER: builtins.int |
| LITERAL_FIELD_NUMBER: builtins.int |
| UNRESOLVED_ATTRIBUTE_FIELD_NUMBER: builtins.int |
| UNRESOLVED_FUNCTION_FIELD_NUMBER: builtins.int |
| EXPRESSION_STRING_FIELD_NUMBER: builtins.int |
| UNRESOLVED_STAR_FIELD_NUMBER: builtins.int |
| ALIAS_FIELD_NUMBER: builtins.int |
| CAST_FIELD_NUMBER: builtins.int |
| UNRESOLVED_REGEX_FIELD_NUMBER: builtins.int |
| SORT_ORDER_FIELD_NUMBER: builtins.int |
| LAMBDA_FUNCTION_FIELD_NUMBER: builtins.int |
| WINDOW_FIELD_NUMBER: builtins.int |
| UNRESOLVED_EXTRACT_VALUE_FIELD_NUMBER: builtins.int |
| UPDATE_FIELDS_FIELD_NUMBER: builtins.int |
| UNRESOLVED_NAMED_LAMBDA_VARIABLE_FIELD_NUMBER: builtins.int |
| COMMON_INLINE_USER_DEFINED_FUNCTION_FIELD_NUMBER: builtins.int |
| CALL_FUNCTION_FIELD_NUMBER: builtins.int |
| NAMED_ARGUMENT_EXPRESSION_FIELD_NUMBER: builtins.int |
| MERGE_ACTION_FIELD_NUMBER: builtins.int |
| TYPED_AGGREGATE_EXPRESSION_FIELD_NUMBER: builtins.int |
| SUBQUERY_EXPRESSION_FIELD_NUMBER: builtins.int |
| DIRECT_SHUFFLE_PARTITION_ID_FIELD_NUMBER: builtins.int |
| EXTENSION_FIELD_NUMBER: builtins.int |
| @property |
| def common(self) -> global___ExpressionCommon: ... |
| @property |
| def literal(self) -> global___Expression.Literal: ... |
| @property |
| def unresolved_attribute(self) -> global___Expression.UnresolvedAttribute: ... |
| @property |
| def unresolved_function(self) -> global___Expression.UnresolvedFunction: ... |
| @property |
| def expression_string(self) -> global___Expression.ExpressionString: ... |
| @property |
| def unresolved_star(self) -> global___Expression.UnresolvedStar: ... |
| @property |
| def alias(self) -> global___Expression.Alias: ... |
| @property |
| def cast(self) -> global___Expression.Cast: ... |
| @property |
| def unresolved_regex(self) -> global___Expression.UnresolvedRegex: ... |
| @property |
| def sort_order(self) -> global___Expression.SortOrder: ... |
| @property |
| def lambda_function(self) -> global___Expression.LambdaFunction: ... |
| @property |
| def window(self) -> global___Expression.Window: ... |
| @property |
| def unresolved_extract_value(self) -> global___Expression.UnresolvedExtractValue: ... |
| @property |
| def update_fields(self) -> global___Expression.UpdateFields: ... |
| @property |
| def unresolved_named_lambda_variable( |
| self, |
| ) -> global___Expression.UnresolvedNamedLambdaVariable: ... |
| @property |
| def common_inline_user_defined_function(self) -> global___CommonInlineUserDefinedFunction: ... |
| @property |
| def call_function(self) -> global___CallFunction: ... |
| @property |
| def named_argument_expression(self) -> global___NamedArgumentExpression: ... |
| @property |
| def merge_action(self) -> global___MergeAction: ... |
| @property |
| def typed_aggregate_expression(self) -> global___TypedAggregateExpression: ... |
| @property |
| def subquery_expression(self) -> global___SubqueryExpression: ... |
| @property |
| def direct_shuffle_partition_id(self) -> global___Expression.DirectShufflePartitionID: ... |
| @property |
| def extension(self) -> google.protobuf.any_pb2.Any: |
| """This field is used to mark extensions to the protocol. When plugins generate arbitrary |
| relations they can add them here. During the planning the correct resolution is done. |
| """ |
| def __init__( |
| self, |
| *, |
| common: global___ExpressionCommon | None = ..., |
| literal: global___Expression.Literal | None = ..., |
| unresolved_attribute: global___Expression.UnresolvedAttribute | None = ..., |
| unresolved_function: global___Expression.UnresolvedFunction | None = ..., |
| expression_string: global___Expression.ExpressionString | None = ..., |
| unresolved_star: global___Expression.UnresolvedStar | None = ..., |
| alias: global___Expression.Alias | None = ..., |
| cast: global___Expression.Cast | None = ..., |
| unresolved_regex: global___Expression.UnresolvedRegex | None = ..., |
| sort_order: global___Expression.SortOrder | None = ..., |
| lambda_function: global___Expression.LambdaFunction | None = ..., |
| window: global___Expression.Window | None = ..., |
| unresolved_extract_value: global___Expression.UnresolvedExtractValue | None = ..., |
| update_fields: global___Expression.UpdateFields | None = ..., |
| unresolved_named_lambda_variable: global___Expression.UnresolvedNamedLambdaVariable |
| | None = ..., |
| common_inline_user_defined_function: global___CommonInlineUserDefinedFunction | None = ..., |
| call_function: global___CallFunction | None = ..., |
| named_argument_expression: global___NamedArgumentExpression | None = ..., |
| merge_action: global___MergeAction | None = ..., |
| typed_aggregate_expression: global___TypedAggregateExpression | None = ..., |
| subquery_expression: global___SubqueryExpression | None = ..., |
| direct_shuffle_partition_id: global___Expression.DirectShufflePartitionID | None = ..., |
| extension: google.protobuf.any_pb2.Any | None = ..., |
| ) -> None: ... |
| def HasField( |
| self, |
| field_name: typing_extensions.Literal[ |
| "alias", |
| b"alias", |
| "call_function", |
| b"call_function", |
| "cast", |
| b"cast", |
| "common", |
| b"common", |
| "common_inline_user_defined_function", |
| b"common_inline_user_defined_function", |
| "direct_shuffle_partition_id", |
| b"direct_shuffle_partition_id", |
| "expr_type", |
| b"expr_type", |
| "expression_string", |
| b"expression_string", |
| "extension", |
| b"extension", |
| "lambda_function", |
| b"lambda_function", |
| "literal", |
| b"literal", |
| "merge_action", |
| b"merge_action", |
| "named_argument_expression", |
| b"named_argument_expression", |
| "sort_order", |
| b"sort_order", |
| "subquery_expression", |
| b"subquery_expression", |
| "typed_aggregate_expression", |
| b"typed_aggregate_expression", |
| "unresolved_attribute", |
| b"unresolved_attribute", |
| "unresolved_extract_value", |
| b"unresolved_extract_value", |
| "unresolved_function", |
| b"unresolved_function", |
| "unresolved_named_lambda_variable", |
| b"unresolved_named_lambda_variable", |
| "unresolved_regex", |
| b"unresolved_regex", |
| "unresolved_star", |
| b"unresolved_star", |
| "update_fields", |
| b"update_fields", |
| "window", |
| b"window", |
| ], |
| ) -> builtins.bool: ... |
| def ClearField( |
| self, |
| field_name: typing_extensions.Literal[ |
| "alias", |
| b"alias", |
| "call_function", |
| b"call_function", |
| "cast", |
| b"cast", |
| "common", |
| b"common", |
| "common_inline_user_defined_function", |
| b"common_inline_user_defined_function", |
| "direct_shuffle_partition_id", |
| b"direct_shuffle_partition_id", |
| "expr_type", |
| b"expr_type", |
| "expression_string", |
| b"expression_string", |
| "extension", |
| b"extension", |
| "lambda_function", |
| b"lambda_function", |
| "literal", |
| b"literal", |
| "merge_action", |
| b"merge_action", |
| "named_argument_expression", |
| b"named_argument_expression", |
| "sort_order", |
| b"sort_order", |
| "subquery_expression", |
| b"subquery_expression", |
| "typed_aggregate_expression", |
| b"typed_aggregate_expression", |
| "unresolved_attribute", |
| b"unresolved_attribute", |
| "unresolved_extract_value", |
| b"unresolved_extract_value", |
| "unresolved_function", |
| b"unresolved_function", |
| "unresolved_named_lambda_variable", |
| b"unresolved_named_lambda_variable", |
| "unresolved_regex", |
| b"unresolved_regex", |
| "unresolved_star", |
| b"unresolved_star", |
| "update_fields", |
| b"update_fields", |
| "window", |
| b"window", |
| ], |
| ) -> None: ... |
| def WhichOneof( |
| self, oneof_group: typing_extensions.Literal["expr_type", b"expr_type"] |
| ) -> ( |
| typing_extensions.Literal[ |
| "literal", |
| "unresolved_attribute", |
| "unresolved_function", |
| "expression_string", |
| "unresolved_star", |
| "alias", |
| "cast", |
| "unresolved_regex", |
| "sort_order", |
| "lambda_function", |
| "window", |
| "unresolved_extract_value", |
| "update_fields", |
| "unresolved_named_lambda_variable", |
| "common_inline_user_defined_function", |
| "call_function", |
| "named_argument_expression", |
| "merge_action", |
| "typed_aggregate_expression", |
| "subquery_expression", |
| "direct_shuffle_partition_id", |
| "extension", |
| ] |
| | None |
| ): ... |
| |
| global___Expression = Expression |
| |
| class ExpressionCommon(google.protobuf.message.Message): |
| DESCRIPTOR: google.protobuf.descriptor.Descriptor |
| |
| ORIGIN_FIELD_NUMBER: builtins.int |
| @property |
| def origin(self) -> pyspark.sql.connect.proto.common_pb2.Origin: |
| """(Required) Keep the information of the origin for this expression such as stacktrace.""" |
| def __init__( |
| self, |
| *, |
| origin: pyspark.sql.connect.proto.common_pb2.Origin | None = ..., |
| ) -> None: ... |
| def HasField( |
| self, field_name: typing_extensions.Literal["origin", b"origin"] |
| ) -> builtins.bool: ... |
| def ClearField(self, field_name: typing_extensions.Literal["origin", b"origin"]) -> None: ... |
| |
| global___ExpressionCommon = ExpressionCommon |
| |
| class CommonInlineUserDefinedFunction(google.protobuf.message.Message): |
| DESCRIPTOR: google.protobuf.descriptor.Descriptor |
| |
| FUNCTION_NAME_FIELD_NUMBER: builtins.int |
| DETERMINISTIC_FIELD_NUMBER: builtins.int |
| ARGUMENTS_FIELD_NUMBER: builtins.int |
| PYTHON_UDF_FIELD_NUMBER: builtins.int |
| SCALAR_SCALA_UDF_FIELD_NUMBER: builtins.int |
| JAVA_UDF_FIELD_NUMBER: builtins.int |
| IS_DISTINCT_FIELD_NUMBER: builtins.int |
| function_name: builtins.str |
| """(Required) Name of the user-defined function.""" |
| deterministic: builtins.bool |
| """(Optional) Indicate if the user-defined function is deterministic.""" |
| @property |
| def arguments( |
| self, |
| ) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[global___Expression]: |
| """(Optional) Function arguments. Empty arguments are allowed.""" |
| @property |
| def python_udf(self) -> global___PythonUDF: ... |
| @property |
| def scalar_scala_udf(self) -> global___ScalarScalaUDF: ... |
| @property |
| def java_udf(self) -> global___JavaUDF: ... |
| is_distinct: builtins.bool |
| """(Required) Indicate if this function should be applied on distinct values.""" |
| def __init__( |
| self, |
| *, |
| function_name: builtins.str = ..., |
| deterministic: builtins.bool = ..., |
| arguments: collections.abc.Iterable[global___Expression] | None = ..., |
| python_udf: global___PythonUDF | None = ..., |
| scalar_scala_udf: global___ScalarScalaUDF | None = ..., |
| java_udf: global___JavaUDF | None = ..., |
| is_distinct: builtins.bool = ..., |
| ) -> None: ... |
| def HasField( |
| self, |
| field_name: typing_extensions.Literal[ |
| "function", |
| b"function", |
| "java_udf", |
| b"java_udf", |
| "python_udf", |
| b"python_udf", |
| "scalar_scala_udf", |
| b"scalar_scala_udf", |
| ], |
| ) -> builtins.bool: ... |
| def ClearField( |
| self, |
| field_name: typing_extensions.Literal[ |
| "arguments", |
| b"arguments", |
| "deterministic", |
| b"deterministic", |
| "function", |
| b"function", |
| "function_name", |
| b"function_name", |
| "is_distinct", |
| b"is_distinct", |
| "java_udf", |
| b"java_udf", |
| "python_udf", |
| b"python_udf", |
| "scalar_scala_udf", |
| b"scalar_scala_udf", |
| ], |
| ) -> None: ... |
| def WhichOneof( |
| self, oneof_group: typing_extensions.Literal["function", b"function"] |
| ) -> typing_extensions.Literal["python_udf", "scalar_scala_udf", "java_udf"] | None: ... |
| |
| global___CommonInlineUserDefinedFunction = CommonInlineUserDefinedFunction |
| |
| class PythonUDF(google.protobuf.message.Message): |
| DESCRIPTOR: google.protobuf.descriptor.Descriptor |
| |
| OUTPUT_TYPE_FIELD_NUMBER: builtins.int |
| EVAL_TYPE_FIELD_NUMBER: builtins.int |
| COMMAND_FIELD_NUMBER: builtins.int |
| PYTHON_VER_FIELD_NUMBER: builtins.int |
| ADDITIONAL_INCLUDES_FIELD_NUMBER: builtins.int |
| @property |
| def output_type(self) -> pyspark.sql.connect.proto.types_pb2.DataType: |
| """(Required) Output type of the Python UDF""" |
| eval_type: builtins.int |
| """(Required) EvalType of the Python UDF""" |
| command: builtins.bytes |
| """(Required) The encoded commands of the Python UDF""" |
| python_ver: builtins.str |
| """(Required) Python version being used in the client.""" |
| @property |
| def additional_includes( |
| self, |
| ) -> google.protobuf.internal.containers.RepeatedScalarFieldContainer[builtins.str]: |
| """(Optional) Additional includes for the Python UDF.""" |
| def __init__( |
| self, |
| *, |
| output_type: pyspark.sql.connect.proto.types_pb2.DataType | None = ..., |
| eval_type: builtins.int = ..., |
| command: builtins.bytes = ..., |
| python_ver: builtins.str = ..., |
| additional_includes: collections.abc.Iterable[builtins.str] | None = ..., |
| ) -> None: ... |
| def HasField( |
| self, field_name: typing_extensions.Literal["output_type", b"output_type"] |
| ) -> builtins.bool: ... |
| def ClearField( |
| self, |
| field_name: typing_extensions.Literal[ |
| "additional_includes", |
| b"additional_includes", |
| "command", |
| b"command", |
| "eval_type", |
| b"eval_type", |
| "output_type", |
| b"output_type", |
| "python_ver", |
| b"python_ver", |
| ], |
| ) -> None: ... |
| |
| global___PythonUDF = PythonUDF |
| |
| class ScalarScalaUDF(google.protobuf.message.Message): |
| DESCRIPTOR: google.protobuf.descriptor.Descriptor |
| |
| PAYLOAD_FIELD_NUMBER: builtins.int |
| INPUTTYPES_FIELD_NUMBER: builtins.int |
| OUTPUTTYPE_FIELD_NUMBER: builtins.int |
| NULLABLE_FIELD_NUMBER: builtins.int |
| AGGREGATE_FIELD_NUMBER: builtins.int |
| payload: builtins.bytes |
| """(Required) Serialized JVM object containing UDF definition, input encoders and output encoder""" |
| @property |
| def inputTypes( |
| self, |
| ) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[ |
| pyspark.sql.connect.proto.types_pb2.DataType |
| ]: |
| """(Optional) Input type(s) of the UDF""" |
| @property |
| def outputType(self) -> pyspark.sql.connect.proto.types_pb2.DataType: |
| """(Required) Output type of the UDF""" |
| nullable: builtins.bool |
| """(Required) True if the UDF can return null value""" |
| aggregate: builtins.bool |
| """(Required) Indicate if the UDF is an aggregate function""" |
| def __init__( |
| self, |
| *, |
| payload: builtins.bytes = ..., |
| inputTypes: collections.abc.Iterable[pyspark.sql.connect.proto.types_pb2.DataType] |
| | None = ..., |
| outputType: pyspark.sql.connect.proto.types_pb2.DataType | None = ..., |
| nullable: builtins.bool = ..., |
| aggregate: builtins.bool = ..., |
| ) -> None: ... |
| def HasField( |
| self, field_name: typing_extensions.Literal["outputType", b"outputType"] |
| ) -> builtins.bool: ... |
| def ClearField( |
| self, |
| field_name: typing_extensions.Literal[ |
| "aggregate", |
| b"aggregate", |
| "inputTypes", |
| b"inputTypes", |
| "nullable", |
| b"nullable", |
| "outputType", |
| b"outputType", |
| "payload", |
| b"payload", |
| ], |
| ) -> None: ... |
| |
| global___ScalarScalaUDF = ScalarScalaUDF |
| |
| class JavaUDF(google.protobuf.message.Message): |
| DESCRIPTOR: google.protobuf.descriptor.Descriptor |
| |
| CLASS_NAME_FIELD_NUMBER: builtins.int |
| OUTPUT_TYPE_FIELD_NUMBER: builtins.int |
| AGGREGATE_FIELD_NUMBER: builtins.int |
| class_name: builtins.str |
| """(Required) Fully qualified name of Java class""" |
| @property |
| def output_type(self) -> pyspark.sql.connect.proto.types_pb2.DataType: |
| """(Optional) Output type of the Java UDF""" |
| aggregate: builtins.bool |
| """(Required) Indicate if the Java user-defined function is an aggregate function""" |
| def __init__( |
| self, |
| *, |
| class_name: builtins.str = ..., |
| output_type: pyspark.sql.connect.proto.types_pb2.DataType | None = ..., |
| aggregate: builtins.bool = ..., |
| ) -> None: ... |
| def HasField( |
| self, |
| field_name: typing_extensions.Literal[ |
| "_output_type", b"_output_type", "output_type", b"output_type" |
| ], |
| ) -> builtins.bool: ... |
| def ClearField( |
| self, |
| field_name: typing_extensions.Literal[ |
| "_output_type", |
| b"_output_type", |
| "aggregate", |
| b"aggregate", |
| "class_name", |
| b"class_name", |
| "output_type", |
| b"output_type", |
| ], |
| ) -> None: ... |
| def WhichOneof( |
| self, oneof_group: typing_extensions.Literal["_output_type", b"_output_type"] |
| ) -> typing_extensions.Literal["output_type"] | None: ... |
| |
| global___JavaUDF = JavaUDF |
| |
| class TypedAggregateExpression(google.protobuf.message.Message): |
| DESCRIPTOR: google.protobuf.descriptor.Descriptor |
| |
| SCALAR_SCALA_UDF_FIELD_NUMBER: builtins.int |
| @property |
| def scalar_scala_udf(self) -> global___ScalarScalaUDF: |
| """(Required) The aggregate function object packed into bytes.""" |
| def __init__( |
| self, |
| *, |
| scalar_scala_udf: global___ScalarScalaUDF | None = ..., |
| ) -> None: ... |
| def HasField( |
| self, field_name: typing_extensions.Literal["scalar_scala_udf", b"scalar_scala_udf"] |
| ) -> builtins.bool: ... |
| def ClearField( |
| self, field_name: typing_extensions.Literal["scalar_scala_udf", b"scalar_scala_udf"] |
| ) -> None: ... |
| |
| global___TypedAggregateExpression = TypedAggregateExpression |
| |
| class CallFunction(google.protobuf.message.Message): |
| DESCRIPTOR: google.protobuf.descriptor.Descriptor |
| |
| FUNCTION_NAME_FIELD_NUMBER: builtins.int |
| ARGUMENTS_FIELD_NUMBER: builtins.int |
| function_name: builtins.str |
| """(Required) Unparsed name of the SQL function.""" |
| @property |
| def arguments( |
| self, |
| ) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[global___Expression]: |
| """(Optional) Function arguments. Empty arguments are allowed.""" |
| def __init__( |
| self, |
| *, |
| function_name: builtins.str = ..., |
| arguments: collections.abc.Iterable[global___Expression] | None = ..., |
| ) -> None: ... |
| def ClearField( |
| self, |
| field_name: typing_extensions.Literal[ |
| "arguments", b"arguments", "function_name", b"function_name" |
| ], |
| ) -> None: ... |
| |
| global___CallFunction = CallFunction |
| |
| class NamedArgumentExpression(google.protobuf.message.Message): |
| DESCRIPTOR: google.protobuf.descriptor.Descriptor |
| |
| KEY_FIELD_NUMBER: builtins.int |
| VALUE_FIELD_NUMBER: builtins.int |
| key: builtins.str |
| """(Required) The key of the named argument.""" |
| @property |
| def value(self) -> global___Expression: |
| """(Required) The value expression of the named argument.""" |
| def __init__( |
| self, |
| *, |
| key: builtins.str = ..., |
| value: global___Expression | None = ..., |
| ) -> None: ... |
| def HasField( |
| self, field_name: typing_extensions.Literal["value", b"value"] |
| ) -> builtins.bool: ... |
| def ClearField( |
| self, field_name: typing_extensions.Literal["key", b"key", "value", b"value"] |
| ) -> None: ... |
| |
| global___NamedArgumentExpression = NamedArgumentExpression |
| |
| class MergeAction(google.protobuf.message.Message): |
| DESCRIPTOR: google.protobuf.descriptor.Descriptor |
| |
| class _ActionType: |
| ValueType = typing.NewType("ValueType", builtins.int) |
| V: typing_extensions.TypeAlias = ValueType |
| |
| class _ActionTypeEnumTypeWrapper( |
| google.protobuf.internal.enum_type_wrapper._EnumTypeWrapper[ |
| MergeAction._ActionType.ValueType |
| ], |
| builtins.type, |
| ): # noqa: F821 |
| DESCRIPTOR: google.protobuf.descriptor.EnumDescriptor |
| ACTION_TYPE_INVALID: MergeAction._ActionType.ValueType # 0 |
| ACTION_TYPE_DELETE: MergeAction._ActionType.ValueType # 1 |
| ACTION_TYPE_INSERT: MergeAction._ActionType.ValueType # 2 |
| ACTION_TYPE_INSERT_STAR: MergeAction._ActionType.ValueType # 3 |
| ACTION_TYPE_UPDATE: MergeAction._ActionType.ValueType # 4 |
| ACTION_TYPE_UPDATE_STAR: MergeAction._ActionType.ValueType # 5 |
| |
| class ActionType(_ActionType, metaclass=_ActionTypeEnumTypeWrapper): ... |
| ACTION_TYPE_INVALID: MergeAction.ActionType.ValueType # 0 |
| ACTION_TYPE_DELETE: MergeAction.ActionType.ValueType # 1 |
| ACTION_TYPE_INSERT: MergeAction.ActionType.ValueType # 2 |
| ACTION_TYPE_INSERT_STAR: MergeAction.ActionType.ValueType # 3 |
| ACTION_TYPE_UPDATE: MergeAction.ActionType.ValueType # 4 |
| ACTION_TYPE_UPDATE_STAR: MergeAction.ActionType.ValueType # 5 |
| |
| class Assignment(google.protobuf.message.Message): |
| DESCRIPTOR: google.protobuf.descriptor.Descriptor |
| |
| KEY_FIELD_NUMBER: builtins.int |
| VALUE_FIELD_NUMBER: builtins.int |
| @property |
| def key(self) -> global___Expression: |
| """(Required) The key of the assignment.""" |
| @property |
| def value(self) -> global___Expression: |
| """(Required) The value of the assignment.""" |
| def __init__( |
| self, |
| *, |
| key: global___Expression | None = ..., |
| value: global___Expression | None = ..., |
| ) -> None: ... |
| def HasField( |
| self, field_name: typing_extensions.Literal["key", b"key", "value", b"value"] |
| ) -> builtins.bool: ... |
| def ClearField( |
| self, field_name: typing_extensions.Literal["key", b"key", "value", b"value"] |
| ) -> None: ... |
| |
| ACTION_TYPE_FIELD_NUMBER: builtins.int |
| CONDITION_FIELD_NUMBER: builtins.int |
| ASSIGNMENTS_FIELD_NUMBER: builtins.int |
| action_type: global___MergeAction.ActionType.ValueType |
| """(Required) The action type of the merge action.""" |
| @property |
| def condition(self) -> global___Expression: |
| """(Optional) The condition expression of the merge action.""" |
| @property |
| def assignments( |
| self, |
| ) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[ |
| global___MergeAction.Assignment |
| ]: |
| """(Optional) The assignments of the merge action. Required for ActionTypes INSERT and UPDATE.""" |
| def __init__( |
| self, |
| *, |
| action_type: global___MergeAction.ActionType.ValueType = ..., |
| condition: global___Expression | None = ..., |
| assignments: collections.abc.Iterable[global___MergeAction.Assignment] | None = ..., |
| ) -> None: ... |
| def HasField( |
| self, |
| field_name: typing_extensions.Literal[ |
| "_condition", b"_condition", "condition", b"condition" |
| ], |
| ) -> builtins.bool: ... |
| def ClearField( |
| self, |
| field_name: typing_extensions.Literal[ |
| "_condition", |
| b"_condition", |
| "action_type", |
| b"action_type", |
| "assignments", |
| b"assignments", |
| "condition", |
| b"condition", |
| ], |
| ) -> None: ... |
| def WhichOneof( |
| self, oneof_group: typing_extensions.Literal["_condition", b"_condition"] |
| ) -> typing_extensions.Literal["condition"] | None: ... |
| |
| global___MergeAction = MergeAction |
| |
| class SubqueryExpression(google.protobuf.message.Message): |
| DESCRIPTOR: google.protobuf.descriptor.Descriptor |
| |
| class _SubqueryType: |
| ValueType = typing.NewType("ValueType", builtins.int) |
| V: typing_extensions.TypeAlias = ValueType |
| |
| class _SubqueryTypeEnumTypeWrapper( |
| google.protobuf.internal.enum_type_wrapper._EnumTypeWrapper[ |
| SubqueryExpression._SubqueryType.ValueType |
| ], |
| builtins.type, |
| ): # noqa: F821 |
| DESCRIPTOR: google.protobuf.descriptor.EnumDescriptor |
| SUBQUERY_TYPE_UNKNOWN: SubqueryExpression._SubqueryType.ValueType # 0 |
| SUBQUERY_TYPE_SCALAR: SubqueryExpression._SubqueryType.ValueType # 1 |
| SUBQUERY_TYPE_EXISTS: SubqueryExpression._SubqueryType.ValueType # 2 |
| SUBQUERY_TYPE_TABLE_ARG: SubqueryExpression._SubqueryType.ValueType # 3 |
| SUBQUERY_TYPE_IN: SubqueryExpression._SubqueryType.ValueType # 4 |
| |
| class SubqueryType(_SubqueryType, metaclass=_SubqueryTypeEnumTypeWrapper): ... |
| SUBQUERY_TYPE_UNKNOWN: SubqueryExpression.SubqueryType.ValueType # 0 |
| SUBQUERY_TYPE_SCALAR: SubqueryExpression.SubqueryType.ValueType # 1 |
| SUBQUERY_TYPE_EXISTS: SubqueryExpression.SubqueryType.ValueType # 2 |
| SUBQUERY_TYPE_TABLE_ARG: SubqueryExpression.SubqueryType.ValueType # 3 |
| SUBQUERY_TYPE_IN: SubqueryExpression.SubqueryType.ValueType # 4 |
| |
| class TableArgOptions(google.protobuf.message.Message): |
| """Nested message for table argument options.""" |
| |
| DESCRIPTOR: google.protobuf.descriptor.Descriptor |
| |
| PARTITION_SPEC_FIELD_NUMBER: builtins.int |
| ORDER_SPEC_FIELD_NUMBER: builtins.int |
| WITH_SINGLE_PARTITION_FIELD_NUMBER: builtins.int |
| @property |
| def partition_spec( |
| self, |
| ) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[ |
| global___Expression |
| ]: |
| """(Optional) The way that input rows are partitioned.""" |
| @property |
| def order_spec( |
| self, |
| ) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[ |
| global___Expression.SortOrder |
| ]: |
| """(Optional) Ordering of rows in a partition.""" |
| with_single_partition: builtins.bool |
| """(Optional) Whether this is a single partition.""" |
| def __init__( |
| self, |
| *, |
| partition_spec: collections.abc.Iterable[global___Expression] | None = ..., |
| order_spec: collections.abc.Iterable[global___Expression.SortOrder] | None = ..., |
| with_single_partition: builtins.bool | None = ..., |
| ) -> None: ... |
| def HasField( |
| self, |
| field_name: typing_extensions.Literal[ |
| "_with_single_partition", |
| b"_with_single_partition", |
| "with_single_partition", |
| b"with_single_partition", |
| ], |
| ) -> builtins.bool: ... |
| def ClearField( |
| self, |
| field_name: typing_extensions.Literal[ |
| "_with_single_partition", |
| b"_with_single_partition", |
| "order_spec", |
| b"order_spec", |
| "partition_spec", |
| b"partition_spec", |
| "with_single_partition", |
| b"with_single_partition", |
| ], |
| ) -> None: ... |
| def WhichOneof( |
| self, |
| oneof_group: typing_extensions.Literal[ |
| "_with_single_partition", b"_with_single_partition" |
| ], |
| ) -> typing_extensions.Literal["with_single_partition"] | None: ... |
| |
| PLAN_ID_FIELD_NUMBER: builtins.int |
| SUBQUERY_TYPE_FIELD_NUMBER: builtins.int |
| TABLE_ARG_OPTIONS_FIELD_NUMBER: builtins.int |
| IN_SUBQUERY_VALUES_FIELD_NUMBER: builtins.int |
| plan_id: builtins.int |
| """(Required) The ID of the corresponding connect plan.""" |
| subquery_type: global___SubqueryExpression.SubqueryType.ValueType |
| """(Required) The type of the subquery.""" |
| @property |
| def table_arg_options(self) -> global___SubqueryExpression.TableArgOptions: |
| """(Optional) Options specific to table arguments.""" |
| @property |
| def in_subquery_values( |
| self, |
| ) -> google.protobuf.internal.containers.RepeatedCompositeFieldContainer[global___Expression]: |
| """(Optional) IN subquery values.""" |
| def __init__( |
| self, |
| *, |
| plan_id: builtins.int = ..., |
| subquery_type: global___SubqueryExpression.SubqueryType.ValueType = ..., |
| table_arg_options: global___SubqueryExpression.TableArgOptions | None = ..., |
| in_subquery_values: collections.abc.Iterable[global___Expression] | None = ..., |
| ) -> None: ... |
| def HasField( |
| self, |
| field_name: typing_extensions.Literal[ |
| "_table_arg_options", b"_table_arg_options", "table_arg_options", b"table_arg_options" |
| ], |
| ) -> builtins.bool: ... |
| def ClearField( |
| self, |
| field_name: typing_extensions.Literal[ |
| "_table_arg_options", |
| b"_table_arg_options", |
| "in_subquery_values", |
| b"in_subquery_values", |
| "plan_id", |
| b"plan_id", |
| "subquery_type", |
| b"subquery_type", |
| "table_arg_options", |
| b"table_arg_options", |
| ], |
| ) -> None: ... |
| def WhichOneof( |
| self, oneof_group: typing_extensions.Literal["_table_arg_options", b"_table_arg_options"] |
| ) -> typing_extensions.Literal["table_arg_options"] | None: ... |
| |
| global___SubqueryExpression = SubqueryExpression |