Skip to content
This repository has been archived by the owner on Nov 22, 2022. It is now read-only.

Backport DataFrameWriter argument changes #556

Merged
merged 1 commit into from
Dec 16, 2021
Merged
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
21 changes: 11 additions & 10 deletions third_party/3/pyspark/sql/readwriter.pyi
Original file line number Diff line number Diff line change
Expand Up @@ -2,13 +2,14 @@
#

from typing import overload
from typing import Any, Dict, List, Optional, Tuple, Union
from typing import Dict, List, Optional, Tuple, Union

from pyspark.sql._typing import OptionalPrimitiveType
from pyspark.sql.dataframe import DataFrame
from pyspark.rdd import RDD
from pyspark.sql.column import Column
from pyspark.sql.context import SQLContext
from pyspark.sql.types import *
from pyspark.sql.types import StructType

PathOrPaths = Union[str, List[str]]
TupleOrListOfString = Union[List[str], Tuple[str, ...]]
Expand All @@ -25,7 +26,7 @@ class DataFrameReader(OptionUtils):
self,
path: Optional[PathOrPaths] = ...,
format: Optional[str] = ...,
schema: Optional[StructType] = ...,
schema: Optional[Union[StructType, str]] = ...,
**options: OptionalPrimitiveType
) -> DataFrame: ...
def json(
Expand Down Expand Up @@ -109,8 +110,8 @@ class DataFrameReader(OptionUtils):
url: str,
table: str,
column: str,
lowerBound: int,
upperBound: int,
lowerBound: Union[int, str],
upperBound: Union[int, str],
numPartitions: int,
*,
properties: Optional[Dict[str, str]] = ...
Expand Down Expand Up @@ -150,7 +151,7 @@ class DataFrameWriter(OptionUtils):
path: Optional[str] = ...,
format: Optional[str] = ...,
mode: Optional[str] = ...,
partitionBy: Optional[List[str]] = ...,
partitionBy: Optional[Union[str, List[str]]] = ...,
**options: OptionalPrimitiveType
) -> None: ...
def insertInto(self, tableName: str, overwrite: Optional[bool] = ...) -> None: ...
Expand All @@ -159,7 +160,7 @@ class DataFrameWriter(OptionUtils):
name: str,
format: Optional[str] = ...,
mode: Optional[str] = ...,
partitionBy: Optional[List[str]] = ...,
partitionBy: Optional[Union[str, List[str]]] = ...,
**options: OptionalPrimitiveType
) -> None: ...
def json(
Expand All @@ -171,13 +172,13 @@ class DataFrameWriter(OptionUtils):
timestampFormat: Optional[str] = ...,
lineSep: Optional[str] = ...,
encoding: Optional[str] = ...,
ignoreNullFields: Optional[bool] = ...,
ignoreNullFields: Optional[Union[bool, str]] = ...,
) -> None: ...
def parquet(
self,
path: str,
mode: Optional[str] = ...,
partitionBy: Optional[List[str]] = ...,
partitionBy: Optional[Union[str, List[str]]] = ...,
compression: Optional[str] = ...,
) -> None: ...
def text(
Expand Down Expand Up @@ -208,7 +209,7 @@ class DataFrameWriter(OptionUtils):
self,
path: str,
mode: Optional[str] = ...,
partitionBy: Optional[List[str]] = ...,
partitionBy: Optional[Union[str, List[str]]] = ...,
compression: Optional[str] = ...,
) -> None: ...
def jdbc(
Expand Down