SparkFilesystemDatasource
- SparkFilesystemDatasource is a subclass of SparkDatasource which connects to the filesystem. 
- Add a csv asset to the datasource. 
- Add a delta asset to the datasource. 
- Add a directory_csv asset to the datasource. 
- Add a directory_delta asset to the datasource. 
- Add a directory_json asset to the datasource. 
- Add a directory_orc asset to the datasource. 
- Add a directory_parquet asset to the datasource. 
- Add a directory_text asset to the datasource. 
- Add a json asset to the datasource. 
- Add an orc asset to the datasource. 
- Add a parquet asset to the datasource. 
- Add a text asset to the datasource. 
- Removes the DataAsset referred to by asset_name from internal list of available DataAsset objects. - Parameters- Name - Description - name - name of DataAsset to be deleted. 
- Returns the DataAsset referred to by asset_name - Parameters- Name - Description - name - name of DataAsset sought. - Returns- Type - Description - great_expectations.datasource.fluent.interfaces._DataAssetT - if named "DataAsset" object exists; otherwise, exception is raised. 
class great_expectations.datasource.fluent.SparkFilesystemDatasource(
 *,
 type: Literal['spark_filesystem'] = 'spark_filesystem',
 name: str,
 id: Optional[uuid.UUID] = None,
 assets: List[Union[great_expectations.datasource.fluent.data_asset.path.spark.csv_asset.CSVAsset,
 great_expectations.datasource.fluent.data_asset.path.spark.csv_asset.DirectoryCSVAsset,
 great_expectations.datasource.fluent.data_asset.path.spark.parquet_asset.ParquetAsset,
 great_expectations.datasource.fluent.data_asset.path.spark.parquet_asset.DirectoryParquetAsset,
 great_expectations.datasource.fluent.data_asset.path.spark.orc_asset.ORCAsset,
 great_expectations.datasource.fluent.data_asset.path.spark.orc_asset.DirectoryORCAsset,
 great_expectations.datasource.fluent.data_asset.path.spark.json_asset.JSONAsset,
 great_expectations.datasource.fluent.data_asset.path.spark.json_asset.DirectoryJSONAsset,
 great_expectations.datasource.fluent.data_asset.path.spark.text_asset.TextAsset,
 great_expectations.datasource.fluent.data_asset.path.spark.text_asset.DirectoryTextAsset,
 great_expectations.datasource.fluent.data_asset.path.spark.delta_asset.DeltaAsset,
 great_expectations.datasource.fluent.data_asset.path.spark.delta_asset.DirectoryDeltaAsset]] = [],
 spark_config: Optional[Dict[pydantic.v1.types.StrictStr,
 Union[pydantic.v1.types.StrictStr,
 pydantic.v1.types.StrictInt,
 pydantic.v1.types.StrictFloat,
 pydantic.v1.types.StrictBool]]] = None,
 force_reuse_spark_context: bool = True,
 persist: bool = True,
 base_directory: pathlib.Path,
 data_context_root_directory: Optional[pathlib.Path] = None
)
Methods
add_csv_asset
add_csv_asset(
 name: str,
 *,
 id: <pydantic.v1.fields.DeferredType object at 0x7f8b65b7c080> = None,
 order_by: <pydantic.v1.fields.DeferredType object at 0x7f8b65b7c140> = None,
 batch_metadata: <pydantic.v1.fields.DeferredType object at 0x7f8b65b7c290> = None,
 batch_definitions: <pydantic.v1.fields.DeferredType object at 0x7f8b65b7c440> = None,
 connect_options: <pydantic.v1.fields.DeferredType object at 0x7f8b65b7c500> = None,
 pathGlobFilter: typing.Optional[typing.Union[bool,
 str]] = None,
 recursiveFileLookup: typing.Optional[typing.Union[bool,
 str]] = None,
 modifiedBefore: typing.Optional[typing.Union[bool,
 str]] = None,
 modifiedAfter: typing.Optional[typing.Union[bool,
 str]] = None,
 schema: typing.Optional[typing.Union[great_expectations.datasource.fluent.serializable_types.pyspark.SerializableStructType,
 str]] = None,
 sep: typing.Optional[str] = None,
 encoding: typing.Optional[str] = None,
 quote: typing.Optional[str] = None,
 escape: typing.Optional[str] = None,
 comment: typing.Optional[str] = None,
 header: typing.Optional[typing.Union[bool,
 str]] = None,
 inferSchema: typing.Optional[typing.Union[bool,
 str]] = None,
 ignoreLeadingWhiteSpace: typing.Optional[typing.Union[bool,
 str]] = None,
 ignoreTrailingWhiteSpace: typing.Optional[typing.Union[bool,
 str]] = None,
 nullValue: typing.Optional[str] = None,
 nanValue: typing.Optional[str] = None,
 positiveInf: typing.Optional[str] = None,
 negativeInf: typing.Optional[str] = None,
 dateFormat: typing.Optional[str] = None,
 timestampFormat: typing.Optional[str] = None,
 maxColumns: typing.Optional[typing.Union[int,
 str]] = None,
 maxCharsPerColumn: typing.Optional[typing.Union[int,
 str]] = None,
 maxMalformedLogPerPartition: typing.Optional[typing.Union[int,
 str]] = None,
 mode: typing.Optional[typing.Literal['PERMISSIVE',
 'DROPMALFORMED',
 'FAILFAST']] = None,
 columnNameOfCorruptRecord: typing.Optional[str] = None,
 multiLine: typing.Optional[typing.Union[bool,
 str]] = None,
 charToEscapeQuoteEscaping: typing.Optional[str] = None,
 samplingRatio: typing.Optional[typing.Union[float,
 str]] = None,
 enforceSchema: typing.Optional[typing.Union[bool,
 str]] = None,
 emptyValue: typing.Optional[str] = None,
 locale: typing.Optional[str] = None,
 lineSep: typing.Optional[str] = None,
 unescapedQuoteHandling: typing.Optional[typing.Literal['STOP_AT_CLOSING_QUOTE',
 'BACK_TO_DELIMITER',
 'STOP_AT_DELIMITER',
 'SKIP_VALUE',
 'RAISE_ERROR']] = None
) → pydantic.BaseModel
add_delta_asset
add_delta_asset(
 name: str,
 *,
 id: <pydantic.v1.fields.DeferredType object at 0x7f8b65b7ffb0> = None,
 order_by: <pydantic.v1.fields.DeferredType object at 0x7f8b65b940b0> = None,
 batch_metadata: <pydantic.v1.fields.DeferredType object at 0x7f8b65b94200> = None,
 batch_definitions: <pydantic.v1.fields.DeferredType object at 0x7f8b65b943b0> = None,
 connect_options: <pydantic.v1.fields.DeferredType object at 0x7f8b65b94470> = None,
 timestampAsOf: typing.Optional[str] = None,
 versionAsOf: typing.Optional[str] = None
) → pydantic.BaseModel
add_directory_csv_asset
add_directory_csv_asset(
 name: str,
 *,
 id: <pydantic.v1.fields.DeferredType object at 0x7f8b65b7e7b0> = None,
 order_by: <pydantic.v1.fields.DeferredType object at 0x7f8b65b7e870> = None,
 batch_metadata: <pydantic.v1.fields.DeferredType object at 0x7f8b65b7e9c0> = None,
 batch_definitions: <pydantic.v1.fields.DeferredType object at 0x7f8b65b7eb70> = None,
 connect_options: <pydantic.v1.fields.DeferredType object at 0x7f8b65b7ec30> = None,
 pathGlobFilter: typing.Optional[typing.Union[bool,
 str]] = None,
 recursiveFileLookup: typing.Optional[typing.Union[bool,
 str]] = None,
 modifiedBefore: typing.Optional[typing.Union[bool,
 str]] = None,
 modifiedAfter: typing.Optional[typing.Union[bool,
 str]] = None,
 schema: typing.Optional[typing.Union[great_expectations.datasource.fluent.serializable_types.pyspark.SerializableStructType,
 str]] = None,
 sep: typing.Optional[str] = None,
 encoding: typing.Optional[str] = None,
 quote: typing.Optional[str] = None,
 escape: typing.Optional[str] = None,
 comment: typing.Optional[str] = None,
 header: typing.Optional[typing.Union[bool,
 str]] = None,
 inferSchema: typing.Optional[typing.Union[bool,
 str]] = None,
 ignoreLeadingWhiteSpace: typing.Optional[typing.Union[bool,
 str]] = None,
 ignoreTrailingWhiteSpace: typing.Optional[typing.Union[bool,
 str]] = None,
 nullValue: typing.Optional[str] = None,
 nanValue: typing.Optional[str] = None,
 positiveInf: typing.Optional[str] = None,
 negativeInf: typing.Optional[str] = None,
 dateFormat: typing.Optional[str] = None,
 timestampFormat: typing.Optional[str] = None,
 maxColumns: typing.Optional[typing.Union[int,
 str]] = None,
 maxCharsPerColumn: typing.Optional[typing.Union[int,
 str]] = None,
 maxMalformedLogPerPartition: typing.Optional[typing.Union[int,
 str]] = None,
 mode: typing.Optional[typing.Literal['PERMISSIVE',
 'DROPMALFORMED',
 'FAILFAST']] = None,
 columnNameOfCorruptRecord: typing.Optional[str] = None,
 multiLine: typing.Optional[typing.Union[bool,
 str]] = None,
 charToEscapeQuoteEscaping: typing.Optional[str] = None,
 samplingRatio: typing.Optional[typing.Union[float,
 str]] = None,
 enforceSchema: typing.Optional[typing.Union[bool,
 str]] = None,
 emptyValue: typing.Optional[str] = None,
 locale: typing.Optional[str] = None,
 lineSep: typing.Optional[str] = None,
 unescapedQuoteHandling: typing.Optional[typing.Literal['STOP_AT_CLOSING_QUOTE',
 'BACK_TO_DELIMITER',
 'STOP_AT_DELIMITER',
 'SKIP_VALUE',
 'RAISE_ERROR']] = None,
 data_directory: pathlib.Path
) → pydantic.BaseModel
add_directory_delta_asset
add_directory_delta_asset(
 name: str,
 *,
 id: <pydantic.v1.fields.DeferredType object at 0x7f8b665eb530> = None,
 order_by: <pydantic.v1.fields.DeferredType object at 0x7f8b665eb590> = None,
 batch_metadata: <pydantic.v1.fields.DeferredType object at 0x7f8b665eb620> = None,
 batch_definitions: <pydantic.v1.fields.DeferredType object at 0x7f8b665eb950> = None,
 connect_options: <pydantic.v1.fields.DeferredType object at 0x7f8b665ebb00> = None,
 timestampAsOf: typing.Optional[str] = None,
 versionAsOf: typing.Optional[str] = None,
 data_directory: pathlib.Path
) → pydantic.BaseModel
add_directory_json_asset
add_directory_json_asset(
 name: str,
 *,
 id: <pydantic.v1.fields.DeferredType object at 0x7f8b66085df0> = None,
 order_by: <pydantic.v1.fields.DeferredType object at 0x7f8b66086870> = None,
 batch_metadata: <pydantic.v1.fields.DeferredType object at 0x7f8b66087e90> = None,
 batch_definitions: <pydantic.v1.fields.DeferredType object at 0x7f8b660877a0> = None,
 connect_options: <pydantic.v1.fields.DeferredType object at 0x7f8b66087c50> = None,
 pathGlobFilter: typing.Optional[typing.Union[bool,
 str]] = None,
 recursiveFileLookup: typing.Optional[typing.Union[bool,
 str]] = None,
 modifiedBefore: typing.Optional[typing.Union[bool,
 str]] = None,
 modifiedAfter: typing.Optional[typing.Union[bool,
 str]] = None,
 schema: typing.Optional[typing.Union[great_expectations.datasource.fluent.serializable_types.pyspark.SerializableStructType,
 str]] = None,
 primitivesAsString: typing.Optional[typing.Union[bool,
 str]] = None,
 prefersDecimal: typing.Optional[typing.Union[bool,
 str]] = None,
 allowComments: typing.Optional[typing.Union[bool,
 str]] = None,
 allowUnquotedFieldNames: typing.Optional[typing.Union[bool,
 str]] = None,
 allowSingleQuotes: typing.Optional[typing.Union[bool,
 str]] = None,
 allowNumericLeadingZero: typing.Optional[typing.Union[bool,
 str]] = None,
 allowBackslashEscapingAnyCharacter: typing.Optional[typing.Union[bool,
 str]] = None,
 mode: typing.Optional[typing.Literal['PERMISSIVE',
 'DROPMALFORMED',
 'FAILFAST']] = None,
 columnNameOfCorruptRecord: typing.Optional[str] = None,
 dateFormat: typing.Optional[str] = None,
 timestampFormat: typing.Optional[str] = None,
 multiLine: typing.Optional[typing.Union[bool,
 str]] = None,
 allowUnquotedControlChars: typing.Optional[typing.Union[bool,
 str]] = None,
 lineSep: typing.Optional[str] = None,
 samplingRatio: typing.Optional[typing.Union[float,
 str]] = None,
 dropFieldIfAllNull: typing.Optional[typing.Union[bool,
 str]] = None,
 encoding: typing.Optional[str] = None,
 locale: typing.Optional[str] = None,
 allowNonNumericNumbers: typing.Optional[typing.Union[bool,
 str]] = None,
 data_directory: pathlib.Path
) → pydantic.BaseModel
add_directory_orc_asset
add_directory_orc_asset(
 name: str,
 *,
 id: <pydantic.v1.fields.DeferredType object at 0x7f8b65c930e0> = None,
 order_by: <pydantic.v1.fields.DeferredType object at 0x7f8b65c92ed0> = None,
 batch_metadata: <pydantic.v1.fields.DeferredType object at 0x7f8b65c933b0> = None,
 batch_definitions: <pydantic.v1.fields.DeferredType object at 0x7f8b65c93560> = None,
 connect_options: <pydantic.v1.fields.DeferredType object at 0x7f8b65c93590> = None,
 pathGlobFilter: typing.Optional[typing.Union[bool,
 str]] = None,
 recursiveFileLookup: typing.Optional[typing.Union[bool,
 str]] = None,
 modifiedBefore: typing.Optional[typing.Union[bool,
 str]] = None,
 modifiedAfter: typing.Optional[typing.Union[bool,
 str]] = None,
 mergeSchema: typing.Optional[typing.Union[bool,
 str]] = False,
 data_directory: pathlib.Path
) → pydantic.BaseModel
add_directory_parquet_asset
add_directory_parquet_asset(
 name: str,
 *,
 id: <pydantic.v1.fields.DeferredType object at 0x7f8b65b7ed20> = None,
 order_by: <pydantic.v1.fields.DeferredType object at 0x7f8b65b7ebd0> = None,
 batch_metadata: <pydantic.v1.fields.DeferredType object at 0x7f8b65b7f1d0> = None,
 batch_definitions: <pydantic.v1.fields.DeferredType object at 0x7f8b65b7f470> = None,
 connect_options: <pydantic.v1.fields.DeferredType object at 0x7f8b65b7f3e0> = None,
 pathGlobFilter: typing.Optional[typing.Union[bool,
 str]] = None,
 recursiveFileLookup: typing.Optional[typing.Union[bool,
 str]] = None,
 modifiedBefore: typing.Optional[typing.Union[bool,
 str]] = None,
 modifiedAfter: typing.Optional[typing.Union[bool,
 str]] = None,
 mergeSchema: typing.Optional[typing.Union[bool,
 str]] = None,
 datetimeRebaseMode: typing.Optional[typing.Literal['EXCEPTION',
 'CORRECTED',
 'LEGACY']] = None,
 int96RebaseMode: typing.Optional[typing.Literal['EXCEPTION',
 'CORRECTED',
 'LEGACY']] = None,
 data_directory: pathlib.Path
) → pydantic.BaseModel
add_directory_text_asset
add_directory_text_asset(
 name: str,
 *,
 id: <pydantic.v1.fields.DeferredType object at 0x7f8b65b96f30> = None,
 order_by: <pydantic.v1.fields.DeferredType object at 0x7f8b65b96ff0> = None,
 batch_metadata: <pydantic.v1.fields.DeferredType object at 0x7f8b65b97140> = None,
 batch_definitions: <pydantic.v1.fields.DeferredType object at 0x7f8b65b972f0> = None,
 connect_options: <pydantic.v1.fields.DeferredType object at 0x7f8b65b973b0> = None,
 pathGlobFilter: typing.Optional[typing.Union[bool,
 str]] = None,
 recursiveFileLookup: typing.Optional[typing.Union[bool,
 str]] = None,
 modifiedBefore: typing.Optional[typing.Union[bool,
 str]] = None,
 modifiedAfter: typing.Optional[typing.Union[bool,
 str]] = None,
 wholetext: bool = False,
 lineSep: typing.Optional[str] = None,
 data_directory: pathlib.Path
) → pydantic.BaseModel
add_json_asset
add_json_asset(
 name: str,
 *,
 id: <pydantic.v1.fields.DeferredType object at 0x7f8b66423bc0> = None,
 order_by: <pydantic.v1.fields.DeferredType object at 0x7f8b6605c2f0> = None,
 batch_metadata: <pydantic.v1.fields.DeferredType object at 0x7f8b6605d9a0> = None,
 batch_definitions: <pydantic.v1.fields.DeferredType object at 0x7f8b6605e3f0> = None,
 connect_options: <pydantic.v1.fields.DeferredType object at 0x7f8b6605e510> = None,
 pathGlobFilter: typing.Optional[typing.Union[bool,
 str]] = None,
 recursiveFileLookup: typing.Optional[typing.Union[bool,
 str]] = None,
 modifiedBefore: typing.Optional[typing.Union[bool,
 str]] = None,
 modifiedAfter: typing.Optional[typing.Union[bool,
 str]] = None,
 schema: typing.Optional[typing.Union[great_expectations.datasource.fluent.serializable_types.pyspark.SerializableStructType,
 str]] = None,
 primitivesAsString: typing.Optional[typing.Union[bool,
 str]] = None,
 prefersDecimal: typing.Optional[typing.Union[bool,
 str]] = None,
 allowComments: typing.Optional[typing.Union[bool,
 str]] = None,
 allowUnquotedFieldNames: typing.Optional[typing.Union[bool,
 str]] = None,
 allowSingleQuotes: typing.Optional[typing.Union[bool,
 str]] = None,
 allowNumericLeadingZero: typing.Optional[typing.Union[bool,
 str]] = None,
 allowBackslashEscapingAnyCharacter: typing.Optional[typing.Union[bool,
 str]] = None,
 mode: typing.Optional[typing.Literal['PERMISSIVE',
 'DROPMALFORMED',
 'FAILFAST']] = None,
 columnNameOfCorruptRecord: typing.Optional[str] = None,
 dateFormat: typing.Optional[str] = None,
 timestampFormat: typing.Optional[str] = None,
 multiLine: typing.Optional[typing.Union[bool,
 str]] = None,
 allowUnquotedControlChars: typing.Optional[typing.Union[bool,
 str]] = None,
 lineSep: typing.Optional[str] = None,
 samplingRatio: typing.Optional[typing.Union[float,
 str]] = None,
 dropFieldIfAllNull: typing.Optional[typing.Union[bool,
 str]] = None,
 encoding: typing.Optional[str] = None,
 locale: typing.Optional[str] = None,
 allowNonNumericNumbers: typing.Optional[typing.Union[bool,
 str]] = None
) → pydantic.BaseModel
add_orc_asset
add_orc_asset(
 name: str,
 *,
 id: <pydantic.v1.fields.DeferredType object at 0x7f8b65cadfa0> = None,
 order_by: <pydantic.v1.fields.DeferredType object at 0x7f8b65cae630> = None,
 batch_metadata: <pydantic.v1.fields.DeferredType object at 0x7f8b65caf0e0> = None,
 batch_definitions: <pydantic.v1.fields.DeferredType object at 0x7f8b65caf980> = None,
 connect_options: <pydantic.v1.fields.DeferredType object at 0x7f8b65c90ec0> = None,
 pathGlobFilter: typing.Optional[typing.Union[bool,
 str]] = None,
 recursiveFileLookup: typing.Optional[typing.Union[bool,
 str]] = None,
 modifiedBefore: typing.Optional[typing.Union[bool,
 str]] = None,
 modifiedAfter: typing.Optional[typing.Union[bool,
 str]] = None,
 mergeSchema: typing.Optional[typing.Union[bool,
 str]] = False
) → pydantic.BaseModel
add_parquet_asset
add_parquet_asset(
 name: str,
 *,
 id: <pydantic.v1.fields.DeferredType object at 0x7f8b65b7d2e0> = None,
 order_by: <pydantic.v1.fields.DeferredType object at 0x7f8b65b7d460> = None,
 batch_metadata: <pydantic.v1.fields.DeferredType object at 0x7f8b65b7d490> = None,
 batch_definitions: <pydantic.v1.fields.DeferredType object at 0x7f8b65b7d850> = None,
 connect_options: <pydantic.v1.fields.DeferredType object at 0x7f8b65b7d910> = None,
 pathGlobFilter: typing.Optional[typing.Union[bool,
 str]] = None,
 recursiveFileLookup: typing.Optional[typing.Union[bool,
 str]] = None,
 modifiedBefore: typing.Optional[typing.Union[bool,
 str]] = None,
 modifiedAfter: typing.Optional[typing.Union[bool,
 str]] = None,
 mergeSchema: typing.Optional[typing.Union[bool,
 str]] = None,
 datetimeRebaseMode: typing.Optional[typing.Literal['EXCEPTION',
 'CORRECTED',
 'LEGACY']] = None,
 int96RebaseMode: typing.Optional[typing.Literal['EXCEPTION',
 'CORRECTED',
 'LEGACY']] = None
) → pydantic.BaseModel
add_text_asset
add_text_asset(
 name: str,
 *,
 id: <pydantic.v1.fields.DeferredType object at 0x7f8b65b959d0> = None,
 order_by: <pydantic.v1.fields.DeferredType object at 0x7f8b65b95a90> = None,
 batch_metadata: <pydantic.v1.fields.DeferredType object at 0x7f8b65b95be0> = None,
 batch_definitions: <pydantic.v1.fields.DeferredType object at 0x7f8b65b95d90> = None,
 connect_options: <pydantic.v1.fields.DeferredType object at 0x7f8b65b95e50> = None,
 pathGlobFilter: typing.Optional[typing.Union[bool,
 str]] = None,
 recursiveFileLookup: typing.Optional[typing.Union[bool,
 str]] = None,
 modifiedBefore: typing.Optional[typing.Union[bool,
 str]] = None,
 modifiedAfter: typing.Optional[typing.Union[bool,
 str]] = None,
 wholetext: bool = False,
 lineSep: typing.Optional[str] = None
) → pydantic.BaseModel
delete_asset
delete_asset(
 name: str
) → None
get_asset
get_asset(
 name: str
) → great_expectations.datasource.fluent.interfaces._DataAssetT