class DefaultSource extends TextBasedFileFormat with DataSourceRegister
Linear Supertypes
Ordering
- Alphabetic
- By Inheritance
Inherited
- DefaultSource
- DataSourceRegister
- TextBasedFileFormat
- FileFormat
- AnyRef
- Any
- Hide All
- Show All
Visibility
- Public
- All
Instance Constructors
- new DefaultSource()
Value Members
-
final
def
!=(arg0: Any): Boolean
- Definition Classes
- AnyRef → Any
-
final
def
##(): Int
- Definition Classes
- AnyRef → Any
-
final
def
==(arg0: Any): Boolean
- Definition Classes
- AnyRef → Any
-
final
def
asInstanceOf[T0]: T0
- Definition Classes
- Any
-
def
buildReader(sparkSession: SparkSession, dataSchema: StructType, partitionSchema: StructType, requiredSchema: StructType, filters: Seq[Filter], options: Map[String, String], hadoopConf: Configuration): (PartitionedFile) ⇒ Iterator[InternalRow]
code based on org.apache.spark:spark-core JsonFileFormat
code based on org.apache.spark:spark-core JsonFileFormat
- sparkSession
sparkSession
- dataSchema
dataSchema
- partitionSchema
partitionSchema
- requiredSchema
requiredSchema
- filters
filters
- options
options
- hadoopConf
hadoopConf
- returns
PartitionedFile => Iterator[InternalRow]
- Attributes
- protected
- Definition Classes
- DefaultSource → FileFormat
-
def
buildReaderWithPartitionValues(sparkSession: SparkSession, dataSchema: StructType, partitionSchema: StructType, requiredSchema: StructType, filters: Seq[Filter], options: Map[String, String], hadoopConf: Configuration): (PartitionedFile) ⇒ Iterator[InternalRow]
- Definition Classes
- FileFormat
-
def
clone(): AnyRef
- Attributes
- protected[lang]
- Definition Classes
- AnyRef
- Annotations
- @throws( ... ) @native()
-
final
def
eq(arg0: AnyRef): Boolean
- Definition Classes
- AnyRef
-
def
equals(arg0: Any): Boolean
- Definition Classes
- AnyRef → Any
-
def
finalize(): Unit
- Attributes
- protected[lang]
- Definition Classes
- AnyRef
- Annotations
- @throws( classOf[java.lang.Throwable] )
-
final
def
getClass(): Class[_]
- Definition Classes
- AnyRef → Any
- Annotations
- @native()
-
def
hashCode(): Int
- Definition Classes
- AnyRef → Any
- Annotations
- @native()
-
def
inferSchema(sparkSession: SparkSession, options: Map[String, String], files: Seq[FileStatus]): Option[StructType]
- Definition Classes
- DefaultSource → FileFormat
-
final
def
isInstanceOf[T0]: Boolean
- Definition Classes
- Any
-
def
isSplitable(sparkSession: SparkSession, options: Map[String, String], path: Path): Boolean
- Definition Classes
- TextBasedFileFormat → FileFormat
-
final
def
ne(arg0: AnyRef): Boolean
- Definition Classes
- AnyRef
-
final
def
notify(): Unit
- Definition Classes
- AnyRef
- Annotations
- @native()
-
final
def
notifyAll(): Unit
- Definition Classes
- AnyRef
- Annotations
- @native()
-
def
prepareWrite(sparkSession: SparkSession, job: Job, options: Map[String, String], dataSchema: StructType): OutputWriterFactory
- Definition Classes
- DefaultSource → FileFormat
-
val
shortName: String
- Definition Classes
- DefaultSource → DataSourceRegister
-
def
supportBatch(sparkSession: SparkSession, dataSchema: StructType): Boolean
- Definition Classes
- FileFormat
-
def
supportDataType(dataType: DataType): Boolean
- Definition Classes
- FileFormat
-
def
supportFieldName(name: String): Boolean
- Definition Classes
- FileFormat
-
final
def
synchronized[T0](arg0: ⇒ T0): T0
- Definition Classes
- AnyRef
-
def
toString(): String
- Definition Classes
- AnyRef → Any
-
def
vectorTypes(requiredSchema: StructType, partitionSchema: StructType, sqlConf: SQLConf): Option[Seq[String]]
- Definition Classes
- FileFormat
-
final
def
wait(): Unit
- Definition Classes
- AnyRef
- Annotations
- @throws( ... )
-
final
def
wait(arg0: Long, arg1: Int): Unit
- Definition Classes
- AnyRef
- Annotations
- @throws( ... )
-
final
def
wait(arg0: Long): Unit
- Definition Classes
- AnyRef
- Annotations
- @throws( ... ) @native()