trait SupportsReportPartitioning extends Scan
A mix in interface for Scan. Data sources can implement this interface to
report data partitioning and try to avoid shuffle at Spark side.
Note that, when a Scan implementation creates exactly one InputPartition,
Spark may avoid adding a shuffle even if the reader does not implement this interface.
- Since
3.0.0
- Alphabetic
- By Inheritance
- SupportsReportPartitioning
- Scan
- AnyRef
- Any
- Hide All
- Show All
- Public
- All
Abstract Value Members
-
abstract
def
outputPartitioning(): Partitioning
Returns the output data partitioning that this reader guarantees.
-
abstract
def
readSchema(): StructType
Returns the actual schema of this data source scan, which may be different from the physical schema of the underlying storage, as column pruning or other optimizations may happen.
Returns the actual schema of this data source scan, which may be different from the physical schema of the underlying storage, as column pruning or other optimizations may happen.
- Definition Classes
- Scan
Concrete Value Members
-
final
def
!=(arg0: Any): Boolean
- Definition Classes
- AnyRef → Any
-
final
def
##(): Int
- Definition Classes
- AnyRef → Any
-
final
def
==(arg0: Any): Boolean
- Definition Classes
- AnyRef → Any
-
final
def
asInstanceOf[T0]: T0
- Definition Classes
- Any
-
def
clone(): AnyRef
- Attributes
- protected[lang]
- Definition Classes
- AnyRef
- Annotations
- @throws( ... ) @native()
-
def
description(): String
A description string of this scan, which may includes information like: what filters are configured for this scan, what's the value of some important options like path, etc.
A description string of this scan, which may includes information like: what filters are configured for this scan, what's the value of some important options like path, etc. The description doesn't need to include
#readSchema(), as Spark already knows it.By default this returns the class name of the implementation. Please override it to provide a meaningful description.
- Definition Classes
- Scan
-
final
def
eq(arg0: AnyRef): Boolean
- Definition Classes
- AnyRef
-
def
equals(arg0: Any): Boolean
- Definition Classes
- AnyRef → Any
-
def
finalize(): Unit
- Attributes
- protected[lang]
- Definition Classes
- AnyRef
- Annotations
- @throws( classOf[java.lang.Throwable] )
-
final
def
getClass(): Class[_]
- Definition Classes
- AnyRef → Any
- Annotations
- @native()
-
def
hashCode(): Int
- Definition Classes
- AnyRef → Any
- Annotations
- @native()
-
final
def
isInstanceOf[T0]: Boolean
- Definition Classes
- Any
-
final
def
ne(arg0: AnyRef): Boolean
- Definition Classes
- AnyRef
-
final
def
notify(): Unit
- Definition Classes
- AnyRef
- Annotations
- @native()
-
final
def
notifyAll(): Unit
- Definition Classes
- AnyRef
- Annotations
- @native()
-
final
def
synchronized[T0](arg0: ⇒ T0): T0
- Definition Classes
- AnyRef
-
def
toBatch(): Batch
Returns the physical representation of this scan for batch query.
Returns the physical representation of this scan for batch query. By default this method throws exception, data sources must overwrite this method to provide an implementation, if the
Tablethat creates this scan returnsTableCapability#BATCH_READsupport in itsTable#capabilities().- Definition Classes
- Scan
- Exceptions thrown
-
def
toContinuousStream(checkpointLocation: String): ContinuousStream
Returns the physical representation of this scan for streaming query with continuous mode.
Returns the physical representation of this scan for streaming query with continuous mode. By default this method throws exception, data sources must overwrite this method to provide an implementation, if the
Tablethat creates this scan returnsTableCapability#CONTINUOUS_READsupport in itsTable#capabilities().- checkpointLocation
a path to Hadoop FS scratch space that can be used for failure recovery. Data streams for the same logical source in the same query will be given the same checkpointLocation.
- Definition Classes
- Scan
- Exceptions thrown
-
def
toMicroBatchStream(checkpointLocation: String): MicroBatchStream
Returns the physical representation of this scan for streaming query with micro-batch mode.
Returns the physical representation of this scan for streaming query with micro-batch mode. By default this method throws exception, data sources must overwrite this method to provide an implementation, if the
Tablethat creates this scan returnsTableCapability#MICRO_BATCH_READsupport in itsTable#capabilities().- checkpointLocation
a path to Hadoop FS scratch space that can be used for failure recovery. Data streams for the same logical source in the same query will be given the same checkpointLocation.
- Definition Classes
- Scan
- Exceptions thrown
-
def
toString(): String
- Definition Classes
- AnyRef → Any
-
final
def
wait(): Unit
- Definition Classes
- AnyRef
- Annotations
- @throws( ... )
-
final
def
wait(arg0: Long, arg1: Int): Unit
- Definition Classes
- AnyRef
- Annotations
- @throws( ... )
-
final
def
wait(arg0: Long): Unit
- Definition Classes
- AnyRef
- Annotations
- @throws( ... ) @native()