class FileStreamSource extends SupportsAdmissionControl with SupportsTriggerAvailableNow with Source with Logging
Instance Constructors
- new FileStreamSource(sparkSession: SparkSession, path: String, fileFormatClassName: String, schema: StructType, partitionColumns: Seq[String], metadataPath: String, options: Map[String, String])
Value Members
- final def !=(arg0: Any): Boolean
- final def ##: Int
- final def ==(arg0: Any): Boolean
- final def asInstanceOf[T0]: T0
- def clone(): AnyRef
- def commit(end: Offset): Unit
- def commit(end: connector.read.streaming.Offset): Unit
- def currentLogOffset: Long
- def deserializeOffset(json: String): connector.read.streaming.Offset
- final def eq(arg0: AnyRef): Boolean
- def equals(arg0: AnyRef): Boolean
- def finalize(): Unit
- def getBatch(start: Option[Offset], end: Offset): DataFrame
- final def getClass(): Class[_ <: AnyRef]
- def getDefaultReadLimit(): ReadLimit
- def getOffset: Option[Offset]
- def hashCode(): Int
- def initialOffset(): connector.read.streaming.Offset
- def initializeLogIfNecessary(isInterpreter: Boolean, silent: Boolean): Boolean
- def initializeLogIfNecessary(isInterpreter: Boolean): Unit
- final def isInstanceOf[T0]: Boolean
- def isTraceEnabled(): Boolean
- def latestOffset(startOffset: connector.read.streaming.Offset, limit: ReadLimit): connector.read.streaming.Offset
- def log: Logger
- def logDebug(msg: => String, throwable: Throwable): Unit
- def logDebug(msg: => String): Unit
- def logError(msg: => String, throwable: Throwable): Unit
- def logError(msg: => String): Unit
- def logInfo(msg: => String, throwable: Throwable): Unit
- def logInfo(msg: => String): Unit
- def logName: String
- def logTrace(msg: => String, throwable: Throwable): Unit
- def logTrace(msg: => String): Unit
- def logWarning(msg: => String, throwable: Throwable): Unit
- def logWarning(msg: => String): Unit
- final def ne(arg0: AnyRef): Boolean
- final def notify(): Unit
- final def notifyAll(): Unit
- def prepareForTriggerAvailableNow(): Unit
- def reportLatestOffset(): connector.read.streaming.Offset
- val schema: StructType
- val seenFiles: SeenFilesMap
- def stop(): Unit
- final def synchronized[T0](arg0: => T0): T0
- def toString(): String
- final def wait(): Unit
- final def wait(arg0: Long, arg1: Int): Unit
- final def wait(arg0: Long): Unit
- def withBatchingLocked[T](func: => T): T
Inherited from Logging
Value Members
- def initializeLogIfNecessary(isInterpreter: Boolean, silent: Boolean): Boolean
- def initializeLogIfNecessary(isInterpreter: Boolean): Unit
- def isTraceEnabled(): Boolean
- def log: Logger
- def logDebug(msg: => String, throwable: Throwable): Unit
- def logDebug(msg: => String): Unit
- def logError(msg: => String, throwable: Throwable): Unit
- def logError(msg: => String): Unit
- def logInfo(msg: => String, throwable: Throwable): Unit
- def logInfo(msg: => String): Unit
- def logName: String
- def logTrace(msg: => String, throwable: Throwable): Unit
- def logTrace(msg: => String): Unit
- def logWarning(msg: => String, throwable: Throwable): Unit
- def logWarning(msg: => String): Unit
Inherited from Source
Value Members
- def commit(end: connector.read.streaming.Offset): Unit
- def deserializeOffset(json: String): connector.read.streaming.Offset
- def initialOffset(): connector.read.streaming.Offset
Inherited from SupportsAdmissionControl
Value Members
- def reportLatestOffset(): connector.read.streaming.Offset
Inherited from AnyRef
Value Members
- final def !=(arg0: Any): Boolean
- final def ##: Int
- final def ==(arg0: Any): Boolean
- def clone(): AnyRef
- final def eq(arg0: AnyRef): Boolean
- def equals(arg0: AnyRef): Boolean
- def finalize(): Unit
- final def getClass(): Class[_ <: AnyRef]
- def hashCode(): Int
- final def ne(arg0: AnyRef): Boolean
- final def notify(): Unit
- final def notifyAll(): Unit
- final def synchronized[T0](arg0: => T0): T0
- final def wait(): Unit
- final def wait(arg0: Long, arg1: Int): Unit
- final def wait(arg0: Long): Unit
Inherited from Any
Value Members
- final def asInstanceOf[T0]: T0
- final def isInstanceOf[T0]: Boolean
Ungrouped
- final def !=(arg0: Any): Boolean
- final def ##: Int
- final def ==(arg0: Any): Boolean
- final def asInstanceOf[T0]: T0
- def clone(): AnyRef
- def commit(end: Offset): Unit
- def commit(end: connector.read.streaming.Offset): Unit
- def currentLogOffset: Long
- def deserializeOffset(json: String): connector.read.streaming.Offset
- final def eq(arg0: AnyRef): Boolean
- def equals(arg0: AnyRef): Boolean
- def finalize(): Unit
- def getBatch(start: Option[Offset], end: Offset): DataFrame
- final def getClass(): Class[_ <: AnyRef]
- def getDefaultReadLimit(): ReadLimit
- def getOffset: Option[Offset]
- def hashCode(): Int
- def initialOffset(): connector.read.streaming.Offset
- def initializeLogIfNecessary(isInterpreter: Boolean, silent: Boolean): Boolean
- def initializeLogIfNecessary(isInterpreter: Boolean): Unit
- final def isInstanceOf[T0]: Boolean
- def isTraceEnabled(): Boolean
- def latestOffset(startOffset: connector.read.streaming.Offset, limit: ReadLimit): connector.read.streaming.Offset
- def log: Logger
- def logDebug(msg: => String, throwable: Throwable): Unit
- def logDebug(msg: => String): Unit
- def logError(msg: => String, throwable: Throwable): Unit
- def logError(msg: => String): Unit
- def logInfo(msg: => String, throwable: Throwable): Unit
- def logInfo(msg: => String): Unit
- def logName: String
- def logTrace(msg: => String, throwable: Throwable): Unit
- def logTrace(msg: => String): Unit
- def logWarning(msg: => String, throwable: Throwable): Unit
- def logWarning(msg: => String): Unit
- final def ne(arg0: AnyRef): Boolean
- final def notify(): Unit
- final def notifyAll(): Unit
- def prepareForTriggerAvailableNow(): Unit
- def reportLatestOffset(): connector.read.streaming.Offset
- val schema: StructType
- val seenFiles: SeenFilesMap
- def stop(): Unit
- final def synchronized[T0](arg0: => T0): T0
- def toString(): String
- final def wait(): Unit
- final def wait(arg0: Long, arg1: Int): Unit
- final def wait(arg0: Long): Unit
- def withBatchingLocked[T](func: => T): T