Packages

t

org.apache.spark.connect.proto

WriteOperationV2OrBuilder

trait WriteOperationV2OrBuilder extends MessageOrBuilder

Linear Supertypes
MessageOrBuilder, MessageLiteOrBuilder, AnyRef, Any
Known Subclasses
Ordering
  1. Alphabetic
  2. By Inheritance
Inherited
  1. WriteOperationV2OrBuilder
  2. MessageOrBuilder
  3. MessageLiteOrBuilder
  4. AnyRef
  5. Any
  1. Hide All
  2. Show All
Visibility
  1. Public
  2. Protected

Abstract Value Members

  1. abstract def containsOptions(key: String): Boolean

    (Optional) A list of configuration options.
    

    (Optional) A list of configuration options.
    

    map<string, string> options = 5;

  2. abstract def containsTableProperties(key: String): Boolean

    (Optional) A list of table properties.
    

    (Optional) A list of table properties.
    

    map<string, string> table_properties = 6;

  3. abstract def findInitializationErrors(): List[String]
    Definition Classes
    MessageOrBuilder
  4. abstract def getAllFields(): Map[FieldDescriptor, AnyRef]
    Definition Classes
    MessageOrBuilder
  5. abstract def getClusteringColumns(index: Int): String

    (Optional) Columns used for clustering the table.
    

    (Optional) Columns used for clustering the table.
    

    repeated string clustering_columns = 9;

    index

    The index of the element to return.

    returns

    The clusteringColumns at the given index.

  6. abstract def getClusteringColumnsBytes(index: Int): ByteString

    (Optional) Columns used for clustering the table.
    

    (Optional) Columns used for clustering the table.
    

    repeated string clustering_columns = 9;

    index

    The index of the value to return.

    returns

    The bytes of the clusteringColumns at the given index.

  7. abstract def getClusteringColumnsCount(): Int

    (Optional) Columns used for clustering the table.
    

    (Optional) Columns used for clustering the table.
    

    repeated string clustering_columns = 9;

    returns

    The count of clusteringColumns.

  8. abstract def getClusteringColumnsList(): List[String]

    (Optional) Columns used for clustering the table.
    

    (Optional) Columns used for clustering the table.
    

    repeated string clustering_columns = 9;

    returns

    A list containing the clusteringColumns.

  9. abstract def getDefaultInstanceForType(): Message
    Definition Classes
    MessageOrBuilder → MessageLiteOrBuilder
  10. abstract def getDescriptorForType(): Descriptor
    Definition Classes
    MessageOrBuilder
  11. abstract def getField(field: FieldDescriptor): AnyRef
    Definition Classes
    MessageOrBuilder
  12. abstract def getInitializationErrorString(): String
    Definition Classes
    MessageOrBuilder
  13. abstract def getInput(): Relation

    (Required) The output of the `input` relation will be persisted according to the options.
    

    (Required) The output of the `input` relation will be persisted according to the options.
    

    .spark.connect.Relation input = 1;

    returns

    The input.

  14. abstract def getInputOrBuilder(): RelationOrBuilder

    (Required) The output of the `input` relation will be persisted according to the options.
    

    (Required) The output of the `input` relation will be persisted according to the options.
    

    .spark.connect.Relation input = 1;

  15. abstract def getMode(): Mode

    (Required) Write mode.
    

    (Required) Write mode.
    

    .spark.connect.WriteOperationV2.Mode mode = 7;

    returns

    The mode.

  16. abstract def getModeValue(): Int

    (Required) Write mode.
    

    (Required) Write mode.
    

    .spark.connect.WriteOperationV2.Mode mode = 7;

    returns

    The enum numeric value on the wire for mode.

  17. abstract def getOneofFieldDescriptor(oneof: OneofDescriptor): FieldDescriptor
    Definition Classes
    MessageOrBuilder
  18. abstract def getOptionsCount(): Int

    (Optional) A list of configuration options.
    

    (Optional) A list of configuration options.
    

    map<string, string> options = 5;

  19. abstract def getOptionsMap(): Map[String, String]

    (Optional) A list of configuration options.
    

    (Optional) A list of configuration options.
    

    map<string, string> options = 5;

  20. abstract def getOptionsOrDefault(key: String, defaultValue: String): String

    (Optional) A list of configuration options.
    

    (Optional) A list of configuration options.
    

    map<string, string> options = 5;

  21. abstract def getOptionsOrThrow(key: String): String

    (Optional) A list of configuration options.
    

    (Optional) A list of configuration options.
    

    map<string, string> options = 5;

  22. abstract def getOverwriteCondition(): Expression

    (Optional) A condition for overwrite saving mode
    

    (Optional) A condition for overwrite saving mode
    

    .spark.connect.Expression overwrite_condition = 8;

    returns

    The overwriteCondition.

  23. abstract def getOverwriteConditionOrBuilder(): ExpressionOrBuilder

    (Optional) A condition for overwrite saving mode
    

    (Optional) A condition for overwrite saving mode
    

    .spark.connect.Expression overwrite_condition = 8;

  24. abstract def getPartitioningColumns(index: Int): Expression

    (Optional) List of columns for partitioning for output table created by `create`,
    `createOrReplace`, or `replace`
    

    (Optional) List of columns for partitioning for output table created by `create`,
    `createOrReplace`, or `replace`
    

    repeated .spark.connect.Expression partitioning_columns = 4;

  25. abstract def getPartitioningColumnsCount(): Int

    (Optional) List of columns for partitioning for output table created by `create`,
    `createOrReplace`, or `replace`
    

    (Optional) List of columns for partitioning for output table created by `create`,
    `createOrReplace`, or `replace`
    

    repeated .spark.connect.Expression partitioning_columns = 4;

  26. abstract def getPartitioningColumnsList(): List[Expression]

    (Optional) List of columns for partitioning for output table created by `create`,
    `createOrReplace`, or `replace`
    

    (Optional) List of columns for partitioning for output table created by `create`,
    `createOrReplace`, or `replace`
    

    repeated .spark.connect.Expression partitioning_columns = 4;

  27. abstract def getPartitioningColumnsOrBuilder(index: Int): ExpressionOrBuilder

    (Optional) List of columns for partitioning for output table created by `create`,
    `createOrReplace`, or `replace`
    

    (Optional) List of columns for partitioning for output table created by `create`,
    `createOrReplace`, or `replace`
    

    repeated .spark.connect.Expression partitioning_columns = 4;

  28. abstract def getPartitioningColumnsOrBuilderList(): List[_ <: ExpressionOrBuilder]

    (Optional) List of columns for partitioning for output table created by `create`,
    `createOrReplace`, or `replace`
    

    (Optional) List of columns for partitioning for output table created by `create`,
    `createOrReplace`, or `replace`
    

    repeated .spark.connect.Expression partitioning_columns = 4;

  29. abstract def getProvider(): String

    (Optional) A provider for the underlying output data source. Spark's default catalog supports
    "parquet", "json", etc.
    

    (Optional) A provider for the underlying output data source. Spark's default catalog supports
    "parquet", "json", etc.
    

    optional string provider = 3;

    returns

    The provider.

  30. abstract def getProviderBytes(): ByteString

    (Optional) A provider for the underlying output data source. Spark's default catalog supports
    "parquet", "json", etc.
    

    (Optional) A provider for the underlying output data source. Spark's default catalog supports
    "parquet", "json", etc.
    

    optional string provider = 3;

    returns

    The bytes for provider.

  31. abstract def getRepeatedField(field: FieldDescriptor, index: Int): AnyRef
    Definition Classes
    MessageOrBuilder
  32. abstract def getRepeatedFieldCount(field: FieldDescriptor): Int
    Definition Classes
    MessageOrBuilder
  33. abstract def getTableName(): String

    (Required) The destination of the write operation must be either a path or a table.
    

    (Required) The destination of the write operation must be either a path or a table.
    

    string table_name = 2;

    returns

    The tableName.

  34. abstract def getTableNameBytes(): ByteString

    (Required) The destination of the write operation must be either a path or a table.
    

    (Required) The destination of the write operation must be either a path or a table.
    

    string table_name = 2;

    returns

    The bytes for tableName.

  35. abstract def getTablePropertiesCount(): Int

    (Optional) A list of table properties.
    

    (Optional) A list of table properties.
    

    map<string, string> table_properties = 6;

  36. abstract def getTablePropertiesMap(): Map[String, String]

    (Optional) A list of table properties.
    

    (Optional) A list of table properties.
    

    map<string, string> table_properties = 6;

  37. abstract def getTablePropertiesOrDefault(key: String, defaultValue: String): String

    (Optional) A list of table properties.
    

    (Optional) A list of table properties.
    

    map<string, string> table_properties = 6;

  38. abstract def getTablePropertiesOrThrow(key: String): String

    (Optional) A list of table properties.
    

    (Optional) A list of table properties.
    

    map<string, string> table_properties = 6;

  39. abstract def getUnknownFields(): UnknownFieldSet
    Definition Classes
    MessageOrBuilder
  40. abstract def hasField(field: FieldDescriptor): Boolean
    Definition Classes
    MessageOrBuilder
  41. abstract def hasInput(): Boolean

    (Required) The output of the `input` relation will be persisted according to the options.
    

    (Required) The output of the `input` relation will be persisted according to the options.
    

    .spark.connect.Relation input = 1;

    returns

    Whether the input field is set.

  42. abstract def hasOneof(oneof: OneofDescriptor): Boolean
    Definition Classes
    MessageOrBuilder
  43. abstract def hasOverwriteCondition(): Boolean

    (Optional) A condition for overwrite saving mode
    

    (Optional) A condition for overwrite saving mode
    

    .spark.connect.Expression overwrite_condition = 8;

    returns

    Whether the overwriteCondition field is set.

  44. abstract def hasProvider(): Boolean

    (Optional) A provider for the underlying output data source. Spark's default catalog supports
    "parquet", "json", etc.
    

    (Optional) A provider for the underlying output data source. Spark's default catalog supports
    "parquet", "json", etc.
    

    optional string provider = 3;

    returns

    Whether the provider field is set.

  45. abstract def isInitialized(): Boolean
    Definition Classes
    MessageLiteOrBuilder
  46. abstract def getOptions(): Map[String, String]

    Use #getOptionsMap() instead.

    Use #getOptionsMap() instead.

    Annotations
    @Deprecated
    Deprecated
  47. abstract def getTableProperties(): Map[String, String]

    Use #getTablePropertiesMap() instead.

    Use #getTablePropertiesMap() instead.

    Annotations
    @Deprecated
    Deprecated

Concrete Value Members

  1. final def !=(arg0: Any): Boolean
    Definition Classes
    AnyRef → Any
  2. final def ##: Int
    Definition Classes
    AnyRef → Any
  3. final def ==(arg0: Any): Boolean
    Definition Classes
    AnyRef → Any
  4. final def asInstanceOf[T0]: T0
    Definition Classes
    Any
  5. def clone(): AnyRef
    Attributes
    protected[lang]
    Definition Classes
    AnyRef
    Annotations
    @throws(classOf[java.lang.CloneNotSupportedException]) @IntrinsicCandidate() @native()
  6. final def eq(arg0: AnyRef): Boolean
    Definition Classes
    AnyRef
  7. def equals(arg0: AnyRef): Boolean
    Definition Classes
    AnyRef → Any
  8. final def getClass(): Class[_ <: AnyRef]
    Definition Classes
    AnyRef → Any
    Annotations
    @IntrinsicCandidate() @native()
  9. def hashCode(): Int
    Definition Classes
    AnyRef → Any
    Annotations
    @IntrinsicCandidate() @native()
  10. final def isInstanceOf[T0]: Boolean
    Definition Classes
    Any
  11. final def ne(arg0: AnyRef): Boolean
    Definition Classes
    AnyRef
  12. final def notify(): Unit
    Definition Classes
    AnyRef
    Annotations
    @IntrinsicCandidate() @native()
  13. final def notifyAll(): Unit
    Definition Classes
    AnyRef
    Annotations
    @IntrinsicCandidate() @native()
  14. final def synchronized[T0](arg0: => T0): T0
    Definition Classes
    AnyRef
  15. def toString(): String
    Definition Classes
    AnyRef → Any
  16. final def wait(arg0: Long, arg1: Int): Unit
    Definition Classes
    AnyRef
    Annotations
    @throws(classOf[java.lang.InterruptedException])
  17. final def wait(arg0: Long): Unit
    Definition Classes
    AnyRef
    Annotations
    @throws(classOf[java.lang.InterruptedException]) @native()
  18. final def wait(): Unit
    Definition Classes
    AnyRef
    Annotations
    @throws(classOf[java.lang.InterruptedException])

Deprecated Value Members

  1. def finalize(): Unit
    Attributes
    protected[lang]
    Definition Classes
    AnyRef
    Annotations
    @throws(classOf[java.lang.Throwable]) @Deprecated
    Deprecated

    (Since version 9)

Inherited from MessageOrBuilder

Inherited from MessageLiteOrBuilder

Inherited from AnyRef

Inherited from Any

Ungrouped