com.krux.hyperion.aws

AdpHiveCopyActivity

case class AdpHiveCopyActivity(id: String, name: Option[String], filterSql: Option[String], generatedScriptsPath: Option[String], input: Option[AdpRef[AdpDataNode]], output: Option[AdpRef[AdpDataNode]], hadoopQueue: Option[String], preActivityTaskConfig: Option[AdpRef[AdpShellScriptConfig]], postActivityTaskConfig: Option[AdpRef[AdpShellScriptConfig]], workerGroup: Option[String], runsOn: Option[AdpRef[AdpEmrCluster]], dependsOn: Option[Seq[AdpRef[AdpActivity]]], precondition: Option[Seq[AdpRef[AdpPrecondition]]], onFail: Option[Seq[AdpRef[AdpSnsAlarm]]], onSuccess: Option[Seq[AdpRef[AdpSnsAlarm]]], onLateAction: Option[Seq[AdpRef[AdpSnsAlarm]]], attemptTimeout: Option[String], lateAfterTimeout: Option[String], maximumRetries: Option[String], retryDelay: Option[String], failureAndRerunMode: Option[String]) extends AdpDataPipelineAbstractObject with AdpActivity with Product with Serializable

ref: http://docs.aws.amazon.com/datapipeline/latest/DeveloperGuide/dp-object-hivecopyactivity.html

filterSql

A Hive SQL statement fragment that filters a subset of DynamoDB or Amazon S3 data to copy. The filter should only contain predicates and not begin with a WHERE clause, because AWS Data Pipeline adds it automatically.

generatedScriptsPath

An Amazon S3 path capturing the Hive script that ran after all the expressions in it were evaluated, including staging information. This script is stored for troubleshooting purposes.

input

The input data node. This must be S3DataNode or DynamoDBDataNode. If you use DynamoDBDataNode, specify a DynamoDBExportDataFormat.

output

The output data node. If input is S3DataNode, this must be DynamoDBDataNode. Otherwise, this can be S3DataNode or DynamoDBDataNode. If you use DynamoDBDataNode, specify a DynamoDBExportDataFormat.

Source
AdpActivities.scala
Linear Supertypes
Serializable, Serializable, Product, Equals, AdpActivity, AdpDataPipelineObject, AdpDataPipelineAbstractObject, AdpObject, AnyRef, Any
Ordering
  1. Alphabetic
  2. By inheritance
Inherited
  1. AdpHiveCopyActivity
  2. Serializable
  3. Serializable
  4. Product
  5. Equals
  6. AdpActivity
  7. AdpDataPipelineObject
  8. AdpDataPipelineAbstractObject
  9. AdpObject
  10. AnyRef
  11. Any
  1. Hide All
  2. Show all
Learn more about member selection
Visibility
  1. Public
  2. All

Instance Constructors

  1. new AdpHiveCopyActivity(id: String, name: Option[String], filterSql: Option[String], generatedScriptsPath: Option[String], input: Option[AdpRef[AdpDataNode]], output: Option[AdpRef[AdpDataNode]], hadoopQueue: Option[String], preActivityTaskConfig: Option[AdpRef[AdpShellScriptConfig]], postActivityTaskConfig: Option[AdpRef[AdpShellScriptConfig]], workerGroup: Option[String], runsOn: Option[AdpRef[AdpEmrCluster]], dependsOn: Option[Seq[AdpRef[AdpActivity]]], precondition: Option[Seq[AdpRef[AdpPrecondition]]], onFail: Option[Seq[AdpRef[AdpSnsAlarm]]], onSuccess: Option[Seq[AdpRef[AdpSnsAlarm]]], onLateAction: Option[Seq[AdpRef[AdpSnsAlarm]]], attemptTimeout: Option[String], lateAfterTimeout: Option[String], maximumRetries: Option[String], retryDelay: Option[String], failureAndRerunMode: Option[String])

    filterSql

    A Hive SQL statement fragment that filters a subset of DynamoDB or Amazon S3 data to copy. The filter should only contain predicates and not begin with a WHERE clause, because AWS Data Pipeline adds it automatically.

    generatedScriptsPath

    An Amazon S3 path capturing the Hive script that ran after all the expressions in it were evaluated, including staging information. This script is stored for troubleshooting purposes.

    input

    The input data node. This must be S3DataNode or DynamoDBDataNode. If you use DynamoDBDataNode, specify a DynamoDBExportDataFormat.

    output

    The output data node. If input is S3DataNode, this must be DynamoDBDataNode. Otherwise, this can be S3DataNode or DynamoDBDataNode. If you use DynamoDBDataNode, specify a DynamoDBExportDataFormat.

Value Members

  1. final def !=(arg0: AnyRef): Boolean

    Definition Classes
    AnyRef
  2. final def !=(arg0: Any): Boolean

    Definition Classes
    Any
  3. final def ##(): Int

    Definition Classes
    AnyRef → Any
  4. final def ==(arg0: AnyRef): Boolean

    Definition Classes
    AnyRef
  5. final def ==(arg0: Any): Boolean

    Definition Classes
    Any
  6. final def asInstanceOf[T0]: T0

    Definition Classes
    Any
  7. val attemptTimeout: Option[String]

    The timeout time interval for an object attempt.

    The timeout time interval for an object attempt. If an attempt does not complete within the start time plus this time interval, AWS Data Pipeline marks the attempt as failed and your retry settings determine the next steps taken.

    Definition Classes
    AdpHiveCopyActivityAdpActivity
  8. def clone(): AnyRef

    Attributes
    protected[java.lang]
    Definition Classes
    AnyRef
    Annotations
    @throws( ... )
  9. val dependsOn: Option[Seq[AdpRef[AdpActivity]]]

    One or more references to other Activities that must reach the FINISHED state before this activity will start.

    One or more references to other Activities that must reach the FINISHED state before this activity will start.

    Definition Classes
    AdpHiveCopyActivityAdpActivity
  10. final def eq(arg0: AnyRef): Boolean

    Definition Classes
    AnyRef
  11. val failureAndRerunMode: Option[String]

    Determines whether pipeline object failures and rerun commands cascade through pipeline object dependencies

    Determines whether pipeline object failures and rerun commands cascade through pipeline object dependencies

    Possible values include cascade and none.

    Definition Classes
    AdpHiveCopyActivityAdpActivity
  12. val filterSql: Option[String]

    A Hive SQL statement fragment that filters a subset of DynamoDB or Amazon S3 data to copy.

    A Hive SQL statement fragment that filters a subset of DynamoDB or Amazon S3 data to copy. The filter should only contain predicates and not begin with a WHERE clause, because AWS Data Pipeline adds it automatically.

  13. def finalize(): Unit

    Attributes
    protected[java.lang]
    Definition Classes
    AnyRef
    Annotations
    @throws( classOf[java.lang.Throwable] )
  14. val generatedScriptsPath: Option[String]

    An Amazon S3 path capturing the Hive script that ran after all the expressions in it were evaluated, including staging information.

    An Amazon S3 path capturing the Hive script that ran after all the expressions in it were evaluated, including staging information. This script is stored for troubleshooting purposes.

  15. final def getClass(): Class[_]

    Definition Classes
    AnyRef → Any
  16. val hadoopQueue: Option[String]

  17. val id: String

    The ID of the object, IDs must be unique within a pipeline definition

    The ID of the object, IDs must be unique within a pipeline definition

    Definition Classes
    AdpHiveCopyActivityAdpDataPipelineObjectAdpObject
  18. val input: Option[AdpRef[AdpDataNode]]

    The input data node.

    The input data node. This must be S3DataNode or DynamoDBDataNode. If you use DynamoDBDataNode, specify a DynamoDBExportDataFormat.

  19. final def isInstanceOf[T0]: Boolean

    Definition Classes
    Any
  20. val lateAfterTimeout: Option[String]

    The time period in which the object run must start.

    The time period in which the object run must start. If the object does not start within the scheduled start time plus this time interval, it is considered late

    Definition Classes
    AdpHiveCopyActivityAdpActivity
  21. val maximumRetries: Option[String]

    The maximum number of times to retry the action.

    The maximum number of times to retry the action. The default value is 2, which results in 3 tries total (1 original attempt plus 2 retries). The maximum value is 5 (6 total attempts).

    Definition Classes
    AdpHiveCopyActivityAdpActivity
  22. val name: Option[String]

    The optional, user-defined label of the object.

    The optional, user-defined label of the object. If you do not provide a name for an object in a pipeline definition, AWS Data Pipeline automatically duplicates the value of id.

    Definition Classes
    AdpHiveCopyActivityAdpDataPipelineObjectAdpDataPipelineAbstractObject
  23. final def ne(arg0: AnyRef): Boolean

    Definition Classes
    AnyRef
  24. final def notify(): Unit

    Definition Classes
    AnyRef
  25. final def notifyAll(): Unit

    Definition Classes
    AnyRef
  26. val onFail: Option[Seq[AdpRef[AdpSnsAlarm]]]

    The SNS alarm to raise when the activity fails.

    The SNS alarm to raise when the activity fails.

    Definition Classes
    AdpHiveCopyActivityAdpActivity
  27. val onLateAction: Option[Seq[AdpRef[AdpSnsAlarm]]]

    The SNS alarm to raise when the activity fails to start on time.

    The SNS alarm to raise when the activity fails to start on time.

    Definition Classes
    AdpHiveCopyActivityAdpActivity
  28. val onSuccess: Option[Seq[AdpRef[AdpSnsAlarm]]]

    The SNS alarm to raise when the activity succeeds.

    The SNS alarm to raise when the activity succeeds.

    Definition Classes
    AdpHiveCopyActivityAdpActivity
  29. val output: Option[AdpRef[AdpDataNode]]

    The output data node.

    The output data node. If input is S3DataNode, this must be DynamoDBDataNode. Otherwise, this can be S3DataNode or DynamoDBDataNode. If you use DynamoDBDataNode, specify a DynamoDBExportDataFormat.

  30. val postActivityTaskConfig: Option[AdpRef[AdpShellScriptConfig]]

  31. val preActivityTaskConfig: Option[AdpRef[AdpShellScriptConfig]]

  32. val precondition: Option[Seq[AdpRef[AdpPrecondition]]]

    A condition that must be met before the object can run.

    A condition that must be met before the object can run. To specify multiple conditions, add multiple precondition fields. The activity cannot run until all its conditions are met.

    Definition Classes
    AdpHiveCopyActivityAdpActivity
  33. val retryDelay: Option[String]

    The timeout duration between two retry attempts.

    The timeout duration between two retry attempts. The default is 10 minutes.

    Definition Classes
    AdpHiveCopyActivityAdpActivity
  34. val runsOn: Option[AdpRef[AdpEmrCluster]]

  35. final def synchronized[T0](arg0: ⇒ T0): T0

    Definition Classes
    AnyRef
  36. val type: String

    The type of object.

    The type of object. Use one of the predefined AWS Data Pipeline object types.

    Definition Classes
    AdpHiveCopyActivityAdpDataPipelineObject
  37. final def wait(): Unit

    Definition Classes
    AnyRef
    Annotations
    @throws( ... )
  38. final def wait(arg0: Long, arg1: Int): Unit

    Definition Classes
    AnyRef
    Annotations
    @throws( ... )
  39. final def wait(arg0: Long): Unit

    Definition Classes
    AnyRef
    Annotations
    @throws( ... )
  40. val workerGroup: Option[String]

    The worker group.

    The worker group. This is used for routing tasks. If you provide a runsOn value and workerGroup exists, workerGroup is ignored.

    Definition Classes
    AdpHiveCopyActivityAdpActivity

Inherited from Serializable

Inherited from Serializable

Inherited from Product

Inherited from Equals

Inherited from AdpActivity

Inherited from AdpDataPipelineObject

Inherited from AdpObject

Inherited from AnyRef

Inherited from Any

Ungrouped