Class/Object

org.bdgenomics.adam.api.java

JavaADAMContext

Related Docs: object JavaADAMContext | package java

Permalink

class JavaADAMContext extends Serializable

The JavaADAMContext provides java-friendly functions on top of ADAMContext.

Linear Supertypes
Serializable, Serializable, AnyRef, Any
Ordering
  1. Alphabetic
  2. By Inheritance
Inherited
  1. JavaADAMContext
  2. Serializable
  3. Serializable
  4. AnyRef
  5. Any
  1. Hide All
  2. Show All
Visibility
  1. Public
  2. All

Instance Constructors

  1. new JavaADAMContext(ac: ADAMContext)

    Permalink

    ac

    The ADAMContext to wrap.

Value Members

  1. final def !=(arg0: Any): Boolean

    Permalink
    Definition Classes
    AnyRef → Any
  2. final def ##(): Int

    Permalink
    Definition Classes
    AnyRef → Any
  3. final def ==(arg0: Any): Boolean

    Permalink
    Definition Classes
    AnyRef → Any
  4. val ac: ADAMContext

    Permalink

    The ADAMContext to wrap.

  5. final def asInstanceOf[T0]: T0

    Permalink
    Definition Classes
    Any
  6. def clone(): AnyRef

    Permalink
    Attributes
    protected[java.lang]
    Definition Classes
    AnyRef
    Annotations
    @throws( ... )
  7. final def eq(arg0: AnyRef): Boolean

    Permalink
    Definition Classes
    AnyRef
  8. def equals(arg0: Any): Boolean

    Permalink
    Definition Classes
    AnyRef → Any
  9. def finalize(): Unit

    Permalink
    Attributes
    protected[java.lang]
    Definition Classes
    AnyRef
    Annotations
    @throws( classOf[java.lang.Throwable] )
  10. final def getClass(): Class[_]

    Permalink
    Definition Classes
    AnyRef → Any
  11. def getSparkContext: JavaSparkContext

    Permalink

    returns

    Returns the Java Spark Context associated with this Java ADAM Context.

  12. def hashCode(): Int

    Permalink
    Definition Classes
    AnyRef → Any
  13. final def isInstanceOf[T0]: Boolean

    Permalink
    Definition Classes
    Any
  14. def loadAlignments(pathName: String, stringency: ValidationStringency): AlignmentDataset

    Permalink

    (Java-specific) Load alignments into an AlignmentDataset.

    (Java-specific) Load alignments into an AlignmentDataset.

    Loads path names ending in: * .bam/.cram/.sam as BAM/CRAM/SAM format, * .fa/.fasta as FASTA format, * .fq/.fastq as FASTQ format, and * .ifq as interleaved FASTQ format.

    If none of these match, fall back to Parquet + Avro.

    For FASTA, FASTQ, and interleaved FASTQ formats, compressed files are supported through compression codecs configured in Hadoop, which by default include .gz and .bz2, but can include more.

    pathName

    The path name to load alignments from. Globs/directories are supported, although file extension must be present for BAM/CRAM/SAM, FASTA, and FASTQ formats.

    stringency

    The validation stringency to use when validating BAM/CRAM/SAM or FASTQ formats.

    returns

    Returns an AlignmentDataset which wraps the genomic dataset of alignments, sequence dictionary representing contigs the alignments may be aligned to, and the read group dictionary for the alignments if one is available.

    See also

    ADAMContext#loadAlignments

  15. def loadAlignments(pathName: String): AlignmentDataset

    Permalink

    (Java-specific) Load alignments into an AlignmentDataset.

    (Java-specific) Load alignments into an AlignmentDataset.

    Loads path names ending in: * .bam/.cram/.sam as BAM/CRAM/SAM format, * .fa/.fasta as FASTA format, * .fq/.fastq as FASTQ format, and * .ifq as interleaved FASTQ format.

    If none of these match, fall back to Parquet + Avro.

    For FASTA, FASTQ, and interleaved FASTQ formats, compressed files are supported through compression codecs configured in Hadoop, which by default include .gz and .bz2, but can include more.

    pathName

    The path name to load alignments from. Globs/directories are supported, although file extension must be present for BAM/CRAM/SAM, FASTA, and FASTQ formats.

    returns

    Returns an AlignmentDataset which wraps the genomic dataset of alignments, sequence dictionary representing contigs the alignments may be aligned to, and the read group dictionary for the alignments if one is available.

    See also

    ADAMContext#loadAlignments

  16. def loadCoverage(pathName: String, stringency: ValidationStringency): CoverageDataset

    Permalink

    (Java-specific) Load features into a FeatureDataset and convert to a CoverageDataset.

    (Java-specific) Load features into a FeatureDataset and convert to a CoverageDataset. Coverage is stored in the score field of Feature.

    Loads path names ending in: * .bed as BED6/12 format, * .gff3 as GFF3 format, * .gtf/.gff as GTF/GFF2 format, * .narrow[pP]eak as NarrowPeak format, and * .interval_list as IntervalList format.

    If none of these match, fall back to Parquet + Avro.

    For BED6/12, GFF3, GTF/GFF2, NarrowPeak, and IntervalList formats, compressed files are supported through compression codecs configured in Hadoop, which by default include .gz and .bz2, but can include more.

    pathName

    The path name to load features from. Globs/directories are supported, although file extension must be present for BED6/12, GFF3, GTF/GFF2, NarrowPeak, or IntervalList formats.

    stringency

    The validation stringency to use when validating BED6/12, GFF3, GTF/GFF2, NarrowPeak, or IntervalList formats.

    returns

    Returns a FeatureDataset converted to a CoverageDataset.

    See also

    ADAMContext#loadCoverage

  17. def loadCoverage(pathName: String): CoverageDataset

    Permalink

    (Java-specific) Load features into a FeatureDataset and convert to a CoverageDataset.

    (Java-specific) Load features into a FeatureDataset and convert to a CoverageDataset. Coverage is stored in the score field of Feature.

    Loads path names ending in: * .bed as BED6/12 format, * .gff3 as GFF3 format, * .gtf/.gff as GTF/GFF2 format, * .narrow[pP]eak as NarrowPeak format, and * .interval_list as IntervalList format.

    If none of these match, fall back to Parquet + Avro.

    For BED6/12, GFF3, GTF/GFF2, NarrowPeak, and IntervalList formats, compressed files are supported through compression codecs configured in Hadoop, which by default include .gz and .bz2, but can include more.

    pathName

    The path name to load features from. Globs/directories are supported, although file extension must be present for BED6/12, GFF3, GTF/GFF2, NarrowPeak, or IntervalList formats.

    returns

    Returns a FeatureDataset converted to a CoverageDataset.

    See also

    ADAMContext#loadCoverage

  18. def loadDnaSequences(pathName: String): SequenceDataset

    Permalink

    (Java-specific) Load DNA sequences into a SequenceDataset.

    (Java-specific) Load DNA sequences into a SequenceDataset.

    If the path name has a .fa/.fasta extension, load as FASTA format. Else, fall back to Parquet + Avro.

    For FASTA format, compressed files are supported through compression codecs configured in Hadoop, which by default include .gz and .bz2, but can include more.

    pathName

    The path name to load sequences from. Globs/directories are supported, although file extension must be present for FASTA format.

    returns

    Returns a SequenceDataset containing DNA sequences.

    See also

    ADAMContext#loadParquetSequences

    ADAMContext#loadFastaDna

  19. def loadFeatures(pathName: String, stringency: ValidationStringency): FeatureDataset

    Permalink

    (Java-specific) Load features into a FeatureDataset.

    (Java-specific) Load features into a FeatureDataset.

    Loads path names ending in: * .bed as BED6/12 format, * .gff3 as GFF3 format, * .gtf/.gff as GTF/GFF2 format, * .narrow[pP]eak as NarrowPeak format, and * .interval_list as IntervalList format.

    If none of these match, fall back to Parquet + Avro.

    For BED6/12, GFF3, GTF/GFF2, NarrowPeak, and IntervalList formats, compressed files are supported through compression codecs configured in Hadoop, which by default include .gz and .bz2, but can include more.

    pathName

    The path name to load features from. Globs/directories are supported, although file extension must be present for BED6/12, GFF3, GTF/GFF2, NarrowPeak, or IntervalList formats.

    stringency

    The validation stringency to use when validating BED6/12, GFF3, GTF/GFF2, NarrowPeak, or IntervalList formats.

    returns

    Returns a FeatureDataset.

    See also

    ADAMContext#loadFeatures

  20. def loadFeatures(pathName: String): FeatureDataset

    Permalink

    (Java-specific) Load features into a FeatureDataset.

    (Java-specific) Load features into a FeatureDataset.

    Loads path names ending in: * .bed as BED6/12 format, * .gff3 as GFF3 format, * .gtf/.gff as GTF/GFF2 format, * .narrow[pP]eak as NarrowPeak format, and * .interval_list as IntervalList format.

    If none of these match, fall back to Parquet + Avro.

    For BED6/12, GFF3, GTF/GFF2, NarrowPeak, and IntervalList formats, compressed files are supported through compression codecs configured in Hadoop, which by default include .gz and .bz2, but can include more.

    pathName

    The path name to load features from. Globs/directories are supported, although file extension must be present for BED6/12, GFF3, GTF/GFF2, NarrowPeak, or IntervalList formats.

    returns

    Returns a FeatureDataset.

    See also

    ADAMContext#loadFeatures

  21. def loadFragments(pathName: String, stringency: ValidationStringency): FragmentDataset

    Permalink

    (Java-specific) Load fragments into a FragmentDataset.

    (Java-specific) Load fragments into a FragmentDataset.

    Loads path names ending in: * .bam/.cram/.sam as BAM/CRAM/SAM format and * .ifq as interleaved FASTQ format.

    If none of these match, fall back to Parquet + Avro.

    For interleaved FASTQ format, compressed files are supported through compression codecs configured in Hadoop, which by default include .gz and .bz2, but can include more.

    pathName

    The path name to load fragments from. Globs/directories are supported, although file extension must be present for BAM/CRAM/SAM and FASTQ formats.

    stringency

    The validation stringency to use when validating BAM/CRAM/SAM or FASTQ formats.

    returns

    Returns a FragmentDataset.

    See also

    ADAMContext#loadFragments

  22. def loadFragments(pathName: String): FragmentDataset

    Permalink

    (Java-specific) Load fragments into a FragmentDataset.

    (Java-specific) Load fragments into a FragmentDataset.

    Loads path names ending in: * .bam/.cram/.sam as BAM/CRAM/SAM format and * .ifq as interleaved FASTQ format.

    If none of these match, fall back to Parquet + Avro.

    For interleaved FASTQ format, compressed files are supported through compression codecs configured in Hadoop, which by default include .gz and .bz2, but can include more.

    pathName

    The path name to load fragments from. Globs/directories are supported, although file extension must be present for BAM/CRAM/SAM and FASTQ formats.

    returns

    Returns a FragmentDataset.

    See also

    ADAMContext#loadFragments

  23. def loadGenotypes(pathName: String, stringency: ValidationStringency): GenotypeDataset

    Permalink

    (Java-specific) Load genotypes into a GenotypeDataset.

    (Java-specific) Load genotypes into a GenotypeDataset.

    If the path name has a .vcf/.vcf.gz/.vcf.bgzf/.vcf.bgz extension, load as VCF format. Else, fall back to Parquet + Avro.

    pathName

    The path name to load genotypes from. Globs/directories are supported, although file extension must be present for VCF format.

    stringency

    The validation stringency to use when validating VCF format.

    returns

    Returns a GenotypeDataset.

    See also

    ADAMContext#loadGenotypes

  24. def loadGenotypes(pathName: String): GenotypeDataset

    Permalink

    (Java-specific) Load genotypes into a GenotypeDataset.

    (Java-specific) Load genotypes into a GenotypeDataset.

    If the path name has a .vcf/.vcf.gz/.vcf.bgzf/.vcf.bgz extension, load as VCF format. Else, fall back to Parquet + Avro.

    pathName

    The path name to load genotypes from. Globs/directories are supported, although file extension must be present for VCF format.

    returns

    Returns a GenotypeDataset.

    See also

    ADAMContext#loadGenotypes

  25. def loadIndexedBam(pathName: String, viewRegions: List[ReferenceRegion], stringency: ValidationStringency): AlignmentDataset

    Permalink

    (Java-specific) Functions like loadBam, but uses BAM index files to look at fewer blocks, and only returns records within the specified ReferenceRegions.

    (Java-specific) Functions like loadBam, but uses BAM index files to look at fewer blocks, and only returns records within the specified ReferenceRegions. BAM index file required.

    pathName

    The path name to load indexed BAM formatted alignments from. Globs/directories are supported.

    viewRegions

    Iterable of ReferenceRegion we are filtering on.

    stringency

    The validation stringency to use when validating the BAM/CRAM/SAM format header. Defaults to ValidationStringency.STRICT.

    returns

    Returns an AlignmentDataset which wraps the genomic dataset of alignments, sequence dictionary representing contigs the alignments may be aligned to, and the read group dictionary for the alignments if one is available.

  26. def loadProteinSequences(pathName: String): SequenceDataset

    Permalink

    (Java-specific) Load protein sequences into a SequenceDataset.

    (Java-specific) Load protein sequences into a SequenceDataset.

    If the path name has a .fa/.fasta extension, load as FASTA format. Else, fall back to Parquet + Avro.

    For FASTA format, compressed files are supported through compression codecs configured in Hadoop, which by default include .gz and .bz2, but can include more.

    pathName

    The path name to load sequences from. Globs/directories are supported, although file extension must be present for FASTA format.

    returns

    Returns a SequenceDataset containing protein sequences.

    See also

    ADAMContext#loadParquetSequences

    ADAMContext#loadFastaProtein

  27. def loadReferenceFile(pathName: String): ReferenceFile

    Permalink

    (Java-specific) Load reference sequences into a broadcastable ReferenceFile.

    (Java-specific) Load reference sequences into a broadcastable ReferenceFile.

    If the path name has a .2bit extension, loads a 2bit file. Else, uses loadSlices to load the reference as an RDD, which is then collected to the driver. Uses a maximum fragment length of 10kbp.

    pathName

    The path name to load reference sequences from. Globs/directories for 2bit format are not supported.

    returns

    Returns a broadcastable ReferenceFile.

    See also

    ADAMContext#loadSlices

  28. def loadReferenceFile(pathName: String, maximumLength: Long): ReferenceFile

    Permalink

    (Java-specific) Load reference sequences into a broadcastable ReferenceFile.

    (Java-specific) Load reference sequences into a broadcastable ReferenceFile.

    If the path name has a .2bit extension, loads a 2bit file. Else, uses loadSlices to load the reference as an RDD, which is then collected to the driver.

    pathName

    The path name to load reference sequences from. Globs/directories for 2bit format are not supported.

    maximumLength

    Maximum fragment length. Defaults to 10000L. Values greater than 1e9 should be avoided.

    returns

    Returns a broadcastable ReferenceFile.

    See also

    ADAMContext#loadSlices

  29. def loadRnaSequences(pathName: String): SequenceDataset

    Permalink

    (Java-specific) Load RNA sequences into a SequenceDataset.

    (Java-specific) Load RNA sequences into a SequenceDataset.

    If the path name has a .fa/.fasta extension, load as FASTA format. Else, fall back to Parquet + Avro.

    For FASTA format, compressed files are supported through compression codecs configured in Hadoop, which by default include .gz and .bz2, but can include more.

    pathName

    The path name to load sequences from. Globs/directories are supported, although file extension must be present for FASTA format.

    returns

    Returns a SequenceDataset containing RNA sequences.

    See also

    ADAMContext#loadParquetSequences

    ADAMContext#loadFastaRna

  30. def loadSlices(pathName: String, maximumLength: Double): SliceDataset

    Permalink

    (R-specific) Load slices into a SliceDataset.

    (R-specific) Load slices into a SliceDataset.

    If the path name has a .fa/.fasta extension, load as DNA in FASTA format. Else, fall back to Parquet + Avro.

    For FASTA format, compressed files are supported through compression codecs configured in Hadoop, which by default include .gz and .bz2, but can include more.

    pathName

    The path name to load DNA slices from. Globs/directories are supported, although file extension must be present for FASTA format.

    maximumLength

    Maximum fragment length, in Double data type to support dispatch from SparkR.

    returns

    Returns a SliceDataset.

  31. def loadSlices(pathName: String, maximumLength: Integer): SliceDataset

    Permalink

    (Java/Python-specific) Load slices into a SliceDataset.

    (Java/Python-specific) Load slices into a SliceDataset.

    If the path name has a .fa/.fasta extension, load as DNA in FASTA format. Else, fall back to Parquet + Avro.

    For FASTA format, compressed files are supported through compression codecs configured in Hadoop, which by default include .gz and .bz2, but can include more.

    pathName

    The path name to load DNA slices from. Globs/directories are supported, although file extension must be present for FASTA format.

    maximumLength

    Maximum slice length, reduced to Integer data type to support dispatch from Python.

    returns

    Returns a SliceDataset.

  32. def loadVariants(pathName: String, stringency: ValidationStringency): VariantDataset

    Permalink

    (Java-specific) Load variants into a VariantDataset.

    (Java-specific) Load variants into a VariantDataset.

    If the path name has a .vcf/.vcf.gz/.vcf.bgzf/.vcf.bgz extension, load as VCF format. Else, fall back to Parquet + Avro.

    pathName

    The path name to load variants from. Globs/directories are supported, although file extension must be present for VCF format.

    stringency

    The validation stringency to use when validating VCF format.

    returns

    Returns a VariantDataset.

    See also

    ADAMContext#loadVariants

  33. def loadVariants(pathName: String): VariantDataset

    Permalink

    (Java-specific) Load variants into a VariantDataset.

    (Java-specific) Load variants into a VariantDataset.

    If the path name has a .vcf/.vcf.gz/.vcf.bgzf/.vcf.bgz extension, load as VCF format. Else, fall back to Parquet + Avro.

    pathName

    The path name to load variants from. Globs/directories are supported, although file extension must be present for VCF format.

    returns

    Returns a VariantDataset.

    See also

    ADAMContext#loadVariants

  34. final def ne(arg0: AnyRef): Boolean

    Permalink
    Definition Classes
    AnyRef
  35. final def notify(): Unit

    Permalink
    Definition Classes
    AnyRef
  36. final def notifyAll(): Unit

    Permalink
    Definition Classes
    AnyRef
  37. final def synchronized[T0](arg0: ⇒ T0): T0

    Permalink
    Definition Classes
    AnyRef
  38. def toString(): String

    Permalink
    Definition Classes
    AnyRef → Any
  39. final def wait(): Unit

    Permalink
    Definition Classes
    AnyRef
    Annotations
    @throws( ... )
  40. final def wait(arg0: Long, arg1: Int): Unit

    Permalink
    Definition Classes
    AnyRef
    Annotations
    @throws( ... )
  41. final def wait(arg0: Long): Unit

    Permalink
    Definition Classes
    AnyRef
    Annotations
    @throws( ... )

Inherited from Serializable

Inherited from Serializable

Inherited from AnyRef

Inherited from Any

Ungrouped