Class

org.apache.spark.sql.sampling

ColumnFormatSamplingRelation

Related Doc: package sampling

Permalink

case class ColumnFormatSamplingRelation(schema: StructType, baseRelation: BaseColumnFormatRelation, sqlContext: SQLContext, samplingOptions: Map[String, String], origOptions: Map[String, String], sampleTable: String, baseSchema: StructType, baseTable: Option[String], reservoirRegionName: String, basePartitioningColumns: Seq[String], baseNumPartition: Int) extends BaseRelation with SamplingRelation with PartitionedDataSourceScan with SchemaInsertableRelation with DestroyRelation with Logging with Serializable with Product

Linear Supertypes
Ordering
  1. Alphabetic
  2. By Inheritance
Inherited
  1. ColumnFormatSamplingRelation
  2. Product
  3. Equals
  4. Serializable
  5. Serializable
  6. Logging
  7. DestroyRelation
  8. PartitionedDataSourceScan
  9. PrunedUnsafeFilteredScan
  10. SamplingRelation
  11. SchemaInsertableRelation
  12. InsertableRelation
  13. BaseRelation
  14. AnyRef
  15. Any
  1. Hide All
  2. Show All
Visibility
  1. Public
  2. All

Instance Constructors

  1. new ColumnFormatSamplingRelation(schema: StructType, baseRelation: BaseColumnFormatRelation, sqlContext: SQLContext, samplingOptions: Map[String, String], origOptions: Map[String, String], sampleTable: String, baseSchema: StructType, baseTable: Option[String], reservoirRegionName: String, basePartitioningColumns: Seq[String], baseNumPartition: Int)

    Permalink

Value Members

  1. final def !=(arg0: Any): Boolean

    Permalink
    Definition Classes
    AnyRef → Any
  2. final def ##(): Int

    Permalink
    Definition Classes
    AnyRef → Any
  3. final def ==(arg0: Any): Boolean

    Permalink
    Definition Classes
    AnyRef → Any
  4. final val PARTITIONBUFSIZE: Int(10)

    Permalink
  5. def append(rows: RDD[Row], time: Long): Unit

    Permalink

    Append a given RDD or rows into the relation.

    Append a given RDD or rows into the relation.

    Definition Classes
    ColumnFormatSamplingRelationSchemaInsertableRelation
  6. final def asInstanceOf[T0]: T0

    Permalink
    Definition Classes
    Any
  7. val baseNumPartition: Int

    Permalink
  8. val basePartitioningColumns: Seq[String]

    Permalink
  9. val baseRelation: BaseColumnFormatRelation

    Permalink

    The underlying column table used to store data.

    The underlying column table used to store data.

    Definition Classes
    ColumnFormatSamplingRelationSamplingRelation
  10. val baseSchema: StructType

    Permalink
  11. val baseTable: Option[String]

    Permalink

    Base table of this relation.

    Base table of this relation.

    Definition Classes
    ColumnFormatSamplingRelationSamplingRelation
  12. def buildUnsafeScan(requiredColumns: Array[String], filters: Array[Expression]): (RDD[Any], Seq[RDD[InternalRow]])

    Permalink
  13. def canBeOnBuildSide: Boolean

    Permalink
  14. def clone(): AnyRef

    Permalink
    Attributes
    protected[java.lang]
    Definition Classes
    AnyRef
    Annotations
    @throws( ... )
  15. def connectionType: execution.columnar.ConnectionType.Value

    Permalink
  16. def destroy(ifExists: Boolean): Unit

    Permalink

    Destroy and cleanup this relation.

    Destroy and cleanup this relation. It may include, but not limited to, dropping the external table that this relation represents.

    Definition Classes
    ColumnFormatSamplingRelationDestroyRelation
  17. final def eq(arg0: AnyRef): Boolean

    Permalink
    Definition Classes
    AnyRef
  18. def finalize(): Unit

    Permalink
    Attributes
    protected[java.lang]
    Definition Classes
    AnyRef
    Annotations
    @throws( classOf[java.lang.Throwable] )
  19. def flushReservoir(): Unit

    Permalink
  20. final def getClass(): Class[_]

    Permalink
    Definition Classes
    AnyRef → Any
  21. def getColocatedTable: Option[String]

    Permalink
  22. def getExternalStoreMetaData: ExternalTableMetaData

    Permalink
  23. def initializeLogIfNecessary(): Unit

    Permalink
    Attributes
    protected
    Definition Classes
    Logging
  24. def insert(data: DataFrame, overwrite: Boolean): Unit

    Permalink
  25. def insertableRelation(sourceSchema: Seq[Attribute]): Option[InsertableRelation]

    Permalink

    Return the actual relation to be used for insertion into the relation or None if sourceSchema cannot be inserted.

    Return the actual relation to be used for insertion into the relation or None if sourceSchema cannot be inserted.

    Definition Classes
    ColumnFormatSamplingRelationSchemaInsertableRelation
  26. final def isDebugEnabled: Boolean

    Permalink
    Definition Classes
    Logging
  27. final def isInfoEnabled: Boolean

    Permalink
    Definition Classes
    Logging
  28. final def isInstanceOf[T0]: Boolean

    Permalink
    Definition Classes
    Any
  29. val isPartitioned: Boolean

    Permalink

    If underlying sample table is partitioned

    If underlying sample table is partitioned

    Definition Classes
    ColumnFormatSamplingRelationPartitionedDataSourceScanSamplingRelation
  30. val isReservoirAsRegion: Boolean

    Permalink

    True if underlying sample table is using a row table as reservoir store.

    True if underlying sample table is using a row table as reservoir store.

    Definition Classes
    ColumnFormatSamplingRelationSamplingRelation
  31. final def isTraceEnabled: Boolean

    Permalink
    Definition Classes
    Logging
  32. final var levelFlags: Int

    Permalink
    Attributes
    protected
    Definition Classes
    Logging
  33. def log: Logger

    Permalink
    Attributes
    protected
    Definition Classes
    Logging
  34. def logDebug(msg: ⇒ String, throwable: Throwable): Unit

    Permalink
    Definition Classes
    Logging
  35. def logDebug(msg: ⇒ String): Unit

    Permalink
    Definition Classes
    Logging
  36. def logError(msg: ⇒ String, throwable: Throwable): Unit

    Permalink
    Definition Classes
    Logging
  37. def logError(msg: ⇒ String): Unit

    Permalink
    Definition Classes
    Logging
  38. def logInfo(msg: ⇒ String, throwable: Throwable): Unit

    Permalink
    Definition Classes
    Logging
  39. def logInfo(msg: ⇒ String): Unit

    Permalink
    Definition Classes
    Logging
  40. def logName: String

    Permalink
    Attributes
    protected
    Definition Classes
    Logging
  41. def logTrace(msg: ⇒ String, throwable: Throwable): Unit

    Permalink
    Definition Classes
    Logging
  42. def logTrace(msg: ⇒ String): Unit

    Permalink
    Definition Classes
    Logging
  43. def logWarning(msg: ⇒ String, throwable: Throwable): Unit

    Permalink
    Definition Classes
    Logging
  44. def logWarning(msg: ⇒ String): Unit

    Permalink
    Definition Classes
    Logging
  45. final var log_: Logger

    Permalink
    Attributes
    protected
    Definition Classes
    Logging
  46. final def ne(arg0: AnyRef): Boolean

    Permalink
    Definition Classes
    AnyRef
  47. val needConversion: Boolean

    Permalink

    Whether does it need to convert the objects in Row to internal representation, for example: java.lang.String to UTF8String java.lang.Decimal to Decimal

    Whether does it need to convert the objects in Row to internal representation, for example: java.lang.String to UTF8String java.lang.Decimal to Decimal

    If needConversion is false, buildScan() should return an RDD of InternalRow

    Definition Classes
    ColumnFormatSamplingRelationBaseRelation
    Since

    1.4.0

    Note

    The internal representation is not stable across releases and thus data sources outside of Spark SQL should leave this as true.

  48. final def notify(): Unit

    Permalink
    Definition Classes
    AnyRef
  49. final def notifyAll(): Unit

    Permalink
    Definition Classes
    AnyRef
  50. def numBuckets: Int

    Permalink
  51. lazy val options: SampleOptions

    Permalink
  52. val origOptions: Map[String, String]

    Permalink
  53. def partitionColumns: Seq[String]

    Permalink
  54. def qcs: Array[String]

    Permalink

    The QCS columns for the sample.

    The QCS columns for the sample.

    Definition Classes
    ColumnFormatSamplingRelationSamplingRelation
  55. lazy val region: LocalRegion

    Permalink
  56. val reservoirRegionName: String

    Permalink
  57. def resetLogger(): Unit

    Permalink
    Attributes
    protected
    Definition Classes
    Logging
  58. val sampleTable: String

    Permalink
  59. val samplingOptions: Map[String, String]

    Permalink

    Options set for this sampling relation.

    Options set for this sampling relation.

    Definition Classes
    ColumnFormatSamplingRelationSamplingRelation
  60. val schema: StructType

    Permalink
  61. def sizeInBytes: Long

    Permalink

    Returns an estimated size of this relation in bytes.

    Returns an estimated size of this relation in bytes. This information is used by the planner to decide when it is safe to broadcast a relation and can be overridden by sources that know the size ahead of time. By default, the system will assume that tables are too large to broadcast. This method will be called multiple times during query planning and thus should not perform expensive operations for each invocation.

    Definition Classes
    ColumnFormatSamplingRelationBaseRelation
    Since

    1.3.0

    Note

    It is always better to overestimate size than underestimate, because underestimation could lead to execution plans that are suboptimal (i.e. broadcasting a very large table).

  62. val sqlContext: SQLContext

    Permalink
  63. final def synchronized[T0](arg0: ⇒ T0): T0

    Permalink
    Definition Classes
    AnyRef
  64. def table: String

    Permalink
  65. def toString(): String

    Permalink
    Definition Classes
    ColumnFormatSamplingRelation → AnyRef → Any
  66. def truncate(): Unit

    Permalink

    Truncate the table represented by this relation.

    Truncate the table represented by this relation.

    Definition Classes
    ColumnFormatSamplingRelationDestroyRelation
  67. def unhandledFilters(filters: Seq[Expression]): Seq[Expression]

    Permalink

    Returns the list of Expressions that this datasource may not be able to handle.

    Returns the list of Expressions that this datasource may not be able to handle. By default, this function will return all filters, as it is always safe to double evaluate an Expression.

    Definition Classes
    ColumnFormatSamplingRelationPrunedUnsafeFilteredScan
  68. def unhandledFilters(filters: Array[Filter]): Array[Filter]

    Permalink

    Returns the list of Filters that this datasource may not be able to handle.

    Returns the list of Filters that this datasource may not be able to handle. These returned Filters will be evaluated by Spark SQL after data is output by a scan. By default, this function will return all filters, as it is always safe to double evaluate a Filter. However, specific implementations can override this function to avoid double filtering when they are capable of processing a filter internally.

    Definition Classes
    BaseRelation
    Since

    1.6.0

  69. final def wait(): Unit

    Permalink
    Definition Classes
    AnyRef
    Annotations
    @throws( ... )
  70. final def wait(arg0: Long, arg1: Int): Unit

    Permalink
    Definition Classes
    AnyRef
    Annotations
    @throws( ... )
  71. final def wait(arg0: Long): Unit

    Permalink
    Definition Classes
    AnyRef
    Annotations
    @throws( ... )

Inherited from Product

Inherited from Equals

Inherited from Serializable

Inherited from Serializable

Inherited from Logging

Inherited from DestroyRelation

Inherited from PartitionedDataSourceScan

Inherited from PrunedUnsafeFilteredScan

Inherited from SamplingRelation

Inherited from SchemaInsertableRelation

Inherited from InsertableRelation

Inherited from BaseRelation

Inherited from AnyRef

Inherited from Any

Ungrouped