The primary workflow for executing relational queries using Spark.
Caches the specified table in-memory.
:: Experimental ::
Creates an empty parquet file with the schema of class A
, which can be registered as a table.
:: Experimental ::
Creates an empty parquet file with the schema of class A
, which can be registered as a table.
This registered table can be used as the target of future insertInto
operations.
val sqlContext = new SQLContext(...) import sqlContext._ case class Person(name: String, age: Int) createParquetFile[Person]("path/to/file.parquet").registerAsTable("people") sql("INSERT INTO people SELECT 'michael', 29")
A case class type that describes the desired schema of the parquet file to be created.
The path where the directory containing parquet metadata should be created. Data inserted into this table will also be stored at this location.
When false, an exception will be thrown if this directory already exists.
A Hadoop configuration object that can be used to specify options to the parquet output format.
Creates a SchemaRDD from an RDD of case classes.
:: DeveloperApi :: Allows catalyst LogicalPlans to be executed as a SchemaRDD.
:: DeveloperApi :: Allows catalyst LogicalPlans to be executed as a SchemaRDD. Note that the LogicalPlan interface is considered internal, and thus not guaranteed to be stable. As a result, using them directly is not recommended.
Loads a Parquet file, returning the result as a SchemaRDD.
Prepares a planned SparkPlan for execution by binding references to specific ordinals, and inserting shuffle operations as needed.
Prepares a planned SparkPlan for execution by binding references to specific ordinals, and inserting shuffle operations as needed.
Registers the given RDD as a temporary table in the catalog.
Registers the given RDD as a temporary table in the catalog. Temporary tables exist only during the lifetime of this instance of SQLContext.
Executes a SQL query using Spark, returning the result as a SchemaRDD.
Returns the specified table as a SchemaRDD
Removes the specified table from the in-memory cache.
:: AlphaComponent :: The entry point for running relational queries using Spark. Allows the creation of SchemaRDD objects and the execution of SQL queries.