object SparkCatalogUtils extends Logging
A shim that defines the interface interact with Spark's catalogs
Linear Supertypes
Ordering
- Alphabetic
- By Inheritance
Inherited
- SparkCatalogUtils
- Logging
- AnyRef
- Any
- Hide All
- Show All
Visibility
- Public
- All
Value Members
-
final
def
!=(arg0: Any): Boolean
- Definition Classes
- AnyRef → Any
-
final
def
##(): Int
- Definition Classes
- AnyRef → Any
-
final
def
==(arg0: Any): Boolean
- Definition Classes
- AnyRef → Any
- val SESSION_CATALOG: String
-
final
def
asInstanceOf[T0]: T0
- Definition Classes
- Any
-
def
clone(): AnyRef
- Attributes
- protected[lang]
- Definition Classes
- AnyRef
- Annotations
- @throws( ... ) @native()
-
def
debug(message: ⇒ Any, t: Throwable): Unit
- Definition Classes
- Logging
-
def
debug(message: ⇒ Any): Unit
- Definition Classes
- Logging
-
final
def
eq(arg0: AnyRef): Boolean
- Definition Classes
- AnyRef
-
def
equals(arg0: Any): Boolean
- Definition Classes
- AnyRef → Any
-
def
error(message: ⇒ Any): Unit
- Definition Classes
- Logging
-
def
error(message: ⇒ Any, t: Throwable): Unit
- Definition Classes
- Logging
-
def
finalize(): Unit
- Attributes
- protected[lang]
- Definition Classes
- AnyRef
- Annotations
- @throws( classOf[java.lang.Throwable] )
- def getCatalog(spark: SparkSession, catalogName: String): CatalogPlugin
- def getCatalogTablesOrViews(spark: SparkSession, catalogName: String, schemaPattern: String, tablePattern: String, tableTypes: Set[String], ignoreTableProperties: Boolean = false): Seq[Row]
-
def
getCatalogs(spark: SparkSession): Seq[Row]
Get all register catalogs in Spark's
CatalogManager -
final
def
getClass(): Class[_]
- Definition Classes
- AnyRef → Any
- Annotations
- @native()
- def getColumns(spark: SparkSession, catalogName: String, schemaPattern: String, tablePattern: String, columnPattern: String): Seq[Row]
-
def
getSchemas(spark: SparkSession, catalogName: String, schemaPattern: String): Seq[Row]
a list of Rows, with 2 fields
schemaName: String, catalogName: String - def getTempViews(spark: SparkSession, catalogName: String, schemaPattern: String, tablePattern: String): Seq[Row]
-
def
hashCode(): Int
- Definition Classes
- AnyRef → Any
- Annotations
- @native()
-
def
info(message: ⇒ Any, t: Throwable): Unit
- Definition Classes
- Logging
-
def
info(message: ⇒ Any): Unit
- Definition Classes
- Logging
-
def
initializeLoggerIfNecessary(isInterpreter: Boolean): Unit
- Attributes
- protected
- Definition Classes
- Logging
-
final
def
isInstanceOf[T0]: Boolean
- Definition Classes
- Any
-
def
logger: Logger
- Attributes
- protected
- Definition Classes
- Logging
-
def
loggerName: String
- Attributes
- protected
- Definition Classes
- Logging
-
final
def
ne(arg0: AnyRef): Boolean
- Definition Classes
- AnyRef
-
final
def
notify(): Unit
- Definition Classes
- AnyRef
- Annotations
- @native()
-
final
def
notifyAll(): Unit
- Definition Classes
- AnyRef
- Annotations
- @native()
- def quoteIdentifier(name: String): String
- def quoteIfNeeded(part: String): String
- def setCurrentCatalog(spark: SparkSession, catalog: String): Unit
- val sparkTableTypes: Set[String]
-
final
def
synchronized[T0](arg0: ⇒ T0): T0
- Definition Classes
- AnyRef
-
def
toString(): String
- Definition Classes
- AnyRef → Any
-
final
def
wait(): Unit
- Definition Classes
- AnyRef
- Annotations
- @throws( ... )
-
final
def
wait(arg0: Long, arg1: Int): Unit
- Definition Classes
- AnyRef
- Annotations
- @throws( ... )
-
final
def
wait(arg0: Long): Unit
- Definition Classes
- AnyRef
- Annotations
- @throws( ... ) @native()
-
def
warn(message: ⇒ Any, t: Throwable): Unit
- Definition Classes
- Logging
-
def
warn(message: ⇒ Any): Unit
- Definition Classes
- Logging