I\'m using HiveContext with SparkSQL and I\'m trying to connect to a remote Hive metastore, the only way to set the hive metastore is through including the hive-site.xml on
In Hadoop 3 Spark and Hive catalogs are separated so:
For spark-shell (it comes with .enableHiveSupport()
by default) just try:
pyspark-shell --conf spark.hadoop.metastore.catalog.default=hive
For spark-submit job create you spark session like this:
SparkSession.builder.appName("Test").enableHiveSupport().getOrCreate()
then add this conf on your spark-submit command:
--conf spark.hadoop.metastore.catalog.default=hive
But for ORC table(and more generally internal table) it is recommended to use HiveWareHouse Connector.