I used to be following this documentation and had some experiences I wished to doc.  I wished to perform this.

Determine 1, Exterior Apache Hive metastore utilizing Azure Databricks and Azure SQL

image

Determine 2, Exterior Apache Hive metastore utilizing Azure Databricks and Azure SQL

The cluster model I used was essentially the most present one.

image

Determine 3, Azure Databricks Runtime: 9.1 LTS (Scala 2.12, Spark 3.1.2) cluster

Right here is how my closing Spark config ended up wanting like.

datanucleus.schema.autoCreateTables true

spark.hadoop.javax.jdo.choice.ConnectionUserName benperk@csharpguitar

datanucleus.fixedDatastore false

spark.hadoop.javax.jdo.choice.ConnectionURL jdbc:sqlserver://csharpguitar.database.home windows.internet:1433;database=brainjammer

spark.hadoop.javax.jdo.choice.ConnectionPassword *************

spark.hadoop.javax.jdo.choice.ConnectionDriverName com.microsoft.sqlserver.jdbc.SQLServerDriver

Principally, the datanucleus.autoCreateSchema configuration title is not right.  So I modified datanucleus.autoCreateSchema true to datanucleus.schema.autoCreateTables true

  • Attributable to: MetaException(message:Model info not present in metastore.)
  • Attributable to: javax.jdo.JDODataStoreException: Required desk lacking : “VERSION” in Catalog “” Schema “”. DataNucleus requires this desk to carry out its persistence operations. Both your MetaData is inaccurate, or it’s worthwhile to allow “datanucleus.schema.autoCreateTables

All of those messages occurred as a result of I included these two entries in my Spark config

  • spark.sql.hive.metastore.model 3.1, additionally tried spark.sql.hive.metastore.model 2.3.7
  • spark.sql.hive.metastore.jars builtin
  • AnalysisException: java.lang.RuntimeException: org.apache.hadoop.hive.ql.metadata.HiveException: java.lang.RuntimeException: Unable to instantiate org.apache.hadoop.hive.ql.metadata.SessionHiveMetaStoreClient;
  • Error in SQL assertion: IllegalArgumentException: Builtin jars can solely be used when hive execution model == hive metastore model. Execution: 2.3.7 != Metastore: 3.1. Specify a legitimate path to the proper hive jars utilizing spark.sql.hive.metastore.jars or change spark.sql.hive.metastore.model to 2.3.7.
  • Builtin jars can solely be used when hive execution model == hive metastore model. Execution: 2.3.7 != Metastore: 0.13.0. Specify a legitimate path to the proper hive jars utilizing spark.sql.hive.metastore.jars or change spark.sql.hive.metastore.model to 2.3.7.
  • Builtin jars can solely be used when hive execution model == hive metastore model. Execution: 2.3.7 != Metastore: 0.13.0. Specify a legitimate path to the proper hive jars utilizing spark.sql.hive.metastore.jars or change spark.sql.hive.metastore.model to 2.3.7.

I did discover some info on StackOverflow about including these two strains to the Spark config, which supplied some good info, seems, apparently the title has modified

  • datanucleus.autoCreateSchema true
  • datanucleus.fixedDatastore false