Calling JDBC to impala/hive from within a spark job and creating a table

前端 未结 1 1337
悲哀的现实
悲哀的现实 2021-01-01 02:09

I am trying to write a spark job in scala that would open a jdbc connection with Impala and let me create a table and perform other operations.

How do I do this?

相关标签:
1条回答
  • 2021-01-01 02:42
    val JDBCDriver = "com.cloudera.impala.jdbc41.Driver"
    val ConnectionURL = "jdbc:impala://url.server.net:21050/default;auth=noSasl"
    
    Class.forName(JDBCDriver).newInstance
    val con = DriverManager.getConnection(ConnectionURL)
    val stmt = con.createStatement()
    val rs = stmt.executeQuery(query)
    
    val resultSetList = Iterator.continually((rs.next(), rs)).takeWhile(_._1).map(r => {
        getRowFromResultSet(r._2) // (ResultSet) => (spark.sql.Row)
    }).toList
    
    sc.parallelize(resultSetList)
    
    0 讨论(0)
提交回复
热议问题