How to extract application ID from the PySpark context

后端 未结 3 1352
广开言路
广开言路 2020-12-17 16:10

A previous question recommends sc.applicationId, but it is not present in PySpark, only in scala.

So, how do I figure

相关标签:
3条回答
  • 2020-12-17 16:32

    In Spark 1.6 (probably 1.5 according to @wladymyrov in comment on the other answer)

    In [1]: sc.applicationId
    Out[1]: u'local-1455827907865'
    
    0 讨论(0)
  • 2020-12-17 16:42

    You could use Java SparkContext object through the Py4J RPC gateway:

    >>> sc._jsc.sc().applicationId()
    u'application_1433865536131_34483'
    

    Please note that sc._jsc is internal variable and not the part of public API - so there is (rather small) chance that it may be changed in the future.

    I'll submit pull request to add public API call for this.

    0 讨论(0)
  • 2020-12-17 16:57

    For PySpark 2.0.0+

    spark_session = SparkSession \
        .builder \
        .enableHiveSupport() \
        .getOrCreate()
    
    app_id = spark_session._sc.applicationId
    
    0 讨论(0)
提交回复
热议问题