hive spark yarn-cluster job fails with: “ClassNotFoundException: org.datanucleus.api.jdo.JDOPersistenceManagerFactory”

点点圈 提交于 2019-12-02 07:21:31
Chris Snow

The solution to this error was to add the jars:

[biadmin@bi4c-xxxxxx-mastermanager ~]$ spark-submit \
    --master yarn-cluster \
    --deploy-mode cluster \
    --jars /usr/iop/4.2.0.0/hive/lib/datanucleus-api-jdo-3.2.6.jar, \
           /usr/iop/4.2.0.0/hive/lib/datanucleus-core-3.2.10.jar, \
           /usr/iop/4.2.0.0/hive/lib/datanucleus-rdbms-3.2.9.jar \
    test_pokes.py

However, I then get a different error:

pyspark.sql.utils.AnalysisException: u'Table not found: pokes; line 1 pos 14'

I've added the other question here: Spark Hive reporting pyspark.sql.utils.AnalysisException: u'Table not found: XXX' when run on yarn cluster

The final solution is captured here: https://stackoverflow.com/a/41272260/1033422

易学教程内所有资源均来自网络或用户发布的内容,如有违反法律规定的内容欢迎反馈
该文章没有解决你所遇到的问题?点击提问,说说你的问题,让更多的人一起探讨吧!