简体   繁体   English

胶水作业失败,随机出现“未找到 JohnSnowLabs spark-nlp 依赖项”错误

[英]Glue job failed with `JohnSnowLabs spark-nlp dependency not found` error randomly

I'm using AWS Glue to run some pyspark python code, sometimes it succeeded but sometimes failed with a dependency error: Resource Setup Error: Exception in thread "main" java.lang.RuntimeException: [unresolved dependency: JohnSnowLabs#spark-nlp;2.5.4: not found] , here is the error logs: I'm using AWS Glue to run some pyspark python code, sometimes it succeeded but sometimes failed with a dependency error: Resource Setup Error: Exception in thread "main" java.lang.RuntimeException: [unresolved dependency: JohnSnowLabs#spark-nlp;2.5.4: not found] ,这里是错误日志:

:: problems summary ::
:::: WARNINGS
        module not found: JohnSnowLabs#spark-nlp;2.5.4

    ==== local-m2-cache: tried

      file:/root/.m2/repository/JohnSnowLabs/spark-nlp/2.5.4/spark-nlp-2.5.4.pom

      -- artifact JohnSnowLabs#spark-nlp;2.5.4!spark-nlp.jar:

      file:/root/.m2/repository/JohnSnowLabs/spark-nlp/2.5.4/spark-nlp-2.5.4.jar

    ==== local-ivy-cache: tried

      /root/.ivy2/local/JohnSnowLabs/spark-nlp/2.5.4/ivys/ivy.xml

      -- artifact JohnSnowLabs#spark-nlp;2.5.4!spark-nlp.jar:

      /root/.ivy2/local/JohnSnowLabs/spark-nlp/2.5.4/jars/spark-nlp.jar

    ==== central: tried

      https://repo1.maven.org/maven2/JohnSnowLabs/spark-nlp/2.5.4/spark-nlp-2.5.4.pom

      -- artifact JohnSnowLabs#spark-nlp;2.5.4!spark-nlp.jar:

      https://repo1.maven.org/maven2/JohnSnowLabs/spark-nlp/2.5.4/spark-nlp-2.5.4.jar

    ==== spark-packages: tried

      https://dl.bintray.com/spark-packages/maven/JohnSnowLabs/spark-nlp/2.5.4/spark-nlp-2.5.4.pom

      -- artifact JohnSnowLabs#spark-nlp;2.5.4!spark-nlp.jar:

      https://dl.bintray.com/spark-packages/maven/JohnSnowLabs/spark-nlp/2.5.4/spark-nlp-2.5.4.jar

        ::::::::::::::::::::::::::::::::::::::::::::::

        ::          UNRESOLVED DEPENDENCIES         ::

        ::::::::::::::::::::::::::::::::::::::::::::::

        :: JohnSnowLabs#spark-nlp;2.5.4: not found

        ::::::::::::::::::::::::::::::::::::::::::::::



:: USE VERBOSE OR DEBUG MESSAGE LEVEL FOR MORE DETAILS
Exception in thread "main" java.lang.RuntimeException: [unresolved dependency: JohnSnowLabs#spark-nlp;2.5.4: not found]

From the logs of a successful run, I can see that glue was able to download the dependency from https://dl.bintray.com/spark-packages/maven/JohnSnowLabs/spark-nlp/2.5.4/spark-nlp-2.5.4.pom , where the failed job has also tried to download from, but failed.从成功运行的日志中,我可以看到胶水能够从https://dl.bintray.com/spark-packages/maven/JohnSnowLabs/spark-nlp/2.5.4/spark-nlp-2.5.4.pom下载依赖项https://dl.bintray.com/spark-packages/maven/JohnSnowLabs/spark-nlp/2.5.4/spark-nlp-2.5.4.pom ,失败的作业也尝试从中下载,但失败了。

This issue seems to resolve itself last week, but in the last couple of days it showed up again, and hasn't resolved itself so far.这个问题上周似乎自行解决了,但最近几天又出现了,到目前为止还没有自行解决。 Has anyone ever seen this weird issue?有没有人见过这个奇怪的问题? Thanks.谢谢。

spark-packages moved on May 1 2021. In my scala project I had to add a different resolver like so. spark-packages 于 2021 年 5 月 1 日移动。在我的 scala 项目中,我必须像这样添加一个不同的解析器。 It's got to be similar in java.在 java 中必须是相似的。

resolvers in ThisBuild ++= Seq(
  "SparkPackages" at "https://repos.spark-packages.org"
 ## remove -> "MVNRepository"  at "https://dl.bintray.com/spark-packages/maven"
)

Go look yourself, that package isn't on that resolver you were looking for. Go 看看你自己,package 不在你要找的解析器上。 Mine wasn't either.我的也不是。

https://dl.bintray.com/spark-packages/ https://dl.bintray.com/spark-packages/

暂无
暂无

声明:本站的技术帖子网页,遵循CC BY-SA 4.0协议,如果您需要转载,请注明本站网址或者原文地址。任何问题请咨询:yoyou2525@163.com.

相关问题 java.lang.ClassNotFoundException: com.johnsnowlabs.nlp.DocumentAssembler spark in Pycharm with conda env - java.lang.ClassNotFoundException: com.johnsnowlabs.nlp.DocumentAssembler spark in Pycharm with conda env 在火花中使用stanford nlp,错误“找不到类java.util.function.Function-继续存根。” - using stanford nlp in spark, error “ Class java.util.function.Function not found - continuing with a stub.” 由于不可序列化的对象,Spark作业失败 - Spark job failed due to not serializable objects Spark 安装错误 => 无法初始化编译器:未找到编译器镜像中的对象 java.lang.Object - Spark installation error => Failed to initialize compiler: object java.lang.Object in compiler mirror not found 在管道作业期间找不到托管在 GitLab 的 maven 存储库中的依赖项 - Dependency hosted in GitLab's maven repository is not found during pipelines job 如何解决Spark Streaming程序中的“运行作业流作业时出错”? - How to resolved “Error running job streaming job” in Spark streaming program? 在Spark 0.9.0上运行作业会引发错误 - Running a Job on Spark 0.9.0 throws error kubernetes 上的 Spark 作业失败,没有特定错误 - Spark job on kubernetes fails without specific error 随机无法识别Maven依赖 - Maven dependency randomly not recognized 由于java.io.NotSerializableException:org.apache.spark.SparkContext,Spark作业失败 - Spark job is failed due to java.io.NotSerializableException: org.apache.spark.SparkContext
 
粤ICP备18138465号  © 2020-2024 STACKOOM.COM