Failed to create Spark client for Spark session

flink58 发布于 03/25 18:39
阅读 16
收藏 0

Total jobs = 1
Launching Job 1 out of 1
In order to change the average load for a reducer (in bytes):
  set hive.exec.reducers.bytes.per.reducer=<number>
In order to limit the maximum number of reducers:
  set hive.exec.reducers.max=<number>
In order to set a constant number of reducers:
  set mapreduce.job.reduces=<number>
FAILED: Execution Error, return code 30041 from org.apache.hadoop.hive.ql.exec.spark.SparkTask. Failed to create Spark client for Spark session 25449a9c-954c-4860-a0f8-8a625e7826d7: java.util.concurrent.TimeoutException: Client '25449a9c-954c-4860-a0f8-8a625e7826d7' timed out waiting for connection from the Remote Spark Driver
 

加载中
返回顶部
顶部