hadoop-spark-pig-hive
hadoop-spark-pig-hive copied to clipboard
Failed to get reduce task report after execution of pig in mapreduce mode
I want to run pig in mapreduce mode. But I am getting this error.
2020-12-26 00:18:55,435 [main] INFO org.apache.hadoop.mapred.ClientServiceDelegate - Application state is completed. FinalApplicationStatus=SUCCEEDED. Redirecting to job history server
2020-12-26 00:18:56,440 [main] INFO org.apache.hadoop.ipc.Client - Retrying connect to server: 0.0.0.0/0.0.0.0:10020. Already tried 0 time(s); retry policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1000 MILLISECONDS)
2020-12-26 00:18:57,442 [main] INFO org.apache.hadoop.ipc.Client - Retrying connect to server: 0.0.0.0/0.0.0.0:10020. Already tried 1 time(s); retry policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1000 MILLISECONDS)
2020-12-26 00:19:05,560 [main] WARN org.apache.pig.tools.pigstats.mapreduce.MRJobStats - Failed to get reduce task report
Is this is an configuration issue? Please support
It seems like you're encountering an issue related to connecting to the Hadoop JobHistory Server, which is causing Pig to fail when trying to retrieve the complete status of the MapReduce tasks. This typically happens when the JobHistory Server is not running or hasn't been started properly.
To fix this issue, make sure the JobHistory Server is up and running. You can do this by executing the following command within your Hadoop environment:
$HADOOP_HOME/bin/mapred historyserver
This command should start the JobHistory Server, allowing Pig to connect properly and avoid the connection errors you're seeing.
Once the JobHistory Server is running, try executing your Pig script again and see if the issue is resolved.