gpt4 book ai didi

hadoop - 关于 Pig 作业 Jar 文件

转载 作者:可可西里 更新时间:2023-11-01 14:59:32 26 4
gpt4 key购买 nike

我正在使用嵌入式 Pig 来实现图形算法。它在本地模式下工作正常。但是在一个完全分布式的 Hadoop 集群中,总是有如下错误信息:(请看最后几行)

2012-11-23 22:00:00,651 [main] INFO org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.JobControlCompiler - creating jar file Job4116346741117365374.jar
2012-11-23 22:00:09,418 [main] INFO org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.JobControlCompiler - jar file Job4116346741117365374.jar created
2012-11-23 22:00:09,423 [main] INFO org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.JobControlCompiler - Setting up multi store job
2012-11-23 22:00:09,431 [main] INFO org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.JobControlCompiler - BytesPerReducer=1000000000 maxReducers=999 totalInputFileSize=296
2012-11-23 22:00:09,431 [main] INFO org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.JobControlCompiler - Neither PARALLEL nor default parallelism is set for this job. Setting number of reducers to 1
2012-11-23 22:00:09,442 [main] INFO org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MapReduceLauncher - 1 map-reduce job(s) waiting for submission.
2012-11-23 22:00:09,949 [main] INFO org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MapReduceLauncher - job null has failed! Stop running all dependent jobs
2012-11-23 22:00:09,949 [main] INFO org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MapReduceLauncher - 100% complete
2012-11-23 22:00:09,992 [main] ERROR org.apache.pig.tools.pigstats.SimplePigStats - ERROR 6015: During execution, encountered a Hadoop error.
2012-11-23 22:00:09,993 [main] ERROR org.apache.pig.tools.pigstats.PigStatsUtil - 1 map reduce job(s) failed!
2012-11-23 22:00:09,994 [main] INFO org.apache.pig.tools.pigstats.SimplePigStats - Script Statistics:

HadoopVersion PigVersion UserId StartedAt FinishedAt Features
0.20.1 0.10.0 jierus 2012-11-23 21:52:38 2012-11-23 22:00:09 HASH_JOIN,GROUP_BY,DISTINCT,FILTER,UNION

Some jobs have failed! Stop running all dependent jobs
Failed Jobs:
JobId Alias Feature Message Outputs
N/A vec_comp,vec_comp_final,vec_comp_tmp HASH_JOIN,MULTI_QUERY Message: java.io.FileNotFoundException: File /tmp/Job4116346741117365374.jar does not exist.
at org.apache.hadoop.fs.RawLocalFileSystem.getFileStatus(RawLocalFileSystem.java:361)
at org.apache.hadoop.fs.FilterFileSystem.getFileStatus(FilterFileSystem.java:245)
at org.apache.hadoop.fs.FileUtil.copy(FileUtil.java:192)
at org.apache.hadoop.fs.FileSystem.copyFromLocalFile(FileSystem.java:1184)
at org.apache.hadoop.fs.FileSystem.copyFromLocalFile(FileSystem.java:1160)
at org.apache.hadoop.fs.FileSystem.copyFromLocalFile(FileSystem.java:1132)

有人知道我的代码或操作的哪一部分是错误的吗?

最佳答案

你好像没有为 pig 指定作业跟踪器(只指定 hdfs 是不够的!)。例如

<property>
<name>mapred.job.tracker</name>
<value>10.xx.xx.99:9001</value>
</property>

关于hadoop - 关于 Pig 作业 Jar 文件,我们在Stack Overflow上找到一个类似的问题: https://stackoverflow.com/questions/13537898/

26 4 0
Copyright 2021 - 2024 cfsdn All Rights Reserved 蜀ICP备2022000587号
广告合作:1813099741@qq.com 6ren.com