运行 hive 查询时出错
我正在尝试运行 hive 查询,但收到错误如下:
hive> FROM (
> FROM t1
> MAP t1.patient_mrn, t1.encounter_date
> USING 'retrieve'
> AS mp1, mp2
> CLUSTER BY mp1) map_output
> INSERT OVERWRITE TABLE t3
> REDUCE map_output.mp1, map_output.mp2
> USING 'q1.txt'
> AS reducef1, reducef2;
MapReduce 作业总数 = 1
正在启动作业 1 of 1
未指定的减少任务数。根据输入数据大小估计:1
为了更改减速器的平均负载(以字节为单位): set hive.exec.reducers.bytes.per.reducer=
为了限制reducer的最大数量: set hive.exec.reducers.max=
为了设置恒定数量的减速器: 设置mapred.reduce.tasks=
开始作业= job_201112281627_0097,跟踪URL= http://localhost:50030/jobdetails.jsp?jobid=job_201112281627_0097
终止命令 = /home/hadoop/hadoop-0.20.2-cdh3u2 //bin/hadoop 作业 -Dmapred.job .tracker=localhost:54311 -kill job_201112281627_0097
2011-12-31 03:10:46,391 Stage-1 映射 = 0%,reduce = 0%
2011-12-31 03:11:29,794 Stage-1 映射 = 100%,reduce = 100%
已结束作业 = job_201112281627_0097 有错误 失败:执行错误,从 org.apache.hadoop.hive.ql.exec.MapRedTask 返回代码 2 蜂巢>
I am trying to run hive queries but I am getting errors as:
hive> FROM (
> FROM t1
> MAP t1.patient_mrn, t1.encounter_date
> USING 'retrieve'
> AS mp1, mp2
> CLUSTER BY mp1) map_output
> INSERT OVERWRITE TABLE t3
> REDUCE map_output.mp1, map_output.mp2
> USING 'q1.txt'
> AS reducef1, reducef2;
Total MapReduce jobs = 1
Launching Job 1 out of 1
Number of reduce tasks not specified. Estimated from input data size: 1
In order to change the average load for a reducer (in bytes):
set hive.exec.reducers.bytes.per.reducer=
In order to limit the maximum number of reducers:
set hive.exec.reducers.max=
In order to set a constant number of reducers:
set mapred.reduce.tasks=
Starting Job = job_201112281627_0097, Tracking URL = http://localhost:50030/jobdetails.jsp?jobid=job_201112281627_0097
Kill Command = /home/hadoop/hadoop-0.20.2-cdh3u2//bin/hadoop job -Dmapred.job.tracker=localhost:54311 -kill job_201112281627_0097
2011-12-31 03:10:46,391 Stage-1 map = 0%, reduce = 0%
2011-12-31 03:11:29,794 Stage-1 map = 100%, reduce = 100%
Ended Job = job_201112281627_0097 with errors
FAILED: Execution Error, return code 2 from org.apache.hadoop.hive.ql.exec.MapRedTask
hive>
如果你对这篇内容有疑问,欢迎到本站社区发帖提问 参与讨论,获取更多帮助,或者扫码二维码加入 Web 技术交流群。

绑定邮箱获取回复消息
由于您还没有绑定你的真实邮箱,如果其他用户或者作者回复了您的评论,将不能在第一时间通知您!
发布评论
评论(2)
在不了解更多信息的情况下,最好的建议是在哪里找到错误日志。因此,请转到 JobTracker 的网页,找到该作业的页面,然后深入查找错误日志。
查找任何“失败”的任务,单击此处即可访问该特定任务的页面。
您最终将到达包含特定于任务的日志的页面,这应该可以帮助您诊断问题。
Best advice without knowing a lot more is where to find the error logs. So go to your JobTracker's web page, find the page for that job, and drill down to find the error logs.
Look for any "failed" tasks, click there to get to the page for that specific task.
You'll eventually get to the page containing the task-specific log, and that should help you diagnose the problem.
这可能在多种情况下发生。再次重新运行查询并检查作业跟踪器是否有失败/被终止的尝试,并查看日志以了解确切原因。
This could happen in n number of scenarios. Rerun the query once more and check the jobtracker for the failed/killed attempts and go through the logs for exact reason.