码迷,mamicode.com
首页 > 其他好文 > 详细

Hive数据导入导出

时间:2015-05-18 14:52:22      阅读:143      评论:0      收藏:0      [点我收藏+]

标签:

一、     从文件系统导入

数据源存放路径: /root/data

hive> load data local inpath "/root/data" overwrite intotable t1; Loading data to table default.t1Table default.t1 stats: [numFiles=1, numRows=0, totalSize=30,rawDataSize=0]OKTime taken: 1.712 secondshive> select * from t1;OKzhangsan        25lisi    27

wangwu  24

二、     HDFS导入

Hdfs数据存放位置

[root@crxy177 ~]# hadoop dfs-ls /

-rw-r--r--   1 root supergroup         30 2015-05-18 10:39 /data

hive> load data inpath"/data" overwrite into table t1; Loading data to table default.t1Moved:‘hdfs://192.168.1.177:9000/user/hive/warehouse/t1/data‘ to trash at:hdfs://192.168.1.177:9000/user/root/.Trash/CurrentTable default.t1 stats: [numFiles=1,numRows=0, totalSize=30, rawDataSize=0]OKTime taken: 1.551 seconds三、     通过查询导入

创建一张表

hive> create table t2 like t1;

OK

Time taken: 0.246 seconds

导入数据

hive> insert overwrite table t2 select * form t1;

FAILED: NullPointerException null

hive> insert overwrite table t2 select * from t1;

Query ID = root_20150518104747_7922f9d4-2e15-434a-8b9f-076393d73470

Total jobs = 3

Launching Job 1 out of 3

Number of reduce tasks is set to 0 since there‘s no reduce operator

Starting Job = job_1431916152610_0001, Tracking URL = http://crxy177:8088/proxy/application_1431916152610_0001/

Kill Command = /usr/local/hadoop-2.6.0/bin/hadoop job  -kill job_1431916152610_0001

Interrupting... Be patient, this might take some time.

Press Ctrl+C again to kill JVM

killing job with: job_1431916152610_0001

Hadoop job information for Stage-1: number of mappers: 0; number ofreducers: 0

2015-05-18 10:47:40,679 Stage-1 map = 0%,  reduce = 0%

Ended Job = job_1431916152610_0001 with errors

Error during job, obtaining debugging information...

FAILED: Execution Error, return code 2 fromorg.apache.hadoop.hive.ql.exec.mr.MapRedTask

MapReduce Jobs Launched:

Stage-Stage-1:  HDFS Read: 0HDFS Write: 0 FAIL

Total MapReduce CPU Time Spent: 0 msec

四、     多表同时导入

创建t3,t4表

hive> createtable t3 like t1;

OK

Time taken:1.235 seconds

hive> createtable t4 like t1;

OK

Time taken:0.211 seconds

多表数据导入

hive> FROM t1

    > INSERT OVERWRITE TABLE t2 SELECT * WHERE 1=1

    > INSERT OVERWRITE TABLE t3 SELECT * WHERE 1=1

    > INSERT OVERWRITE TABLE t4 SELECT * WHERE 1=1;

Query ID =root_20150518105252_9101659d-0990-4626-a4f7-8bad768af48b

Total jobs = 7

Launching Job 1out of 7

Number of reducetasks is set to 0 since there‘s no reduce operator

Starting Job =job_1431916152610_0002, Tracking URL = http://crxy177:8088/proxy/application_1431916152610_0002/

Kill Command =/usr/local/hadoop-2.6.0/bin/hadoop job -kill job_1431916152610_0002

Hadoop jobinformation for Stage-3: number of mappers: 1; number of reducers: 0

2015-05-1810:52:50,866 Stage-3 map = 0%,  reduce =0%

2015-05-1810:53:02,273 Stage-3 map = 100%,  reduce= 0%, Cumulative CPU 1.41 sec

MapReduce Totalcumulative CPU time: 1 seconds 410 msec

Ended Job =job_1431916152610_0002

Stage-6 isselected by condition resolver.

Stage-5 isfiltered out by condition resolver.

Stage-7 isfiltered out by condition resolver.

Stage-12 isselected by condition resolver.

Stage-11 isfiltered out by condition resolver.

Stage-13 isfiltered out by condition resolver.

Stage-18 isselected by condition resolver.

Stage-17 isfiltered out by condition resolver.

Stage-19 isfiltered out by condition resolver.

Moving data to:hdfs://192.168.1.177:9000/tmp/hive/root/88e075ab-e7da-497d-a56b-74f652f3eae6/hive_2015-05-18_10-52-30_865_4936011539493382740-1/-ext-10000

Moving data to:hdfs://192.168.1.177:9000/tmp/hive/root/88e075ab-e7da-497d-a56b-74f652f3eae6/hive_2015-05-18_10-52-30_865_4936011539493382740-1/-ext-10002

Moving data to:hdfs://192.168.1.177:9000/tmp/hive/root/88e075ab-e7da-497d-a56b-74f652f3eae6/hive_2015-05-18_10-52-30_865_4936011539493382740-1/-ext-10004

Loading data totable default.t2

Loading data totable default.t3

Loading data totable default.t4

Table default.t2stats: [numFiles=1, numRows=0, totalSize=30, rawDataSize=0]

Table default.t3stats: [numFiles=1, numRows=0, totalSize=30, rawDataSize=0]

Table default.t4stats: [numFiles=1, numRows=0, totalSize=30, rawDataSize=0]

MapReduce JobsLaunched:

Stage-Stage-3:Map: 1   Cumulative CPU: 1.41 sec   HDFS Read: 237 HDFS Write: 288 SUCCESS

Total MapReduceCPU Time Spent: 1 seconds 410 msec

OK

Time taken:34.245 seconds

更多精彩内容请关注:http://bbs.superwu.cn

更多精彩内容请关注超人学院微信二维码: 技术分享

Hive数据导入导出

标签:

原文地址:http://my.oschina.net/crxy/blog/416180

(0)
(0)
   
举报
评论 一句话评论(0
登录后才能评论!
© 2014 mamicode.com 版权所有  联系我们:gaon5@hotmail.com
迷上了代码!