备注:
Hive 版本 2.1.1
一.模拟误删表
误删除了这张表
hive>
> drop table ods_fact_sale_orc;
OK
二.从回收站恢复表
查看回收表
[root@hp1 ~]# hadoop fs -ls /user/root/.Trash/Current/user/hive/warehouse/test.db
Found 2 items
drwxrwxrwt - root hive 0 2020-12-02 19:18 /user/root/.Trash/Current/user/hive/warehouse/test.db/dm_sale_orc
drwxrwxrwt - root hive 0 2020-12-02 18:11 /user/root/.Trash/Current/user/hive/warehouse/test.db/ods_fact_sale_orc
从回收站拷贝数据
[root@hp1 ~]# hadoop fs -cp /user/root/.Trash/Current/user/hive/warehouse/test.db/ods_fact_sale_orc /user/hive/warehouse/ods_fact_sale_orc
[root@hp1 ~]#
恢复数据并验证
[root@hp1 ~]# hive
WARNING: Use "yarn jar" to launch YARN applications.
SLF4J: Class path contains multiple SLF4J bindings.
SLF4J: Found binding in [jar:file:/opt/cloudera/parcels/CDH-6.3.1-1.cdh6.3.1.p0.1470567/jars/log4j-slf4j-impl-2.8.2.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: Found binding in [jar:file:/opt/cloudera/parcels/CDH-6.3.1-1.cdh6.3.1.p0.1470567/jars/slf4j-log4j12-1.7.25.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an explanation.
SLF4J: Actual binding is of type [org.apache.logging.slf4j.Log4jLoggerFactory]
Logging initialized using configuration in jar:file:/opt/cloudera/parcels/CDH-6.3.1-1.cdh6.3.1.p0.1470567/jars/hive-common-2.1.1-cdh6.3.1.jar!/hive-log4j2.properties Async: false
WARNING: Hive CLI is deprecated and migration to Beeline is recommended.
hive> use test;
OK
Time taken: 1.184 seconds
hive> load data inpath '/user/hive/warehouse/ods_fact_sale_orc' into table ods_fact_sale_orc;
FAILED: SemanticException [Error 10001]: Line 1:69 Table not found 'ods_fact_sale_orc'
hive>
> CREATE TABLE ods_fact_sale_orc(
> id bigint,
> sale_date string,
> prod_name string,
> sale_nums int)
> STORED AS ORC;
OK
Time taken: 0.396 seconds
hive>
> load data inpath '/user/hive/warehouse/ods_fact_sale_orc' into table ods_fact_sale_orc;
Loading data to table test.ods_fact_sale_orc
OK
Time taken: 1.852 seconds
hive> select count(*) from ods_fact_sale_orc;
Query ID = root_20201202193108_77c436b0-2b3e-47a8-a6e3-23f416f3bfd6
Total jobs = 1
Launching Job 1 out of 1
Number of reduce tasks determined at compile time: 1
In order to change the average load for a reducer (in bytes):
set hive.exec.reducers.bytes.per.reducer=<number>
In order to limit the maximum number of reducers:
set hive.exec.reducers.max=<number>
In order to set a constant number of reducers:
set mapreduce.job.reduces=<number>
Starting Job = job_1606698967173_0016, Tracking URL = http://hp1:8088/proxy/application_1606698967173_0016/
Kill Command = /opt/cloudera/parcels/CDH-6.3.1-1.cdh6.3.1.p0.1470567/lib/hadoop/bin/hadoop job -kill job_1606698967173_0016
Hadoop job information for Stage-1: number of mappers: 9; number of reducers: 1
2020-12-02 19:31:18,259 Stage-1 map = 0%, reduce = 0%
2020-12-02 19:31:26,557 Stage-1 map = 22%, reduce = 0%, Cumulative CPU 10.36 sec
2020-12-02 19:31:30,690 Stage-1 map = 33%, reduce = 0%, Cumulative CPU 14.55 sec
2020-12-02 19:31:31,720 Stage-1 map = 44%, reduce = 0%, Cumulative CPU 19.62 sec
2020-12-02 19:31:36,873 Stage-1 map = 56%, reduce = 0%, Cumulative CPU 24.96 sec
2020-12-02 19:31:37,901 Stage-1 map = 67%, reduce = 0%, Cumulative CPU 30.11 sec
2020-12-02 19:31:43,045 Stage-1 map = 78%, reduce = 0%, Cumulative CPU 35.28 sec
2020-12-02 19:31:44,073 Stage-1 map = 89%, reduce = 0%, Cumulative CPU 39.97 sec
2020-12-02 19:31:49,213 Stage-1 map = 100%, reduce = 0%, Cumulative CPU 45.53 sec
2020-12-02 19:31:51,273 Stage-1 map = 100%, reduce = 100%, Cumulative CPU 48.22 sec
MapReduce Total cumulative CPU time: 48 seconds 220 msec
Ended Job = job_1606698967173_0016
MapReduce Jobs Launched:
Stage-Stage-1: Map: 9 Reduce: 1 Cumulative CPU: 48.22 sec HDFS Read: 1991747 HDFS Write: 109 HDFS EC Read: 0 SUCCESS
Total MapReduce CPU Time Spent: 48 seconds 220 msec
OK
767830000
Time taken: 43.846 seconds, Fetched: 1 row(s)
hive>
网友评论