/tags/release-0.0.0-rc0/hive/external/ql/src/test/results/clientnegative/archive2.q.out
text | 33 lines | 31 code | 2 blank | 0 comment | 0 complexity | 0d4bb5fa53718c9ff3c6a8c58944da54 MD5 | raw file
Possible License(s): Apache-2.0, BSD-3-Clause, JSON, CPL-1.0
- PREHOOK: query: -- Tests trying to unarchive a non-archived partition
- -- EXCLUDE_HADOOP_MAJOR_VERSIONS(0.17, 0.18, 0.19)
- drop table tstsrcpart
- PREHOOK: type: DROPTABLE
- POSTHOOK: query: -- Tests trying to unarchive a non-archived partition
- -- EXCLUDE_HADOOP_MAJOR_VERSIONS(0.17, 0.18, 0.19)
- drop table tstsrcpart
- POSTHOOK: type: DROPTABLE
- PREHOOK: query: create table tstsrcpart like srcpart
- PREHOOK: type: CREATETABLE
- POSTHOOK: query: create table tstsrcpart like srcpart
- POSTHOOK: type: CREATETABLE
- POSTHOOK: Output: default@tstsrcpart
- PREHOOK: query: insert overwrite table tstsrcpart partition (ds='2008-04-08', hr='12')
- select key, value from srcpart where ds='2008-04-08' and hr='12'
- PREHOOK: type: QUERY
- PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12
- PREHOOK: Output: default@tstsrcpart@ds=2008-04-08/hr=12
- POSTHOOK: query: insert overwrite table tstsrcpart partition (ds='2008-04-08', hr='12')
- select key, value from srcpart where ds='2008-04-08' and hr='12'
- POSTHOOK: type: QUERY
- POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12
- POSTHOOK: Output: default@tstsrcpart@ds=2008-04-08/hr=12
- POSTHOOK: Lineage: tstsrcpart PARTITION(ds=2008-04-08,hr=12).key SIMPLE [(srcpart)srcpart.FieldSchema(name:key, type:string, comment:default), ]
- POSTHOOK: Lineage: tstsrcpart PARTITION(ds=2008-04-08,hr=12).value SIMPLE [(srcpart)srcpart.FieldSchema(name:value, type:string, comment:default), ]
- PREHOOK: query: ALTER TABLE tstsrcpart UNARCHIVE PARTITION (ds='2008-04-08', hr='12')
- PREHOOK: type: ALTERTABLE_UNARCHIVE
- PREHOOK: Input: default@tstsrcpart
- PREHOOK: Output: default@tstsrcpart@ds=2008-04-08/hr=12
- FAILED: Error in metadata: Specified partition is not archived
- FAILED: Execution Error, return code 1 from org.apache.hadoop.hive.ql.exec.DDLTask