PageRenderTime 42ms CodeModel.GetById 15ms RepoModel.GetById 0ms app.codeStats 0ms

/tags/release-0.0.0-rc0/hive/external/contrib/src/test/results/clientpositive/serde_typedbytes4.q.out

#
text | 245 lines | 239 code | 6 blank | 0 comment | 0 complexity | 55859825f43598f4e92617a9bb20e5f5 MD5 | raw file
Possible License(s): Apache-2.0, BSD-3-Clause, JSON, CPL-1.0
  1. PREHOOK: query: drop table dest1
  2. PREHOOK: type: DROPTABLE
  3. POSTHOOK: query: drop table dest1
  4. POSTHOOK: type: DROPTABLE
  5. PREHOOK: query: CREATE TABLE dest1(key STRING, value STRING) STORED AS TEXTFILE
  6. PREHOOK: type: CREATETABLE
  7. POSTHOOK: query: CREATE TABLE dest1(key STRING, value STRING) STORED AS TEXTFILE
  8. POSTHOOK: type: CREATETABLE
  9. POSTHOOK: Output: default@dest1
  10. PREHOOK: query: EXPLAIN
  11. FROM (
  12. FROM src
  13. SELECT TRANSFORM(cast(src.key as tinyint), src.value) ROW FORMAT SERDE 'org.apache.hadoop.hive.contrib.serde2.TypedBytesSerDe'
  14. RECORDWRITER 'org.apache.hadoop.hive.contrib.util.typedbytes.TypedBytesRecordWriter'
  15. USING '/bin/cat'
  16. AS (tkey, tvalue) ROW FORMAT SERDE 'org.apache.hadoop.hive.contrib.serde2.TypedBytesSerDe'
  17. RECORDREADER 'org.apache.hadoop.hive.contrib.util.typedbytes.TypedBytesRecordReader'
  18. WHERE key < 100
  19. ) tmap
  20. INSERT OVERWRITE TABLE dest1 SELECT tkey, tvalue ORDER by tkey, tvalue
  21. PREHOOK: type: QUERY
  22. POSTHOOK: query: EXPLAIN
  23. FROM (
  24. FROM src
  25. SELECT TRANSFORM(cast(src.key as tinyint), src.value) ROW FORMAT SERDE 'org.apache.hadoop.hive.contrib.serde2.TypedBytesSerDe'
  26. RECORDWRITER 'org.apache.hadoop.hive.contrib.util.typedbytes.TypedBytesRecordWriter'
  27. USING '/bin/cat'
  28. AS (tkey, tvalue) ROW FORMAT SERDE 'org.apache.hadoop.hive.contrib.serde2.TypedBytesSerDe'
  29. RECORDREADER 'org.apache.hadoop.hive.contrib.util.typedbytes.TypedBytesRecordReader'
  30. WHERE key < 100
  31. ) tmap
  32. INSERT OVERWRITE TABLE dest1 SELECT tkey, tvalue ORDER by tkey, tvalue
  33. POSTHOOK: type: QUERY
  34. ABSTRACT SYNTAX TREE:
  35. (TOK_QUERY (TOK_FROM (TOK_SUBQUERY (TOK_QUERY (TOK_FROM (TOK_TABREF (TOK_TABNAME src))) (TOK_INSERT (TOK_DESTINATION (TOK_DIR TOK_TMP_FILE)) (TOK_SELECT (TOK_SELEXPR (TOK_TRANSFORM (TOK_EXPLIST (TOK_FUNCTION TOK_TINYINT (. (TOK_TABLE_OR_COL src) key)) (. (TOK_TABLE_OR_COL src) value)) (TOK_SERDE (TOK_SERDENAME 'org.apache.hadoop.hive.contrib.serde2.TypedBytesSerDe')) (TOK_RECORDWRITER 'org.apache.hadoop.hive.contrib.util.typedbytes.TypedBytesRecordWriter') '/bin/cat' (TOK_SERDE (TOK_SERDENAME 'org.apache.hadoop.hive.contrib.serde2.TypedBytesSerDe')) (TOK_RECORDREADER 'org.apache.hadoop.hive.contrib.util.typedbytes.TypedBytesRecordReader') (TOK_ALIASLIST tkey tvalue)))) (TOK_WHERE (< (TOK_TABLE_OR_COL key) 100)))) tmap)) (TOK_INSERT (TOK_DESTINATION (TOK_TAB (TOK_TABNAME dest1))) (TOK_SELECT (TOK_SELEXPR (TOK_TABLE_OR_COL tkey)) (TOK_SELEXPR (TOK_TABLE_OR_COL tvalue))) (TOK_ORDERBY (TOK_TABSORTCOLNAMEASC (TOK_TABLE_OR_COL tkey)) (TOK_TABSORTCOLNAMEASC (TOK_TABLE_OR_COL tvalue)))))
  36. STAGE DEPENDENCIES:
  37. Stage-1 is a root stage
  38. Stage-0 depends on stages: Stage-1
  39. Stage-2 depends on stages: Stage-0
  40. STAGE PLANS:
  41. Stage: Stage-1
  42. Map Reduce
  43. Alias -> Map Operator Tree:
  44. tmap:src
  45. TableScan
  46. alias: src
  47. Filter Operator
  48. predicate:
  49. expr: (key < 100)
  50. type: boolean
  51. Filter Operator
  52. predicate:
  53. expr: (key < 100)
  54. type: boolean
  55. Select Operator
  56. expressions:
  57. expr: UDFToByte(key)
  58. type: tinyint
  59. expr: value
  60. type: string
  61. outputColumnNames: _col0, _col1
  62. Transform Operator
  63. command: /bin/cat
  64. output info:
  65. input format: org.apache.hadoop.mapred.TextInputFormat
  66. output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
  67. Select Operator
  68. expressions:
  69. expr: _col0
  70. type: string
  71. expr: _col1
  72. type: string
  73. outputColumnNames: _col0, _col1
  74. Reduce Output Operator
  75. key expressions:
  76. expr: _col0
  77. type: string
  78. expr: _col1
  79. type: string
  80. sort order: ++
  81. tag: -1
  82. value expressions:
  83. expr: _col0
  84. type: string
  85. expr: _col1
  86. type: string
  87. Reduce Operator Tree:
  88. Extract
  89. File Output Operator
  90. compressed: false
  91. GlobalTableId: 1
  92. table:
  93. input format: org.apache.hadoop.mapred.TextInputFormat
  94. output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
  95. serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
  96. name: default.dest1
  97. Stage: Stage-0
  98. Move Operator
  99. tables:
  100. replace: true
  101. table:
  102. input format: org.apache.hadoop.mapred.TextInputFormat
  103. output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat
  104. serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe
  105. name: default.dest1
  106. Stage: Stage-2
  107. Stats-Aggr Operator
  108. PREHOOK: query: FROM (
  109. FROM src
  110. SELECT TRANSFORM(cast(src.key as tinyint), src.value) ROW FORMAT SERDE 'org.apache.hadoop.hive.contrib.serde2.TypedBytesSerDe'
  111. RECORDWRITER 'org.apache.hadoop.hive.contrib.util.typedbytes.TypedBytesRecordWriter'
  112. USING '/bin/cat'
  113. AS (tkey, tvalue) ROW FORMAT SERDE 'org.apache.hadoop.hive.contrib.serde2.TypedBytesSerDe'
  114. RECORDREADER 'org.apache.hadoop.hive.contrib.util.typedbytes.TypedBytesRecordReader'
  115. WHERE key < 100
  116. ) tmap
  117. INSERT OVERWRITE TABLE dest1 SELECT tkey, tvalue ORDER by tkey, tvalue
  118. PREHOOK: type: QUERY
  119. PREHOOK: Input: default@src
  120. PREHOOK: Output: default@dest1
  121. POSTHOOK: query: FROM (
  122. FROM src
  123. SELECT TRANSFORM(cast(src.key as tinyint), src.value) ROW FORMAT SERDE 'org.apache.hadoop.hive.contrib.serde2.TypedBytesSerDe'
  124. RECORDWRITER 'org.apache.hadoop.hive.contrib.util.typedbytes.TypedBytesRecordWriter'
  125. USING '/bin/cat'
  126. AS (tkey, tvalue) ROW FORMAT SERDE 'org.apache.hadoop.hive.contrib.serde2.TypedBytesSerDe'
  127. RECORDREADER 'org.apache.hadoop.hive.contrib.util.typedbytes.TypedBytesRecordReader'
  128. WHERE key < 100
  129. ) tmap
  130. INSERT OVERWRITE TABLE dest1 SELECT tkey, tvalue ORDER by tkey, tvalue
  131. POSTHOOK: type: QUERY
  132. POSTHOOK: Input: default@src
  133. POSTHOOK: Output: default@dest1
  134. POSTHOOK: Lineage: dest1.key SCRIPT [(src)src.FieldSchema(name:key, type:string, comment:default), (src)src.FieldSchema(name:value, type:string, comment:default), ]
  135. POSTHOOK: Lineage: dest1.value SCRIPT [(src)src.FieldSchema(name:key, type:string, comment:default), (src)src.FieldSchema(name:value, type:string, comment:default), ]
  136. PREHOOK: query: SELECT dest1.* FROM dest1
  137. PREHOOK: type: QUERY
  138. PREHOOK: Input: default@dest1
  139. PREHOOK: Output: file:/tmp/sdong/hive_2011-02-16_20-10-19_325_182915218311720960/-mr-10000
  140. POSTHOOK: query: SELECT dest1.* FROM dest1
  141. POSTHOOK: type: QUERY
  142. POSTHOOK: Input: default@dest1
  143. POSTHOOK: Output: file:/tmp/sdong/hive_2011-02-16_20-10-19_325_182915218311720960/-mr-10000
  144. POSTHOOK: Lineage: dest1.key SCRIPT [(src)src.FieldSchema(name:key, type:string, comment:default), (src)src.FieldSchema(name:value, type:string, comment:default), ]
  145. POSTHOOK: Lineage: dest1.value SCRIPT [(src)src.FieldSchema(name:key, type:string, comment:default), (src)src.FieldSchema(name:value, type:string, comment:default), ]
  146. 0 val_0
  147. 0 val_0
  148. 0 val_0
  149. 10 val_10
  150. 11 val_11
  151. 12 val_12
  152. 12 val_12
  153. 15 val_15
  154. 15 val_15
  155. 17 val_17
  156. 18 val_18
  157. 18 val_18
  158. 19 val_19
  159. 2 val_2
  160. 20 val_20
  161. 24 val_24
  162. 24 val_24
  163. 26 val_26
  164. 26 val_26
  165. 27 val_27
  166. 28 val_28
  167. 30 val_30
  168. 33 val_33
  169. 34 val_34
  170. 35 val_35
  171. 35 val_35
  172. 35 val_35
  173. 37 val_37
  174. 37 val_37
  175. 4 val_4
  176. 41 val_41
  177. 42 val_42
  178. 42 val_42
  179. 43 val_43
  180. 44 val_44
  181. 47 val_47
  182. 5 val_5
  183. 5 val_5
  184. 5 val_5
  185. 51 val_51
  186. 51 val_51
  187. 53 val_53
  188. 54 val_54
  189. 57 val_57
  190. 58 val_58
  191. 58 val_58
  192. 64 val_64
  193. 65 val_65
  194. 66 val_66
  195. 67 val_67
  196. 67 val_67
  197. 69 val_69
  198. 70 val_70
  199. 70 val_70
  200. 70 val_70
  201. 72 val_72
  202. 72 val_72
  203. 74 val_74
  204. 76 val_76
  205. 76 val_76
  206. 77 val_77
  207. 78 val_78
  208. 8 val_8
  209. 80 val_80
  210. 82 val_82
  211. 83 val_83
  212. 83 val_83
  213. 84 val_84
  214. 84 val_84
  215. 85 val_85
  216. 86 val_86
  217. 87 val_87
  218. 9 val_9
  219. 90 val_90
  220. 90 val_90
  221. 90 val_90
  222. 92 val_92
  223. 95 val_95
  224. 95 val_95
  225. 96 val_96
  226. 97 val_97
  227. 97 val_97
  228. 98 val_98
  229. 98 val_98
  230. PREHOOK: query: drop table dest1
  231. PREHOOK: type: DROPTABLE
  232. PREHOOK: Input: default@dest1
  233. PREHOOK: Output: default@dest1
  234. POSTHOOK: query: drop table dest1
  235. POSTHOOK: type: DROPTABLE
  236. POSTHOOK: Input: default@dest1
  237. POSTHOOK: Output: default@dest1
  238. POSTHOOK: Lineage: dest1.key SCRIPT [(src)src.FieldSchema(name:key, type:string, comment:default), (src)src.FieldSchema(name:value, type:string, comment:default), ]
  239. POSTHOOK: Lineage: dest1.value SCRIPT [(src)src.FieldSchema(name:key, type:string, comment:default), (src)src.FieldSchema(name:value, type:string, comment:default), ]