hive 之select 中文乱码

此处的中文乱码和mysql的库表 编码 latin utf 无关。

直接上案例。

有时候我们需要自定义一列,有时是汉字有时是字母,结果遇到这种情况了。

说实话看到这真是糟心。这谁受得了。

单独select 没有任何问题。

这是怎么回事呢? 经过一番检查,发现有个地方类似与 "境内" as col但是没乱码,

此时怀疑就是if 函数起了作用,但是一时间不知道是为啥。。

经过多方面测试 concat("境内") concat_ws("","境内")没用,

concat_ws("",arrary("境内")) 有用,此时也不知道如何下手,只有掏出大杀器 explain.

起作用的

Plan optimized by CBO.

""

Vertex dependency in root stage

Map 1 <- Map 3 (BROADCAST_EDGE)

Reducer 2 <- Map 1 (SIMPLE_EDGE)

""

Stage-0

Fetch Operator

limit:-1

Stage-1

Reducer 2

File Output Operator [FS_14]

Select Operator [SEL_13] (rows=105 width=273)

" Output:[""_col0"",""_col1"",""_col2"",""_col3"",""_col4"",""_col5"",""_col6"",""_col7""]"

Group By Operator [GBY_12] (rows=105 width=273)

" Output:[""_col0"",""_col1"",""_col2"",""_col3"",""_col4""],keys:KEY._col0, KEY._col1, KEY._col2, KEY._col3, KEY._col4"

<-Map 1 [SIMPLE_EDGE] vectorized

SHUFFLE [RS_28]

" PartitionCols:_col0, _col1, _col2, _col3, _col4"

Group By Operator [GBY_27] (rows=211 width=273)

" Output:[""_col0"",""_col1"",""_col2"",""_col3"",""_col4""],keys:_col1, _col2, _col3, _col4, _col5"

Map Join Operator [MAPJOIN_26] (rows=211 width=273)

" Conds:SEL_25._col0=RS_23._col0(Inner),Output:[""_col1"",""_col2"",""_col3"",""_col4"",""_col5""]"

<-Map 3 [BROADCAST_EDGE] vectorized

BROADCAST [RS_23]

PartitionCols:_col0

Select Operator [SEL_22] (rows=1 width=736)

" Output:[""_col0"",""_col1"",""_col2"",""_col3""]"

Filter Operator [FIL_21] (rows=1 width=736)

predicate:bank_code is not null

TableScan [TS_3] (rows=1 width=736)

" dwapsdata@dw_conf_ce_bank_dict_v,t1,Tbl:COMPLETE,Col:NONE,Output:[""bank_code"",""bank_name"",""bank_short_name"",""bank_onshore_flag""]"

<-Select Operator [SEL_25] (rows=192 width=273)

" Output:[""_col0"",""_col1""]"

Filter Operator [FIL_24] (rows=192 width=273)

predicate:bank_code is not null

TableScan [TS_0] (rows=192 width=273)

" dwdmdata@dm_ce_f_portrait_credit_line,t,Tbl:COMPLETE,Col:COMPLETE,Output:[""bank_code""]"

""

没有作用的

Plan optimized by CBO.

""

Vertex dependency in root stage

Map 1 <- Map 3 (BROADCAST_EDGE)

Reducer 2 <- Map 1 (SIMPLE_EDGE)

""

Stage-0

Fetch Operator

limit:-1

Stage-1

Reducer 2 vectorized

File Output Operator [FS_31]

Select Operator [SEL_30] (rows=105 width=273)

" Output:[""_col0"",""_col1"",""_col2"",""_col3"",""_col4"",""_col5"",""_col6""]"

Group By Operator [GBY_29] (rows=105 width=273)

" Output:[""_col0"",""_col1"",""_col2"",""_col3"",""_col4""],keys:KEY._col0, KEY._col1, KEY._col2, KEY._col3, KEY._col4"

<-Map 1 [SIMPLE_EDGE] vectorized

SHUFFLE [RS_28]

" PartitionCols:_col0, _col1, _col2, _col3, _col4"

Group By Operator [GBY_27] (rows=211 width=273)

" Output:[""_col0"",""_col1"",""_col2"",""_col3"",""_col4""],keys:_col1, _col2, _col3, _col4, _col5"

Map Join Operator [MAPJOIN_26] (rows=211 width=273)

" Conds:SEL_25._col0=RS_23._col0(Inner),Output:[""_col1"",""_col2"",""_col3"",""_col4"",""_col5""]"

<-Map 3 [BROADCAST_EDGE] vectorized

BROADCAST [RS_23]

PartitionCols:_col0

Select Operator [SEL_22] (rows=1 width=736)

" Output:[""_col0"",""_col1"",""_col2"",""_col3""]"

Filter Operator [FIL_21] (rows=1 width=736)

predicate:bank_code is not null

TableScan [TS_3] (rows=1 width=736)

" dwapsdata@dw_conf_ce_bank_dict_v,t1,Tbl:COMPLETE,Col:NONE,Output:[""bank_code"",""bank_name"",""bank_short_name"",""bank_onshore_flag""]"

<-Select Operator [SEL_25] (rows=192 width=273)

" Output:[""_col0"",""_col1""]"

Filter Operator [FIL_24] (rows=192 width=273)

predicate:bank_code is not null

TableScan [TS_0] (rows=192 width=273)

" dwdmdata@dm_ce_f_portrait_credit_line,t,Tbl:COMPLETE,Col:COMPLETE,Output:[""bank_code""]"

""

对比发现

vectorzied 这个单词一出来我就知道怎么回事了。

hive decimal bug, nvl(decimal,1)=0_cclovezbf的博客-CSDN博客

这个b参数好处没体会到一点,bug到是一堆。

复制代码
set hive.vectorized.execution.enabled=false; 即可解决中文乱码问题!!!!!!!

其实还有别的办法,但是和concat_ws(array(""))一样比较丑陋,我就不说了

相关推荐
写代码的【黑咖啡】2 小时前
HDFS简介及其存储机制详解
大数据·hadoop·hdfs
俊哥大数据4 小时前
【项目实战1】大数据项目开发案例---新闻资讯离线分析|实时分析|大数据仓库|推荐系统|数据可视化项目
数据仓库·hadoop·flink·spark·推荐系统·实时分析·离线分析
忘记9265 小时前
Servlet 生命周期
数据仓库·hive·hadoop
zhixingheyi_tian7 小时前
HDFS 之 Client 调试
大数据·hadoop·hdfs
天天向上杰7 小时前
小识:从理财数仓角度看GaussDB、PostgreSQL、Hive 三区别
hive·hadoop·gaussdb
写代码的【黑咖啡】7 小时前
Hive on Spark:加速大数据分析的新引擎
hive·数据分析·spark
yumgpkpm9 小时前
Hadoop 与AI大模型实战:从Hive、Impala(Cloudera CDH、CDP)海量数据到 AI 决策的落地方法
arm开发·人工智能·hive·zookeeper·flink·kafka·cloudera
路边草随风21 小时前
java操作cosn使用
java·大数据·hadoop
码以致用1 天前
Hive笔记
hive·hadoop·笔记
路边草随风1 天前
通过hive元数据库查询表信息
大数据·数据库·hive·hadoop