메뉴 건너뛰기

Bigdata, Semantic IoT, Hadoop, NoSQL

Bigdata, Hadoop ecosystem, Semantic IoT등의 프로젝트를 진행중에 습득한 내용을 정리하는 곳입니다.
필요한 분을 위해서 공개하고 있습니다. 문의사항은 gooper@gooper.com로 메일을 보내주세요.


Hive가 "GROUP BY"에서 최적화하면서 내부적으로 컬럼을 추가하여 reducer로 보내는데 이 컬럼은 실존하지 않는 컬럼으로 문제가 발생할 수 있음
hive 0.14.0에서 default가 disabled되었는데 이전 버전에서는 default가 enabled로 되어 있어서 문제가 발생하면
set hive.optimize.sort.dynamic.partition=false;
를 추가하여 준다.

-----오류내용----
2015-06-15 14:12:01,345 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskImpl: task_1434344795088_0005_r_000000 Task Transitioned from SCHEDULED to RUNNING
2015-06-15 14:12:02,327 INFO [RMCommunicator Allocator] org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor: getResources() for application_1434344795088_0005: ask=1 release= 0 newContainers=0 finishedContainers=0 resourcelimit=<memory:0, vCores:0> knownNMs=5
2015-06-15 14:12:03,067 INFO [Socket Reader #1 for port 47209] SecurityLogger.org.apache.hadoop.ipc.Server: Auth successful for job_1434344795088_0005 (auth:SIMPLE)
2015-06-15 14:12:03,079 INFO [IPC Server handler 2 on 47209] org.apache.hadoop.mapred.TaskAttemptListenerImpl: JVM with ID : jvm_1434344795088_0005_r_000003 asked for a task
2015-06-15 14:12:03,079 INFO [IPC Server handler 2 on 47209] org.apache.hadoop.mapred.TaskAttemptListenerImpl: JVM with ID: jvm_1434344795088_0005_r_000003 given task: attempt_1434344795088_0005_r_000000_0
2015-06-15 14:12:04,185 INFO [IPC Server handler 1 on 47209] org.apache.hadoop.mapred.TaskAttemptListenerImpl: MapCompletionEvents request from attempt_1434344795088_0005_r_000000_0. startIndex 0 maxEvents 10000
2015-06-15 14:12:04,363 INFO [IPC Server handler 2 on 47209] org.apache.hadoop.mapred.TaskAttemptListenerImpl: Progress of TaskAttempt attempt_1434344795088_0005_r_000000_0 is : 0.0
2015-06-15 14:12:04,382 INFO [IPC Server handler 6 on 47209] org.apache.hadoop.mapred.TaskAttemptListenerImpl: Progress of TaskAttempt attempt_1434344795088_0005_r_000000_0 is : 0.0
2015-06-15 14:12:04,769 INFO [IPC Server handler 7 on 47209] org.apache.hadoop.mapred.TaskAttemptListenerImpl: Progress of TaskAttempt attempt_1434344795088_0005_r_000000_0 is : 0.0
2015-06-15 14:12:04,772 FATAL [IPC Server handler 9 on 47209] org.apache.hadoop.mapred.TaskAttemptListenerImpl: Task: attempt_1434344795088_0005_r_000000_0 - exited : java.lang.RuntimeException: org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime Error: Unable to deserialize reduce input key from x1x50x48x49x53x45x48x54x45x49x53x32x49x52x58x49x49x58x52x53x0x1x50x48x49x53x45x48x54x45x48x51x0x1x50x48x49x53x45x48x54x45x48x51x32x49x53x58x49x57x58x48x49x46x0x1x50x50x48x46x57x48x46x50x49x54x46x55x51x0x1x104x116x116x112x58x47x47x50x50x48x46x57x48x46x50x0x1x71x69x84x0x1x187x154x203x138x255 with properties {columns=reducesinkkey0,reducesinkkey1,reducesinkkey2,reducesinkkey3,reducesinkkey4,reducesinkkey5,reducesinkkey6, serialization.lib=org.apache.hadoop.hive.serde2.binarysortable.BinarySortableSerDe, serialization.sort.order=+++++++, columns.types=string,string,string,string,string,string,string}
	at org.apache.hadoop.hive.ql.exec.mr.ExecReducer.reduce(ExecReducer.java:283)
	at org.apache.hadoop.mapred.ReduceTask.runOldReducer(ReduceTask.java:444)
	at org.apache.hadoop.mapred.ReduceTask.run(ReduceTask.java:392)
	at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:168)
	at java.security.AccessController.doPrivileged(Native Method)
	at javax.security.auth.Subject.doAs(Subject.java:415)
	at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1614)
	at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:163)
Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime Error: Unable to deserialize reduce input key from x1x50x48x49x53x45x48x54x45x49x53x32x49x52x58x49x49x58x52x53x0x1x50x48x49x53x45x48x54x45x48x51x0x1x50x48x49x53x45x48x54x45x48x51x32x49x53x58x49x57x58x48x49x46x0x1x50x50x48x46x57x48x46x50x49x54x46x55x51x0x1x104x116x116x112x58x47x47x50x50x48x46x57x48x46x50x0x1x71x69x84x0x1x187x154x203x138x255 with properties {columns=reducesinkkey0,reducesinkkey1,reducesinkkey2,reducesinkkey3,reducesinkkey4,reducesinkkey5,reducesinkkey6, serialization.lib=org.apache.hadoop.hive.serde2.binarysortable.BinarySortableSerDe, serialization.sort.order=+++++++, columns.types=string,string,string,string,string,string,string}
	at org.apache.hadoop.hive.ql.exec.mr.ExecReducer.reduce(ExecReducer.java:222)
	... 7 more
Caused by: org.apache.hadoop.hive.serde2.SerDeException: java.io.EOFException
	at org.apache.hadoop.hive.serde2.binarysortable.BinarySortableSerDe.deserialize(BinarySortableSerDe.java:189)
	at org.apache.hadoop.hive.ql.exec.mr.ExecReducer.reduce(ExecReducer.java:220)
	... 7 more
Caused by: java.io.EOFException
	at org.apache.hadoop.hive.serde2.binarysortable.InputByteBuffer.read(InputByteBuffer.java:54)
	at org.apache.hadoop.hive.serde2.binarysortable.BinarySortableSerDe.deserializeText(BinarySortableSerDe.java:550)
	at org.apache.hadoop.hive.serde2.binarysortable.BinarySortableSerDe.deserialize(BinarySortableSerDe.java:285)
	at org.apache.hadoop.hive.serde2.binarysortable.BinarySortableSerDe.deserialize(BinarySortableSerDe.java:185)
	... 8 more

2015-06-15 14:12:04,772 INFO [IPC Server handler 9 on 47209] org.apache.hadoop.mapred.TaskAttemptListenerImpl: Diagnostics report from attempt_1434344795088_0005_r_000000_0: Error: java.lang.RuntimeException: org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime Error: Unable to deserialize reduce input key from x1x50x48x49x53x45x48x54x45x49x53x32x49x52x58x49x49x58x52x53x0x1x50x48x49x53x45x48x54x45x48x51x0x1x50x48x49x53x45x48x54x45x48x51x32x49x53x58x49x57x58x48x49x46x0x1x50x50x48x46x57x48x46x50x49x54x46x55x51x0x1x104x116x116x112x58x47x47x50x50x48x46x57x48x46x50x0x1x71x69x84x0x1x187x154x203x138x255 with properties {columns=reducesinkkey0,reducesinkkey1,reducesinkkey2,reducesinkkey3,reducesinkkey4,reducesinkkey5,reducesinkkey6, serialization.lib=org.apache.hadoop.hive.serde2.binarysortable.BinarySortableSerDe, serialization.sort.order=+++++++, columns.types=string,string,string,string,string,string,string}
	at org.apache.hadoop.hive.ql.exec.mr.ExecReducer.reduce(ExecReducer.java:283)
	at org.apache.hadoop.mapred.ReduceTask.runOldReducer(ReduceTask.java:444)
	at org.apache.hadoop.mapred.ReduceTask.run(ReduceTask.java:392)
	at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:168)
	at java.security.AccessController.doPrivileged(Native Method)
	at javax.security.auth.Subject.doAs(Subject.java:415)
	at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1614)
	at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:163)
Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime Error: Unable to deserialize reduce input key from x1x50x48x49x53x45x48x54x45x49x53x32x49x52x58x49x49x58x52x53x0x1x50x48x49x53x45x48x54x45x48x51x0x1x50x48x49x53x45x48x54x45x48x51x32x49x53x58x49x57x58x48x49x46x0x1x50x50x48x46x57x48x46x50x49x54x46x55x51x0x1x104x116x116x112x58x47x47x50x50x48x46x57x48x46x50x0x1x71x69x84x0x1x187x154x203x138x255 with properties {columns=reducesinkkey0,reducesinkkey1,reducesinkkey2,reducesinkkey3,reducesinkkey4,reducesinkkey5,reducesinkkey6, serialization.lib=org.apache.hadoop.hive.serde2.binarysortable.BinarySortableSerDe, serialization.sort.order=+++++++, columns.types=string,string,string,string,string,string,string}
	at org.apache.hadoop.hive.ql.exec.mr.ExecReducer.reduce(ExecReducer.java:222)
	... 7 more
Caused by: org.apache.hadoop.hive.serde2.SerDeException: java.io.EOFException
	at org.apache.hadoop.hive.serde2.binarysortable.BinarySortableSerDe.deserialize(BinarySortableSerDe.java:189)
	at org.apache.hadoop.hive.ql.exec.mr.ExecReducer.reduce(ExecReducer.java:220)
	... 7 more
Caused by: java.io.EOFException
	at org.apache.hadoop.hive.serde2.binarysortable.InputByteBuffer.read(InputByteBuffer.java:54)
	at org.apache.hadoop.hive.serde2.binarysortable.BinarySortableSerDe.deserializeText(BinarySortableSerDe.java:550)
	at org.apache.hadoop.hive.serde2.binarysortable.BinarySortableSerDe.deserialize(BinarySortableSerDe.java:285)
	at org.apache.hadoop.hive.serde2.binarysortable.BinarySortableSerDe.deserialize(BinarySortableSerDe.java:185)
	... 8 more

2015-06-15 14:12:04,773 INFO [AsyncDispatcher event handler] org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl: Diagnostics report from attempt_1434344795088_0005_r_000000_0: Error: java.lang.RuntimeException: org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime Error: Unable to deserialize reduce input key from x1x50x48x49x53x45x48x54x45x49x53x32x49x52x58x49x49x58x52x53x0x1x50x48x49x53x45x48x54x45x48x51x0x1x50x48x49x53x45x48x54x45x48x51x32x49x53x58x49x57x58x48x49x46x0x1x50x50x48x46x57x48x46x50x49x54x46x55x51x0x1x104x116x116x112x58x47x47x50x50x48x46x57x48x46x50x0x1x71x69x84x0x1x187x154x203x138x255 with properties {columns=reducesinkkey0,reducesinkkey1,reducesinkkey2,reducesinkkey3,reducesinkkey4,reducesinkkey5,reducesinkkey6, serialization.lib=org.apache.hadoop.hive.serde2.binarysortable.BinarySortableSerDe, serialization.sort.order=+++++++, columns.types=string,string,string,string,string,string,string}
	at org.apache.hadoop.hive.ql.exec.mr.ExecReducer.reduce(ExecReducer.java:283)
	at org.apache.hadoop.mapred.ReduceTask.runOldReducer(ReduceTask.java:444)
	at org.apache.hadoop.mapred.ReduceTask.run(ReduceTask.java:392)
	at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:168)
	at java.security.AccessController.doPrivileged(Native Method)
	at javax.security.auth.Subject.doAs(Subject.java:415)
	at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1614)
	at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:163)
Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime Error: Unable to deserialize reduce input key from x1x50x48x49x53x45x48x54x45x49x53x32x49x52x58x49x49x58x52x53x0x1x50x48x49x53x45x48x54x45x48x51x0x1x50x48x49x53x45x48x54x45x48x51x32x49x53x58x49x57x58x48x49x46x0x1x50x50x48x46x57x48x46x50x49x54x46x55x51x0x1x104x116x116x112x58x47x47x50x50x48x46x57x48x46x50x0x1x71x69x84x0x1x187x154x203x138x255 with properties {columns=reducesinkkey0,reducesinkkey1,reducesinkkey2,reducesinkkey3,reducesinkkey4,reducesinkkey5,reducesinkkey6, serialization.lib=org.apache.hadoop.hive.serde2.binarysortable.BinarySortableSerDe, serialization.sort.order=+++++++, columns.types=string,string,string,string,string,string,string}
	at org.apache.hadoop.hive.ql.exec.mr.ExecReducer.reduce(ExecReducer.java:222)
	... 7 more
Caused by: org.apache.hadoop.hive.serde2.SerDeException: java.io.EOFException
	at org.apache.hadoop.hive.serde2.binarysortable.BinarySortableSerDe.deserialize(BinarySortableSerDe.java:189)
	at org.apache.hadoop.hive.ql.exec.mr.ExecReducer.reduce(ExecReducer.java:220)
	... 7 more
Caused by: java.io.EOFException
	at org.apache.hadoop.hive.serde2.binarysortable.InputByteBuffer.read(InputByteBuffer.java:54)
	at org.apache.hadoop.hive.serde2.binarysortable.BinarySortableSerDe.deserializeText(BinarySortableSerDe.java:550)
	at org.apache.hadoop.hive.serde2.binarysortable.BinarySortableSerDe.deserialize(BinarySortableSerDe.java:285)
	at org.apache.hadoop.hive.serde2.binarysortable.BinarySortableSerDe.deserialize(BinarySortableSerDe.java:185)
	... 8 more
번호 제목 글쓴이 날짜 조회 수
130 [Impala] alter table구문수행시 "WARNINGS: Impala does not have READ_WRITE access to path 'hdfs://nameservice1/DATA/Temp/DB/source/table01_ccd'" 발생시 조치 gooper 2024.04.26 0
129 [CDP7.1.7, Hive Replication]Hive Replication진행중 "The following columns have types incompatible with the existing columns in their respective positions " 오류 gooper 2023.12.27 7
128 [CDP7.1.7]Oozie job에서 ERROR: Kudu error(s) reported, first error: Timed out: Failed to write batch of 774 ops to tablet 8003f9a064bf4be5890a178439b2ba91가 발생하면서 쿼리가 실패하는 경우 gooper 2024.01.05 7
127 [CDP7.1.7]impala-shell수행시 간헐적으로 "-k requires a valid kerberos ticket but no valid kerberos ticket found." 오류 gooper 2023.11.16 11
126 임시 테이블에서 데이터를 읽어서 partitioned table에 입력하는 impala SQL문 예시 gooper 2023.11.10 16
125 [impala]insert into db명.table명 select a, b from db명.table명 쿼리 수행시 "Memory limit exceeded: Failed to allocate memory for Parquet page index"오류 조치 방법 gooper 2023.05.31 22
124 not leader of this config: current role FOLLOWER 오류 발생시 확인방법 총관리자 2022.01.17 23
123 kudu table와 impala(hive) table정보가 틀어져서 테이블을 읽지 못하는 경우(Error Loading Metadata) 조치방법 gooper 2023.11.10 25
122 [CDP7.1.7]Impala Query의 Memory Spilled 양은 ScratchFileUsedBytes값을 누적해서 구할 수 있다. gooper 2022.07.29 29
121 [Cloudera 6.3.4, Kudu]]Service Monitor에서 사용하는 metric중에 일부를 blacklist로 설정하여 모니터링 정보 수집 제외하는 방법 gooper 2022.07.08 31
120 Failed to write to server: (no server available): 총관리자 2022.01.17 32
119 AnalysisException: Incomplatible return type 'DECIMAL(38,0)' and 'DECIMAL(38,5)' of exprs가 발생시 조치 총관리자 2021.07.26 34
118 spark에서 hive table을 읽어 출력하는 예제 소스 총관리자 2017.03.09 35
117 [TLS/SSL]Kudu Tablet Server설정 총관리자 2022.05.13 35
116 spark에서 hive table을 읽어 출력하는 예제 소스 총관리자 2017.03.09 37
115 [KUDU] kudu tablet server여러가지 원인에 의해서 corrupted상태가 된 경우 복구방법 gooper 2023.03.28 37
114 [CDP7.1.7]impala-shell을 이용하여 kudu table에 insert/update수행시 발생하는 오류(Transport endpoint is not connected (error 107)) 발생시 확인할 내용 gooper 2023.11.30 45
113 spark 온라인 책자링크 (제목 : mastering-apache-spark) 총관리자 2016.05.25 51
112 [hive] hive.tbls테이블의 owner컬럼값은 hadoop.security.auth_to_local에 의해서 filtering된다. 총관리자 2022.04.14 55
111 [Impala jdbc]CDP7.1.7환경에서 java프로그램을 이용하여 kerberized impala cluster에 접근하여 SQL을 수행하는 방법 gooper 2023.08.22 57

A personal place to organize information learned during the development of such Hadoop, Hive, Hbase, Semantic IoT, etc.
We are open to the required minutes. Please send inquiries to gooper@gooper.com.

위로