메뉴 건너뛰기

Bigdata, Semantic IoT, Hadoop, NoSQL

Bigdata, Hadoop ecosystem, Semantic IoT등의 프로젝트를 진행중에 습득한 내용을 정리하는 곳입니다.
필요한 분을 위해서 공개하고 있습니다. 문의사항은 gooper@gooper.com로 메일을 보내주세요.


./schematool -initSchema -dbType derby 를 실행해준다.

(-dbType은  derby|mysql|postgres|oracle중에 하나를 지정해준다.)


schematool실행시 오류가 발생하여 다시 schematool -initSchema를 실행할때 "Error: FUNCTION 'NUCLEUS_ASCII' already exists. (state=X0Y68,code=30000)"오류가 발생하는데 이때는 "ls -l | grep meta"로 metastore_db폴더를 찾아 삭제하고 schematool을 이용하여 initSchema를 수행하면 정상적으로 schema가 초기화 된다.


* 참조 : https://cwiki.apache.org/confluence/display/Hive/Hive+Schema+Tool


-----------------schema설정없이 hive를 최초에 실행했을때 발생하는 오류메세지 ----------

$ bin/hive

hive: line 86: readlink: command not found

which: 0652-141 There is no hbase in /usr/bin /etc /usr/sbin /usr/ucb /engine/bigdata/bin /usr/bin/X11 /sbin . /usr/java7_64/bin /engine/bigdata/hadoop/bin /engine/bigdata/hive/bin.

SLF4J: Class path contains multiple SLF4J bindings.

SLF4J: Found binding in [jar:file:/engine/bigdata/apache-hive-2.1.0-bin/lib/hive-jdbc-2.1.0-standalone.jar!/org/slf4j/impl/StaticLoggerBinder.class]

SLF4J: Found binding in [jar:file:/engine/bigdata/apache-hive-2.1.0-bin/lib/log4j-slf4j-impl-2.4.1.jar!/org/slf4j/impl/StaticLoggerBinder.class]

SLF4J: Found binding in [jar:file:/engine/bigdata/hadoop-2.7.2/share/hadoop/common/lib/slf4j-log4j12-1.7.10.jar!/org/slf4j/impl/StaticLoggerBinder.class]

SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an explanation.

SLF4J: Actual binding is of type [org.apache.logging.slf4j.Log4jLoggerFactory]


Logging initialized using configuration in file:/engine/bigdata/apache-hive-2.1.0-bin/conf/hive-log4j2.properties Async: true

Exception in thread "main" java.lang.RuntimeException: org.apache.hadoop.hive.ql.metadata.HiveException: org.apache.hadoop.hive.ql.metadata.HiveException: MetaException(message:Hive metastore database is not initialized. Please use schematool (e.g. ./schematool -initSchema -dbType ...) to create the schema. If needed, don'                         t forget to include the option to auto-create the underlying database in your JDBC connection string (e.g. ?createDatabaseIfNotExist=true for mysql))

        at org.apache.hadoop.hive.ql.session.SessionState.start(SessionState.java:578)

        at org.apache.hadoop.hive.ql.session.SessionState.beginStart(SessionState.java:518)

        at org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:705)

        at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:641)

        at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)

        at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:95)

        at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:56)

        at java.lang.reflect.Method.invoke(Method.java:620)

        at org.apache.hadoop.util.RunJar.run(RunJar.java:221)

        at org.apache.hadoop.util.RunJar.main(RunJar.java:136)

Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: org.apache.hadoop.hive.ql.metadata.HiveException: MetaException(message:Hive metastore database is not initialized. Please use schematool (e.g. ./schematool -initSchema -dbType ...) to create the schema. If needed, don't forget to include the option to auto-creat                         e the underlying database in your JDBC connection string (e.g. ?createDatabaseIfNotExist=true for mysql))

        at org.apache.hadoop.hive.ql.metadata.Hive.registerAllFunctionsOnce(Hive.java:226)

        at org.apache.hadoop.hive.ql.metadata.Hive.<init>(Hive.java:366)

        at org.apache.hadoop.hive.ql.metadata.Hive.create(Hive.java:310)

        at org.apache.hadoop.hive.ql.metadata.Hive.getInternal(Hive.java:290)

        at org.apache.hadoop.hive.ql.metadata.Hive.get(Hive.java:266)

        at org.apache.hadoop.hive.ql.session.SessionState.start(SessionState.java:545)

        ... 9 more

Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: MetaException(message:Hive metastore database is not initialized. Please use schematool (e.g. ./schematool -initSchema -dbType ...) to create the schema. If needed, don't forget to include the option to auto-create the underlying database in your JDBC connection                          string (e.g. ?createDatabaseIfNotExist=true for mysql))

        at org.apache.hadoop.hive.ql.metadata.Hive.getAllFunctions(Hive.java:3593)

        at org.apache.hadoop.hive.ql.metadata.Hive.reloadFunctions(Hive.java:236)

        at org.apache.hadoop.hive.ql.metadata.Hive.registerAllFunctionsOnce(Hive.java:221)

        ... 14 more

Caused by: MetaException(message:Hive metastore database is not initialized. Please use schematool (e.g. ./schematool -initSchema -dbType ...) to create the schema. If needed, don't forget to include the option to auto-create the underlying database in your JDBC connection string (e.g. ?createDatabaseIfNotExist=true for my                         sql))

        at org.apache.hadoop.hive.ql.metadata.Hive.getMSC(Hive.java:3364)

        at org.apache.hadoop.hive.ql.metadata.Hive.getMSC(Hive.java:3336)

        at org.apache.hadoop.hive.ql.metadata.Hive.getAllFunctions(Hive.java:3590)

        ... 16 more

번호 제목 글쓴이 날짜 조회 수
130 [Impala] alter table구문수행시 "WARNINGS: Impala does not have READ_WRITE access to path 'hdfs://nameservice1/DATA/Temp/DB/source/table01_ccd'" 발생시 조치 gooper 2024.04.26 0
129 [CDP7.1.7, Hive Replication]Hive Replication진행중 "The following columns have types incompatible with the existing columns in their respective positions " 오류 gooper 2023.12.27 7
128 [CDP7.1.7]Oozie job에서 ERROR: Kudu error(s) reported, first error: Timed out: Failed to write batch of 774 ops to tablet 8003f9a064bf4be5890a178439b2ba91가 발생하면서 쿼리가 실패하는 경우 gooper 2024.01.05 7
127 [CDP7.1.7]impala-shell수행시 간헐적으로 "-k requires a valid kerberos ticket but no valid kerberos ticket found." 오류 gooper 2023.11.16 11
126 임시 테이블에서 데이터를 읽어서 partitioned table에 입력하는 impala SQL문 예시 gooper 2023.11.10 16
125 [impala]insert into db명.table명 select a, b from db명.table명 쿼리 수행시 "Memory limit exceeded: Failed to allocate memory for Parquet page index"오류 조치 방법 gooper 2023.05.31 22
124 not leader of this config: current role FOLLOWER 오류 발생시 확인방법 총관리자 2022.01.17 23
123 kudu table와 impala(hive) table정보가 틀어져서 테이블을 읽지 못하는 경우(Error Loading Metadata) 조치방법 gooper 2023.11.10 25
122 [CDP7.1.7]Impala Query의 Memory Spilled 양은 ScratchFileUsedBytes값을 누적해서 구할 수 있다. gooper 2022.07.29 29
121 [Cloudera 6.3.4, Kudu]]Service Monitor에서 사용하는 metric중에 일부를 blacklist로 설정하여 모니터링 정보 수집 제외하는 방법 gooper 2022.07.08 31
120 Failed to write to server: (no server available): 총관리자 2022.01.17 32
119 AnalysisException: Incomplatible return type 'DECIMAL(38,0)' and 'DECIMAL(38,5)' of exprs가 발생시 조치 총관리자 2021.07.26 34
118 spark에서 hive table을 읽어 출력하는 예제 소스 총관리자 2017.03.09 35
117 [TLS/SSL]Kudu Tablet Server설정 총관리자 2022.05.13 35
116 spark에서 hive table을 읽어 출력하는 예제 소스 총관리자 2017.03.09 37
115 [KUDU] kudu tablet server여러가지 원인에 의해서 corrupted상태가 된 경우 복구방법 gooper 2023.03.28 37
114 [CDP7.1.7]impala-shell을 이용하여 kudu table에 insert/update수행시 발생하는 오류(Transport endpoint is not connected (error 107)) 발생시 확인할 내용 gooper 2023.11.30 45
113 spark 온라인 책자링크 (제목 : mastering-apache-spark) 총관리자 2016.05.25 51
112 [hive] hive.tbls테이블의 owner컬럼값은 hadoop.security.auth_to_local에 의해서 filtering된다. 총관리자 2022.04.14 55
111 [Impala jdbc]CDP7.1.7환경에서 java프로그램을 이용하여 kerberized impala cluster에 접근하여 SQL을 수행하는 방법 gooper 2023.08.22 58

A personal place to organize information learned during the development of such Hadoop, Hive, Hbase, Semantic IoT, etc.
We are open to the required minutes. Please send inquiries to gooper@gooper.com.

위로