메뉴 건너뛰기

Cloudera, BigData, Semantic IoT, Hadoop, NoSQL

Cloudera CDH/CDP 및 Hadoop EcoSystem, Semantic IoT등의 개발/운영 기술을 정리합니다. gooper@gooper.com로 문의 주세요.


./schematool -initSchema -dbType derby 를 실행해준다.

(-dbType은  derby|mysql|postgres|oracle중에 하나를 지정해준다.)


schematool실행시 오류가 발생하여 다시 schematool -initSchema를 실행할때 "Error: FUNCTION 'NUCLEUS_ASCII' already exists. (state=X0Y68,code=30000)"오류가 발생하는데 이때는 "ls -l | grep meta"로 metastore_db폴더를 찾아 삭제하고 schematool을 이용하여 initSchema를 수행하면 정상적으로 schema가 초기화 된다.


* 참조 : https://cwiki.apache.org/confluence/display/Hive/Hive+Schema+Tool


-----------------schema설정없이 hive를 최초에 실행했을때 발생하는 오류메세지 ----------

$ bin/hive

hive: line 86: readlink: command not found

which: 0652-141 There is no hbase in /usr/bin /etc /usr/sbin /usr/ucb /engine/bigdata/bin /usr/bin/X11 /sbin . /usr/java7_64/bin /engine/bigdata/hadoop/bin /engine/bigdata/hive/bin.

SLF4J: Class path contains multiple SLF4J bindings.

SLF4J: Found binding in [jar:file:/engine/bigdata/apache-hive-2.1.0-bin/lib/hive-jdbc-2.1.0-standalone.jar!/org/slf4j/impl/StaticLoggerBinder.class]

SLF4J: Found binding in [jar:file:/engine/bigdata/apache-hive-2.1.0-bin/lib/log4j-slf4j-impl-2.4.1.jar!/org/slf4j/impl/StaticLoggerBinder.class]

SLF4J: Found binding in [jar:file:/engine/bigdata/hadoop-2.7.2/share/hadoop/common/lib/slf4j-log4j12-1.7.10.jar!/org/slf4j/impl/StaticLoggerBinder.class]

SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an explanation.

SLF4J: Actual binding is of type [org.apache.logging.slf4j.Log4jLoggerFactory]


Logging initialized using configuration in file:/engine/bigdata/apache-hive-2.1.0-bin/conf/hive-log4j2.properties Async: true

Exception in thread "main" java.lang.RuntimeException: org.apache.hadoop.hive.ql.metadata.HiveException: org.apache.hadoop.hive.ql.metadata.HiveException: MetaException(message:Hive metastore database is not initialized. Please use schematool (e.g. ./schematool -initSchema -dbType ...) to create the schema. If needed, don'                         t forget to include the option to auto-create the underlying database in your JDBC connection string (e.g. ?createDatabaseIfNotExist=true for mysql))

        at org.apache.hadoop.hive.ql.session.SessionState.start(SessionState.java:578)

        at org.apache.hadoop.hive.ql.session.SessionState.beginStart(SessionState.java:518)

        at org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:705)

        at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:641)

        at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)

        at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:95)

        at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:56)

        at java.lang.reflect.Method.invoke(Method.java:620)

        at org.apache.hadoop.util.RunJar.run(RunJar.java:221)

        at org.apache.hadoop.util.RunJar.main(RunJar.java:136)

Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: org.apache.hadoop.hive.ql.metadata.HiveException: MetaException(message:Hive metastore database is not initialized. Please use schematool (e.g. ./schematool -initSchema -dbType ...) to create the schema. If needed, don't forget to include the option to auto-creat                         e the underlying database in your JDBC connection string (e.g. ?createDatabaseIfNotExist=true for mysql))

        at org.apache.hadoop.hive.ql.metadata.Hive.registerAllFunctionsOnce(Hive.java:226)

        at org.apache.hadoop.hive.ql.metadata.Hive.<init>(Hive.java:366)

        at org.apache.hadoop.hive.ql.metadata.Hive.create(Hive.java:310)

        at org.apache.hadoop.hive.ql.metadata.Hive.getInternal(Hive.java:290)

        at org.apache.hadoop.hive.ql.metadata.Hive.get(Hive.java:266)

        at org.apache.hadoop.hive.ql.session.SessionState.start(SessionState.java:545)

        ... 9 more

Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: MetaException(message:Hive metastore database is not initialized. Please use schematool (e.g. ./schematool -initSchema -dbType ...) to create the schema. If needed, don't forget to include the option to auto-create the underlying database in your JDBC connection                          string (e.g. ?createDatabaseIfNotExist=true for mysql))

        at org.apache.hadoop.hive.ql.metadata.Hive.getAllFunctions(Hive.java:3593)

        at org.apache.hadoop.hive.ql.metadata.Hive.reloadFunctions(Hive.java:236)

        at org.apache.hadoop.hive.ql.metadata.Hive.registerAllFunctionsOnce(Hive.java:221)

        ... 14 more

Caused by: MetaException(message:Hive metastore database is not initialized. Please use schematool (e.g. ./schematool -initSchema -dbType ...) to create the schema. If needed, don't forget to include the option to auto-create the underlying database in your JDBC connection string (e.g. ?createDatabaseIfNotExist=true for my                         sql))

        at org.apache.hadoop.hive.ql.metadata.Hive.getMSC(Hive.java:3364)

        at org.apache.hadoop.hive.ql.metadata.Hive.getMSC(Hive.java:3336)

        at org.apache.hadoop.hive.ql.metadata.Hive.getAllFunctions(Hive.java:3590)

        ... 16 more

번호 제목 날짜 조회 수
85 [impala]쿼리 수행중 발생하는 오류(due to memory pressure: the memory usage of this transaction, Failed to write to server) 2022.10.05 4260
84 [TLS/SSL]Kudu Master 설정하기 2022.05.13 4294
83 AIX 7.1에서 hive실행시 "hive: line 86: readlink: command not found" 오류가 발생시 임시 조치사항 2016.09.25 4303
82 Scala를 이용한 Streaming예제 2018.03.08 4303
81 Soft memory limit exceeded (at 100.05% of capacity) 오류 조치 2022.01.17 4324
80 conda를 이용한 jupyterhub(v0.9)및 jupyter설치 (v4.4.0) 2018.07.30 4327
79 [sentry]role부여후 테이블명이 변경되어 오류가 발생할때 조치방법 2018.10.16 4336
78 CentOS 7.x에 Jupyter설치 2018.04.18 4337
77 AnalysisException: Incomplatible return type 'DECIMAL(38,0)' and 'DECIMAL(38,5)' of exprs가 발생시 조치 2021.07.26 4341
76 Apache Spark와 Drools를 이용한 CEP구현 테스트 2016.07.15 4362
75 spark-env.sh에서 사용할 수있는 항목. 2016.05.24 4385
74 kudu hms check 사용법(예시) 2021.10.22 4403
73 [Kudu]Schema별 혹은 테이블별 사용량(Replica포함) 구하는 방법 2022.07.14 4407
72 spark-submit으로 spark application실행하는 다양한 방법 2016.05.25 4414
71 Windows7 64bit 환경에서 Apache Spark 2.2.0 설치하기 2017.07.26 4415
70 [CDP7.1.7]Impala Query의 Memory Spilled 양은 ScratchFileUsedBytes값을 누적해서 구할 수 있다. 2022.07.29 4424
69 hive query에서 mapreduce돌리지 않고 select하는 방법 2014.05.23 4435
68 Failed to write to server: (no server available): 2022.01.17 4435
67 Query Status: Sender xxx.xxx.xxx.xxx timed out waiting for receiver fragment instance: 1234:cdsf, dest node: 10 의 오류 원인및 대응방안 2021.11.03 4479
66 index생성, 삭제, 활용 2014.04.25 4490
위로