메뉴 건너뛰기

Bigdata, Semantic IoT, Hadoop, NoSQL

Bigdata, Hadoop ecosystem, Semantic IoT등의 프로젝트를 진행중에 습득한 내용을 정리하는 곳입니다.
필요한 분을 위해서 공개하고 있습니다. 문의사항은 gooper@gooper.com로 메일을 보내주세요.


*출처1 : https://www.cloudera.com/more/training/certification/cca-spark.html




*출처2 : http://www.hadoopexam.com/Cloudera_Certification/CCA175/CCA175_Cloudera_Hadoop_and_Spark_Developer_Tips_and_Tricks.pdf

1. Preparation: I have gone through all the CCA175 Questions and practice the code provided by
http://www.HadoopExam.com Thanks for your questions and code content. The content was
excellent and it helped me a lot. (Especially I have gone through all the Spark Professional
training module as well)
2. No. Of Questions: Generally you will get 10 questions in real exam: Topic will be coverings are
Sqoop, Hive, Pyspark and Scala and avro-tools to extract schema (All questions are covered in
CCA175 Certification Simulator).
3. Code Snippets: will be provided for Pyspark and Scala. You have to edit the snippets accordingly
as per the problem statement.
4. Real Exam Environment: Gateway node will be accessible for execution of the problems during
the exam. Keep in mind there will not be any on-screen timer available during the exam. You
have to keep asking for the time left. There are three sections for each problem i.e.
· Instructions
· Data Set
· Output Requirements.
Please go through all the three sections carefully before start developing the code.
Note: If you started developing code right after looking at the Instruction part of the question,
then later you will realized the exact details of the table like name of the table and HDFS
directory are also mentioned. This can waste your time if have to redo the code or might as well
cost you a question.
5. Editor: nano, gedit are not available. So if you have to edit any code snippets, you have to use vi
alone. Please make yourself familiar with vi editor if you are not.
6. Fill in blanks: You dont have to write entire code for Python and Scala for Apache Spark,
generally they will ask you to do fill in the blanks.
7. Flume: Very few questions on flume.
8. Difficulty Level: If you have enough knowledge, you will feel exam is quite easy. The questions
were logically easy and can be answered in the first attempt if you read the question carefully
(all three sections).
9. Common mistake in Sqoop: People use connector as localhost which is wrong, you have to use
full name instead of localhost (Avoid wasting your time). Use given hostname
10. Hive: Have initial knowledge of hive as well.
11. Spark: Using basic transform functions to get desired output. For instance filter according
particular scenario, sorting and ranking etc.
12. Avro-tool : avro-tool to get schema of avro file. (Very  nicely covered in CCA175
HadoopExam.com Simulator)
13. Big Mistake: Avoid accidently deleting your data: good practice is necessary to avoid such
mistakes. (Once you delete or drop hive table, you have to create it entirely once again.) Same is
instructed by www.HadoopExam.com during their videos  session provided at
http://cca175cloudera.training4exam.com/ (Please go through sample sessions)
14. Spark-sql: They will not ask questions based on Spark Sql learn importantly aggregate, reduce,
sort.
15. Time management: It is very important, (That’s the reason you need too much practice, use
CCA175 simulator to practice all the questions at least a week or two before your real exam).
16. Data sets in real exam is quite larger, hence it will take 2 to 5 mins for execution.

17. Attempts: try to attempt all questions at least 9/10, hence you must be able to score 70%.
18. File format: In most of questions there was tab delimited file to process.
19. Python or Scala: You will get a preloaded python or scala file to work with, so you don't have a
choice whether you want to attempt a question via scala or pyspark. (I have gone through all the
Video sessions provided by www.HadoopExam.com here
20. Connection Issue: If you got disconnected during exam, you may need to contact the proctor
immediately. If he/she is not available log back into examslocal.com and use their online help.
21. Shell scripts: Have good experience to use shell scripts.
22. Question types as mentioned in syllabus : Questions were from Sqoop(import and export),
Hive(table creation and dynamic partitioning), Pyspark and Scala(Joining, sorting and filtering
data), avro-tools. Snippets of code will be provided for Pyspark and Scala. You have to edit the
snippets accordingly as per the problem statement and can the script file(which is another file
apart from snippet) to get the results.
23. Overall exam is easy, but require lot of practice to complete on time and for accurate
solutions of the problem. Hence go through the all below material for CCA175 (It will not take
more than a month, if you are new and already know the Spark and Hadoop then 2-3 weeks
are good enough.
· CCA175 : Hadoop and Spark Developer Certification practice questions
· Hadoop professional training
· Spark professional training.

Wish you all the best

번호 제목 글쓴이 날짜 조회 수
580 oozie 4.1 설치 - maven을 이용한 source compile on hadoop 2.5.2 with postgresql 9.3 총관리자 2015.04.30 862
579 ubuntu에 maven 3.6.1설치 및 환경변수 설정 총관리자 2019.06.02 856
578 Impala의 Queries탭에서 여러조건으로 쿼리 찾기 총관리자 2018.05.09 856
577 [SBT] SBT 사용법 정리(링크) 총관리자 2016.08.04 839
576 Hive JDBC Connection과 유형별 에러및 필요한 jar파일 총관리자 2021.05.24 827
575 dual table만들기 총관리자 2014.05.16 824
574 [shellscript]엑셀파일에서 여러줄에 존재하는 단어를 한줄의 문자열로 합치는 방법(comma로 구분) 총관리자 2019.07.15 811
573 update 샘플 총관리자 2018.03.12 810
572 hive query에서 mapreduce돌리지 않고 select하는 방법 총관리자 2014.05.23 810
571 [Mybatis]Spring과 연동하지 않고 Java+Mybatis 형태의 프로그램 샘플소스 총관리자 2016.09.01 808
570 oozie job 구동시 JA009: User: hadoop is not allowed to impersonate hadoop 오류나는 경우 총관리자 2014.06.02 807
569 sentry설정 방법및 활성화시 설정이 필요한 파일및 설정값, 계정생성 방법 총관리자 2018.08.16 777
568 mongodb aggregation query를 Java code로 변환한 샘플 총관리자 2016.12.15 777
567 python test.py실행시 "ImportError: No module named pyspark" 혹은 "ImportError: No module named py4j.protocol"등의 오류 발생시 조치사항 총관리자 2017.07.04 765
566 oracle to hive data type정리표 총관리자 2018.08.22 764
565 [Kerberos]Kerberos상태의 클러스터에 JDBC로 접근할때 케이스별 오류내용 총관리자 2020.02.14 761
564 beeline실행시 User: root is not allowed to impersonate오류 발생시 조치사항 총관리자 2016.06.03 759
563 [DBeaver 4.3.0]import/export시 "Client home is not specified for connection" 오류발생시 조치사항 총관리자 2017.12.21 753
562 Cloudera Manager web UI의 언어를 한글에서 영문으로 변경하기 총관리자 2018.04.03 743
561 hive metastore ERD file 총관리자 2018.09.20 729

A personal place to organize information learned during the development of such Hadoop, Hive, Hbase, Semantic IoT, etc.
We are open to the required minutes. Please send inquiries to gooper@gooper.com.

위로