40000+ Learners upgraded/switched career Testimonials
All
Certifications
preparation material is for renowned vendors like
Cloudera, MapR, EMC, Databricks,SAS, Datastax, Oracle,
NetApp etc , which has more value, reliability and
consideration in industry other than any training
institutional certifications.
Note
:
You can choose more than one product to have custome
package created from below and send email to
hadoopexam@gmail.com to get discount.
Do you know?
Hadoop Admin: Apache Ambari interview Questions which include the 118 questions in total and it will prepare you for the Hadoop Administration. It is not necessary this all questions would be asked during the interview process. But HadoopExam tries to cover all possible concepts which needs to learn for knowing the Apache Ambari Hadoop Cluster management tool. These questions and answer would be helpful to understand the various components, operations, monitoring and administering the Hadoop cluster for sure. The benefit of Question and answer format is that, it would allow you to understand the thing in depth and you can get the better insight on the subject. This book was created by the Engineering team of HadoopExam which has in depth knowledge about the Hadoop Cluster Administration and Created HandsOn Hadoop Administration training. The team target is to make you learn the subject as in depth as possible with the minimum effort hence we have material in Question, Answers format, On-demand video trainings, E-Books, Projects and POC etc. We are delighted when learners come and give the feedback about our material and become repeat subscriber because they regularly get new material as well as updated material. Again all the best and please provide the feedback on the admin@hadoopexam.com or hadoopexam@gmail.com . Wherever possible we are trying to help you in your career.
While traveling (Home to Office)
We have training subscriber from TCS, IBM, INFOSYS, ACCENTURE, APPLE, HEWITT, Oracle , NetApp , Capgemini etc.
Question-1: What is Apache Ambari?
Answer: If you wanted to create a BigData Cluster using Hadopp, HBase, Spark and all supported software. You need the Apache Ambari. Ambari is mainly used by Hortonworks® and Cloudera® to setup the Data platform. This helps in managing and monitoring the cluster.
Question-2: What all activity you can accomplish using the Apache Ambari?
Answer: There are multiple ways you can use the Apache Ambari like Web UI and REST APIs. Using that we can deploy new cluster, make the configuration changes, and also monitor the all services across the nodes in the cluster. Lets assume you have 1000 node cluster, and you wanted to monitor the service on all the 1000’s of the nodes. Then you would be using the Apache Ambari to monitor and even manage all the services across the cluster.
Question-3: What all platforms are currently supported using the Apache Ambari?
Answer: Currently HDP (Hortonworks Data platform and Hortonworks Dataflow are supported.
Question-4: Which toolkit is used by the Apache Ambari Metrics monitor utility?
Answer: Ambari Metrics monitor uses the Python Library (psutil) which requires the gcc and python-devel packages.
Question-5: Which command you would be using to check the free memory on any Linux host?
Answer: You would be using free -m command.
Question-6: Is there any dependency between the requirement of free memory and disk space with the number of nodes in the HDP cluster?
Answer: Yes, certainly it has. It depends how many nodes you have in your cluster and based on that you need memory and disk. In general for 1 node 1GB memory and 10GB disk space needed. Similarly for 100 node , 4GB memory and 100GB disk space. You need to check specific version to get the full detail.
Question-7: What is the ulimit and how it related to Ambari?
Answer: ulimit are the utility on Linux system to check, what all resources limits. For example we need to check number of open file descriptor supported you need to run below commands.
Check Soft limit using
ulimit -Sn
Check hard limit using
ulimit -Hn
Question-8: How can i increase the open file limit descriptor values?
Answer: you can do using the command
ulimit -n 10000
In above command it will set the limit to 10000
Question-9: What is the FQDN?
Answer: You wanted to know the Fully Qualified name of the your linux host on which you would like to install the Ambari, so you could use the below command.
hostname -f
Question-10: Should we install Ambari on more than one host?
Answer: Yes, you should install the Apache Ambari at least on 3 nodes, if you are planning in production environment. Where one node will be master and another two node as slaves. So Ambari itself has its own cluster to manage the HDP cluster.
Question-11: Can you give example the kind of data stored by the Apache Ambari for HDP cluster?
Answer: As Ambari is a management tool for the Hadoop (HDP) cluster it needs to store various services data. Like for HDFS it needs to store NameNode, DataNode , Secondary NameNode, Oozie , YARN and Zookeeper data.
Question-12: Can I store this data on the /tmp directory?
Answer: No, you should not use the /tmp directory at all. Because if you use the /tmp directory then data would be deleted. And cluster would fail, hence you should must have persistent storage locations.
Question-13: What is the best way to install Ambari agent on all the 1000 host in HDP cluster?
Answer: As there are 1000 nodes in the cluster, we should not manually install the Ambari agent on all the nodes in the cluster. Rather we should have password less ssh connection setup done between Ambari host and all the nodes in the cluster. And Ambari Server hosts uses the SSH public key authentication to remotely access and install the Ambari Agent.
Question-14: What is the Service account in Ambari?
Answer: As Ambari manages the services like Spark, HDFS, HBase etc on the HDP cluster and to manage these services it requires that you have at least one service account created for each service.
Question-15: What is the default user name and password for Ambari administration?
Answer: The default username and password for the Ambari Administration is created while installing and that is admin/admin. And this user has the full access and control on the Ambari.
Question-16: If you have three HDP cluster created Dev, UAT and Prod. How many Ambari installation you can have?
Answer: Usually if you are beginner, you get confused with that, you need 3 Ambari installation one for each cluster. But that is not true. You can manage more than one cluster using the single Ambari setup.
Question-17: If I have Ambari admin privileges, then what all I can do?
Answer: Become Hadoop Administrator is very demanding profile. You can check all the available Hadoop Admin training for HDP, Cloudera etc on the HadoopExam.com (visit now). If you are Ambari Admin then you can create a cluster, manage the users for that cluster, and create groups etc. And default admin user has all this permission. Even being an Amabari administrator, you can grant the same or other permissions to the another user as well.
Question-18: What do you mean by group in the Ambari?
Answer: In Ambari group represents the collection of users in a group.
Question-19: What all possible type of groups can be created in Ambari?
Answer: You can create two different type of groups in Ambari as below
Question-20: What do you mean by Principal?
Answer: All the users or group which can
be authenticated
and authorized by the LDAP are known as Principal. Because,
you would always be
granting permissions to the either user or
group.
Question-21: How
would
i get access to all 118 questions, because i am already on
the path of
becomming Hadoop Administrator. ?
About the
Apache Ambari
If you are Hadoop administrator or want to become Hadoop Administrator then first of all we would like to tell you this one of the good choices for your career. And for becoming Hadoop administrator Apache Ambari has a great role to play. Apache Ambari is an open source component to manage, monitor and administer the various components and services in the Hadoop Cluster.
With the Ambari Web UI you can add, delete, stop, start services in the Hadoop cluster which is managed by the Ambari. Even you can register the already created or deployed Hadoop Cluster with the Ambari. As we know that Hadoop cluster can have more than 10000 nodes in a single cluster. And managing so many nodes in the cluster is not an easy task, its very complex activity and I can assure when you start using the Apache Ambari you will feel how easy to manage to 10000 nodes cluster or farm of the server. But you must know various operations and feature of the Ambari Web UI before you do or make any changes to the cluster configuration.
Having knowledge about the Apache Ambari in and out can help you become very good Hadoop Administrator and certainly have high impact among the team member. Learning Ambari can help you grow vertically as well as horizontally (like Hadoop Cluster 😊) in your career as well and another important factor is it can help you to fetch the higher salary as well. You can search any job portal and check the demand for this profile your own. There are lack of such administrator in the market and companies are training their own current employee for the same because they are not easily available in the market.
Even we recommend that you go through the documentation about Apache Ambari during your routine administration work but for the interview you have to be very specific and should know the general aspect and use of the Apache Ambari, because interviewer is smart enough he would try to find it out that you really have hands on knowledge of the Ambari service or not. And these question and answers will certainly help you get through at least for Apache Ambari perspective and for other Hadoop component interview questions, you can check the HadoopExam.com website as well. There are various Hadoop Solution provided vendor and these three are very popular Cloudera, Hortonworks, and MapR. However, reading and going through entire documentation is not possible in one sitting. But this book you can finish it at least in one or two sittings for sure, if you wish to. This would help you learn Apache Ambari quickly.
Let’s go through the question and answer and learn various things like Rolling start, maintenance mode, basics of Kerberos keytab files and many more concepts. There are too many things to learn and my experience says that as many things you learn it would make easier to learn further things and quicker as well. It is an adventures activity to learn how to manage 10000+ node cluster using the single Web UI. You would become a bond for such a large cluster management in reality. And HadoopExam.com would try on each and every possible step you learn the things and work.