Skip to main content

Hadoop Interview Questions


1) What is Big Data?

Big Data is nothing but an assortment of such a huge and complex data that it becomes very tedious to capture, store, process, retrieve and analyze it with the help of on-hand database management tools or traditional data processing techniques.

2) How is the analysis of Big Data useful for organization?

Effective analysis of Big Data provides a lot of business advantage as organizations will learn which areas to focus on and which areas are less important. Big data analysis provides some early key indicators that can prevent the company from a huge loss or help in grasping a great opportunity with open hands! A precise analysis of Big Data helps in decision making! For instance, nowadays people rely so much on Facebook and Twitter before buying any product or service. All thanks to the Big Data explosion.

3) What is difference between traditional RDBMS and Hadoop?

Traditional RDBMS is used for transactional systems to report and archive the data, whereas Hadoop is an approach to store huge amount of data in the distributed file system and process it. RDBMS will be useful when you want to seek one record from Big data, whereas, Hadoop will be useful when you want Big data in one shot and perform analysis on that later

4) What is Fault Tolerance?

Suppose you have a file stored in a system, and due to some technical problem that file gets destroyed. Then there is no chance of getting the data back present in that file. To avoid such situations, Hadoop has introduced the feature of fault tolerance in HDFS. In Hadoop, when we store a file, it automatically gets replicated at two other locations also. So even if one or two of the systems collapse, the file is still available on the third system.

5) What is Datanode?

Datanodes are the slaves which are deployed on each machine and provide the actual storage. These are responsible for serving read and write requests for the clients.

6) What is HeartBeat in HDFS?

A heartbeat is a signal indicating that it is alive. A datanode sends heartbeat to Namenode and task tracker will send its heart beat to job tracker. If the Namenode or job tracker does not receive heart beat then they will decide that there is some problem in datanode or task tracker is unable to perform the assigned task.

7) How indexing is done in HDFS?

Hadoop has its own way of indexing. Depending upon the block size, once the data is stored, HDFS will keep on storing the last part of the data which will say where the next part of the data will be.

Thinks The above discussed question may be useful for you . To become expert in Hadoop you can enroll with Hadoop Training @ Besant Technologies. Besant Technologies offers Hadoop Training in Chennai with best placement support






  



Comments

Popular posts from this blog

Why Is Python So Famous?

Have you ever come across word the python anywhere? If no, let me brief you about “Python”. Python is an interpreted, object oriented and high level programming languages more over it has built in data structures which makes it attractive for rapid application development as well suitable for scripting language. The applications of python languages are wider it used to create GUI based desktop applications, web applications and business applications. So the opportunities for python are overwhelming, so grab the opportunity by enrolling Python Training in Chennai Advantages of python Python has numerous third party modules which make it capable for interacting with other language and platform Python is open source software which makes it free to use and distribute Python provides excellent readability which helps beginners to learn the language fastly Python offers user friendly data structures Python increases the speed of application developmen...

Why Certification in Hadoop?

Today the hot word in the IT industry is “ Hadoop ”. Since every companies are dealing with large amount of data. Many of the IT recruiters say there would be larger demand for hadoop trained professional.  Many of the professional who want to make career in hadoop are planning to get trained in Hadoop and they want to have certification in hadoop certification moreover   hadoop is gaining popularity in all over world as discussed many of industries like financial service,utilities,media ,retail and paramedical are moving towards hadoop for processing the big data Today many corporate are looking for hadoop trained professional. Having certification would give them assurance that you are master in hadoop and more over having certification normally would be an added advantage and it would help you to get job easily and you can demand your dream salary since you have certification Let me brief you about the power of certification in hadoop IT recruiter’s p...

Software Testing Interview Questions

1) What is PDCA cycle and where testing fits in? There are four steps in a normal software development process. In short, these steps are referred as PDCA. PDCA stands for Plan, Do, Check, Act. Plan:  It defines the goal and the plan for achieving that goal. Do/ Execute:  It depends on the plan strategy decided during the plan stage. It is done according to this phase. Check:  This is the testing part of the software development phase. It is used to ensure that we are moving according to plan and getting the desired result. Act:  This step is used to solve if there any issue is occurred during the check cycle. It takes appropriate action accordingly and revises the plan again. The developers do the "planning and building" of the project while testers do the "check" part of the project.   2) What is the difference among white box, black box and gray box testing? Black box Testing:  The s...