In Hadoop 2, Map Reduce jobs are executed using the YARN(Yet Another Resource Negotiator). Let us understand the different id’s that are created while executing a mapreduce application.
When a MR job is submitted by the client, the resource manager will first create the application ID. Application ID is composed of the time that the resource manager is started and an incrementing counter maintained by the RM to uniquely identify the application.
In the above example, 1622829088382 refers to the start time format of resource manager (not the application) and 0005 indicates that it is the fifth…
In real world the clusters are busy and the resources are limited, as a result the applications often need to wait to have some of its resources fulfilled. The YARN scheduler takes the responsibility of allocating the resources to applications based on some defined policies. In this article we are going to discuss about the three scheduling options available in YARN.
1. FIFO Scheduler:
In FIFO (FIRST IN FIRST OUT) scheduler, applications are placed in a queue and runs them in the order of submission. This scheduling option is simple understand and doesn’t require any configuration, but it is not…
Before we start our discussion on what exactly is Erasure coding, let us understand the below two terms and see how HDFS achieve them.
In HDFS the durability, reliability, read bandwidth and write bandwidth can be achieved by replication process.
To provide fault tolerance, HDFS replicates blocks of a file on different DataNodes depending on the replication factor. …
Hadoop Distributed File System (HDFS) is file system of Hadoop designed for storing very large files running on clusters of commodity hardware. Generally, when dataset outgrows the storage capacity of a single machine, it is necessary to partition it across number of separate machines. The file system that manages the storage across a network of machines are called distributed file systems.
In this topic, we will discuss about the different HDFS commands with examples. Most of the commands have similar functionality as that of the Unix commands.
This command is similar to that of Unix mkdir and is…
In this blog we are going to discuss about how to integrate Apache Kafka with Spark using Python and its required configuration.
How Kafka works ?
Kafka is a distributed system consisting of servers and clients that communicate via a high-performance TCP network protocol. It can be deployed on bare-metal hardware, virtual machines, and containers in on-premise as well as cloud environments.
Introduction to Parquet
Apache Parquet is a columnar open source storage format that can efficiently store nested data which is widely used in Hadoop and Spark. Initially developed by Twitter and Cloudera. Columnar formats are attractive since they enable greater efficiency, in terms of both file size and query performance. File sizes are usually smaller than row-oriented equivalents since in a columnar format the values from one column are stored next to each other, which usually allows a very efficient encoding. …