hadoop administrator resume example with 7+ years of experience

Jessica Claire
  • , , 609 Johnson Ave., 49204, Tulsa, OK 100 Montgomery St. 10th Floor
  • H: (555) 432-1000
  • C:
  • Date of Birth:
  • India:
  • :
  • single:
  • :
Professional Summary

ü 7+ year’s experience in IT industry including Hadoop Administrator with 5 years of experience in Apache Hadoop by providing design and support on various Big Data Analytical Tools.

ü Extensive knowledge and understanding of the Hadoop architecture and various components like HDFS, YARN, Name Node, Data Node, Hive, Hue, Hbase, Impala, Kafka………...(continue)

ü Installing, configuring and using ecosystem components like HDFS, Hive, spark, sqoop, Kafka, MapReduce, Oozie, YARN, Zookeeper on CDH/CDP.

ü Experience on Commissioning, Decommissioning, Balancing, and Managing Nodes and tuning server for optimal performance of the cluster.

ü Experience in maintaining high availability of Name Node metadata and other important sensitive data residing on cluster by building functional BDR cluster.

ü Hands on experience in Hadoop cluster capacity planning, performance tuning, cluster monitoring, troubleshooting.

ü Experience in minor and major upgrades of Hadoop and Hadoop Ecosystem.

ü Enabling High Availability to HDFS, YARN, Hive, Impala, Services to improve redundancy from failures.

ü Manage, coordinate and implement software upgrades, patches, and error fixes on servers, workstations, network as well as hardware.

ü Experience to configure security system by implementing Kerberos for Authentication and Sentry/Ranger for Authorization.

ü Hands on experience on data governance, data encryption and audit by configuring Cloudera Navigator/vormatric and SSL/TLS.

ü Authorize users and groups to the databases, tables, views and column using sentry/ranger and FACL system.

ü Responsible for scheduling and upgrading these servers throughout the year to the latest versions of software.

ü Experience in Data modeling for Data Warehouse/Data Mart development, Data Analysis for Online Transaction Processing (OLTP) and Data Warehousing (OLAP)/Business Intelligence (BI) applications.

ü Involved in various projects related to Data Modeling, System/Data Analysis, Design and Development for both OLTP and Data warehousing environments.

ü Facilitated data requirement meetings with business and technical stakeholders and resolved conflicts to drive decisions.

ü Comprehensive knowledge and experience in process improvement, normalization/de-normalization, data extraction, data cleansing, data manipulation.

ü Excellent understanding and knowledge of NOSQL databases like MongoDB, HBase, and Cassandra.

ü Solid understanding of all phases of development using multiple methodologies i.e. Agile with JIRA, Kanban board along with ticketing tool.

ü Real-time experience on data ingestion tools- such as Sqoop ,Talend, Informatica and configure ETL processes as well as manage them on daily bases .

ü Experience in importing and exporting data between HDFS and Relational Database Management systems using different ETL .

ü Experience in monitoring the health of Hadoop cluster and also performing multiple administrative cluster maintenance such as commissioning /decommissioning data nodes.

ü Experienced about converting Hive/SQL queries into Spark transformations using Spark RDDs.

ü Hands on experience in analyzing Log files for Hadoop and Eco-System services and finding root cause.

ü Practical knowledge on functionalities of Hadoop Daemons, interactions between them, resource utilization and dynamic tuning of resources to ensure efficiency of cluster performance.

ü Creating job pools, assigning users to pools and restricting production job submissions based on pools to ensure optimal performance.

ü Experience with replicating data across data centers for disaster recovery scenarios through BDR and HDFS snapshots.

Provided 24x7 supports to ensure round the clock availability.


    SQL & NOSQL Data Storage: MYSQL, MSSQL, HBase

    Big Data Ecosystem: Hadoop, MapReduce, HDFS, HBase, Zookeeper, Hive, Pig, SQOOP, Oozie, YARN, Impala, Kerberos, Sentry and Spark

    Management Tool: Cloudera Manager and Hortonworks.

    ETL Tool: Talend and Sqoop.

    Security: SSL/TLS, AD,LDAP, Kerberos, Sentry and Ranger.

    Cluster Management Tools: Kubernetes.

    Operating Systems/Platforms: Linux, Unix, Ubuntu, Windows, MacOS .

    Cloud Computing Services: VMware and AWS.

Work History
Cognizant Technology Solutions Dover, DE,

ü Collaborate in identifying the current problems, constraints and root causes with data sets to identify the descriptive and predictive solution with support of the Hadoop HDFS, MapReduce, Hive.

ü Installed/Configured/Maintained Apache Hadoop clusters for application development and Hadoop tools like Hive, HBase, Zookeeper and Sqoop.

ü Installation and Configuration of Hadoop Cluster for both CDH & CDP.

ü Working with Cloudera Support Team to Fine tune Cluster.

ü Moving the Services (Re-distribution) from one Host to another host within the Cluster to facilitate securing the cluster and ensuring High availability of the services.

ü Developed Map Reduce jobs to analyze data and provide heuristics reports

ü Worked on installing cluster, commissioning & decommissioning of Data Nodes, NameNode recovery, capacity planning, and slots configuration.

ü Involved in Installing and configuring Kerberos for the authentication of users and Hadoop daemons, AD integration (LDAP) and Ranger authorization.

ü Configured various property files like core-site.xml, hdfs-site.xml, mapred-site.xml based upon the job requirement.

ü Collaborating with project BA to ensure connectivity and database access via BI tools like DbVisualizer, Talend.

ü Performed transformations, cleaning and filtering on imported data using Hive, Map Reduce, and loaded final data into HDFS.

ü Setting up Linux users, setting up Kerberos principals and testing HDFS, Hive, Pig and MapReduce access for the new users

ü Configuring, Monitoring and Management Tools using Cloudera

ü Applying Patches and Perform Version Upgrades

ü Incident Management, Problem Management and Change Management

ü Performance Management and Reporting

ü Recover from Name Node failures

ü Installation and Configuration of other Open Source Software like Hive, HBASE, Flume and Sqoop.

ü Integration with RDBMS using Sqoop and JDBC Connectors.

ü Working with Dev Team to tune Job Knowledge of Writing Hive Jobs.

Environment: RHEL, puppet, CDH distribution, Tableau, Datameer, HBase,

Cloudera Manager, YARN, Hive, Flume.

HADOOP ADMINISTRATOR, 01/2018 - 07/2020
Cognizant Technology Solutions Downers Grove, IL,

ü Working on Hadoop infrastructure and playing a key role in supporting Hadoop cluster.

ü Utilizing components such as Yarn, Zookeeper, Journal Nodes, Sqoop, Hive, Tez, Hue, Hbase.

ü Worked with system engineering team to plan and deployed new Hadoop environments and expand existing Hadoop clusters.

ü Involved in Installing and configuring Kerberos for the authentication of users and Hadoop daemons, AD integration (LDAP) and Ranger authorization.

ü Responsible for Cluster maintenance, commissioning and decommissioning Data nodes, Cluster Monitoring, Troubleshooting, Manage and review data backups, Manage & review Hadoop log files.

ü Configured various property files like core-site.xml, hdfs-site.xml, mapred-site.xml based upon the job requirement.

ü Responsible for data extraction and data ingestion from different data sources into Hadoop Data Lake by creating ETL pipelines using Hbase and Hive.

ü Upgraded Apache Ambari and HDP stack from older versions to latest versions.

ü Design and Configure the Cluster with the services required (Ranger, Hive server2, Kerberos, HDFS, Hue, Hive, Zookeeper).

ü Design and maintain the Name node and Data nodes with appropriate processing capacity and disk space.

ü Responsible for troubleshooting issues in the execution of MapReduce jobs by inspecting and reviewing log files.

ü Monitored workload, job performance and capacity planning using Apache Ambari.

ü Automated Setup Hadoop Cluster, Implemented Kerberos security for various Hadoop services using HDP distribution.

ü Handled importing of data from various data sources, performed transformations using Hive, MapReduce.

ü Experience in Importing and Exporting data using Sqoop from HDFS TO RDBMS and vice-versa.

ü Performed HDFS cluster support and maintenance tasks like adding and removing nodes without any downtime in cluster and node label.

ü Configured Resource management in Hadoop through dynamic resource allocation.

ü Worked with application teams to install operating system, Hadoop updates, patches, version upgrades as required.

ü Installed Oozie workflow engine to run multiple Hive Jobs.

ü Application development using RDBMS, and Linux shell scripting.

ü Experience in installing, configuring and optimizing Hortonworks HDP in a Multi Clustered environment.

ü Set up KDC Kerberos trust between multiple Hadoop clusters and tested validated adding peers and business development representative job set up.

Environment: Cloudera, Cloudera Manager, HDFS, Map Reduce, YARN, Ambari, Pig, Hive, Sqoop, Oozie, Zookeeper, Kerberos, Apache Ranger.

HADOOP ADMINISTRATOR, 08/2015 - 12/2017
Cognizant Technology Solutions Eagan, MN,

ü Installed/Configured/Maintained Apache Hadoop clusters for application development and Hadoop tools like Hive, HBase, Zookeeper and Sqoop.

ü Installing and Upgrading Cloudera CDH & Hortonworks HDP Versions on lower and poc environments.

ü Installed and Configured Sqoop to import and export the data into HDFS and Hive from Relational databases.

ü Close monitoring and analysis of the MapReduce job executions on cluster at task level and optimized Hadoop clusters components to achieve high performance.

ü Integrated CDH clusters with Active Directory and enabled Kerberos for Authentication.

ü Worked on commissioning & decommissioning of Data Nodes, NameNode recovery, capacity planning and installed Oozie workflow engine to run multiple Hive Jobs.

ü Implemented High Availability and automatic failover infrastructure to overcome single point of failure for Name node utilizing zookeeper services.

ü Used Hive and created Hive tables and involved in data loading and writing Hive UDFs and worked with Linux server admin team in administering the server hardware and operating system.

ü Worked closely with data analysts to construct creative solutions for their analysis tasks and managed and reviewed Hadoop and Hive log files.

ü Collaborating with application teams to install operating system and Hadoop updates, version upgrades when required.

ü Automated workflows using shell scripts pull data from various databases into Hadoop.

Linux Administrator, 02/2014 - 07/2015
Criterion Systems Beltsville, MD,

ü Performed software installation, upgrades/patches, performance tuning and troubleshooting of all the Linux servers in the environment.

ü Performed vulnerability testing of all the Linux servers and providing right solutions.

ü Experienced in Red Hat Linux package administration using YUM

ü Linux Installation and configuration from scratch and regular monitoring.

ü Experienced in Package management using RPM, YUM and up to date in Red Hat Linux.

ü Performed automated installations of Operating Systems using kickstart for Linux. Remote monitoring and management of server hardware.

ü Escalation Management: identify, manage, and lead escalations through L3. Work with others to help manage escalations through L5 Quality may provide feedback/influence change in internal and/or vendor-provided products/service offerings

ü Working knowledge of network infrastructure in relationship to server infrastructure

ü Support distributed-development and test environment management activities for large enterprise-wide deployments

ü Server administration of local and remote systems via Command Line Interface (CLI), monitoring logs for stop processes, disk space utilization, hardware failures, applying patches to comply with AF TCNOs/NOTAMs/MTOs/etc

ü Perform common system administration tasks including but not limited to: adding users, configuring volumes, creating filesystems, configuring various system services, configuring software firewall, monitor system health

ü Engineer and automate the cloud computing environment, particularly defining the strategy and then execute the integration of different large scale computing environments

Bachelor of Science: Computer Science, Expected in
Darul Ihsan University - Dhaka ,Bangladesh,
Status -
bangla :
Native or Bilingual:
Negotiated :

By clicking Customize This Resume, you agree to our Terms of Use and Privacy Policy

Your data is safe with us

Any information uploaded, such as a resume, or input by the user is owned solely by the user, not LiveCareer. For further information, please visit our Terms of Use.

Resume Overview

School Attended

  • Darul Ihsan University

Job Titles Held:

  • Linux Administrator


  • Bachelor of Science

By clicking Customize This Resume, you agree to our Terms of Use and Privacy Policy

*As seen in:As seen in: