Good knowledge of installing, configuring, managing, capacity planning and administering of Hadoop clusters. Experience in Hadoop eco system component MapReduce, Hive, Pig, Sqoop, Flume, and Oozie. Working knowledge of R to statistically analyze and explore data. Knowledge of various Machine Learning Techniques and their implementation using Apache Mahout. Experience of designing and writing MapReduce programs & UDFs for both Hive & Pig. Experience in setting up clusters in Amazon EC2 & S3. Knowledge of Hadoop security: Kerberos, authorization security deployments. Experience in writing scripts using Pig Latin Scripts & Perl in Linux. Experience in developing PL/SQL packages, stored procedures, functions & triggers. Experience in writing PERL scripts to move files from internal systems to AWS cloud, customer sites and vice versa. Expertise in designing, developing and implementing many Java based applications. Led multiple off-shore teams for developing various applications.
Solid experience as Big Data Consultant, and Big Data solution architect
Expert knowledge of big data technologies with extensive hands on experience. Led design & implementation of big data projects in the education and financial domains. Experience in setting up Hadoop clusters in AWS. Also has very deep expertise in designing & implementing large enterprise systems in the financial domain.
Familiar with Machine Learning technologies including the language R and algorithms K-Means, Random Forest and Naïve Bayes.
Cloudera certified Hadoop developer.
Around 4+ years of experience in designing big data applications with extensive hands on experience in Hadoop echo system components. Led the effort from creating POC, demoing it to stakeholders, designing, developing and implementing in production environment.
Around 8+ years of experience in design and development of various java and RDBMS based applications.
Around 2+ years of experience in the development of various Mainframe based applications using PL/1.
Led development teams, used agile methodologies and software development life cycle (SDLC) to deliver applications on time and within the budget.
Big Data Architect, 03/2015
to Current Infosys/DIRECTV – Los Angeles,
Supplier management system at DIRECTV was designed to make payments to its content providers.
It was developed using Oracle Pro C.
This application has been redesigned, developed and implemented using Hadoop eco system components and Data Torrent; is an enterprise-grade unified stream and batch processing engine integrated with YARN.
Responsibilities: Architected and led the development of SMS application using Data Torrent and various Hadoop Eco system components such as HDFS, Sqoop, Hive and various frame works using shell scripts, and Java.
Leading and guiding off-shore development team of fifteen members to implement the solution.
Responsible for providing and presenting solutions, and POCs for CRs and new requirements to the client.
Designed and involved in developing schemas in Hive.
Involved in writing SQL queries, creating views, triggers and audit tables in Oracle for data migration and data retention.
Built, tested and deployed scalable, highly available and modular software products.
Streamlined deployment process by developing continuous integration tools.
Strengthened developmental methodologies by introducing a code quality document.
Wrote and implemented scripts to enhance user experience and integrated scripts with the CMS.
Designed and developed transactional and analytical data structures.
Software Development Lead, 06/2003
to 12/2014 CTB/McGraw-Hill Education – Monterey,
CTB is one of the main players in education domain to conduct student assessment tests in multiple states for K-12 students.
to 04/2003 Bank of America – San Francisco,
Bank of America is one of the leading financial institutes.
It serves individual customers, businesses, corporates and governments with a range of banking needs.
Programmer Analyst, 01/2001
to 05/2002 Airlines Software Development Consultancy TCS and Singapore Airlines – Chennai,
Extensively involved in maintaining and implementing new business requirements in Singapore Airlines HR applications using Mainframe based technologies.