data engineer resume example with 6+ years of experience

Jessica Claire
, , 609 Johnson Ave., 49204, Tulsa, OK 100 Montgomery St. 10th Floor
Home: (555) 432-1000 - Cell: - - : - -
Professional Summary
  • Over all 4+ years of IT experience in Data Engineering, Analytics and Software development for Banking and Retail customers.
  • Strong Experience in data engineering and building ETL pipelines on batch and streaming data using Pyspark, SparkSQL.
  • Good working exposure on Cloud technolgies of AWS - EC2, EMR, S3, Lambda, SQS, SNS.
  • AWS Certified Solutions Architect - Associate-2020.
  • Strong programming knowledge of Java 7, Java 8 and experience with scripting languages - Python 3.
  • Strong experience in Python libraries - Numpy, Pandas, Matplotlib.
  • Strong experience in Data Engineering technologies including Hadoop 2, Spark , Elastic Map Reduce.
  • Strong Working experience in Java J2EE,Spring Rest, Spring Boot, Spring HIbernate, RESTful Services, Spring Data , JSON , XML for web application server side scripting.
  • Proficient in performing EDA (Exploratory data analysis), Root Cause Analysis, Impact Analysis on large volume of datasets.
  • Experienced in querying Snowflake, Oracle, Redshift, MS SQL server databases for OLTP and OLAP.
  • Solid understanding of RDBMS database concepts including performance tuning and Query optimization
  • Experience in complete Software Development Life Cycle (SDLC) involving Analysis, Design, Development and Testing.
  • Advanced working SQL knowledge and experience working with relational databases, query optimization (SQL) as well as working familiarity with a variety of databases.
  • Experience building and optimizing ‘big data’ data pipelines, architectures and data sets on building Data Lakes.
  • Experience performing root cause analysis on internal and external data and processes to answer specific business questions.
  • Strong analytic skills related to working with unstructured datasets.
  • Experience with building processes supporting data transformation, data structures, metadata, dependency and workload management.
  • Working knowledge of message queuing, stream processing, and highly scalable ‘big data’ data stores.
  • Experience creating APIs and interfacing with front-end systems
  • Experience with stream-processing systems using PySpark.
  • Strong knowledge of data structures, algorithms, operating systems, and distributed systems fundamentals.
  • Working familiarity with networking protocols (TCP/IP, HTTP) and standard network architectures.
  • Team player with Strong sense of ownership.
Expected in 05/2019 to to
Master of Science: Computer Science
New Jersey Institute of Technology - Newark, NJ
  • Graduated with 3.5 GPA
  • Majored in Data Science
Expected in 06/2013 to to
Bachelor of Technology: Electrical & Electronics Engineering
Sastra University - Tanjore, Tamil Nadu,India,
  • Programming Languages : Python 3, Java 8, SQL, NoSQL
  • Databases : Snowflake, AWS Redshift, Cassandra, Oracle, MSSQL
  • Cloud Technologies : AWS EC2, EMR, AWS Lambda, S3, IAM,AWS Glue
  • Web Technologies: Java J2EE, Spring Stack
  • Interface Development Environment (IDE) : Eclipse, IntelliJ, Anaconda Jupyter notebooks,Spyder, Oracle developer
  • Operating System : Mac, Windows
  • ETL Tools : SQL Server SSIS, Snow pipe, Athena
  • Reporting tools : Tableau, Power BI
  • Apache Spark, Spark SQL
Work History
08/2019 to Current
Data Engineer Assurant Lakeland, FL,
  • Working currently as Data Engineer in Risk Management Team in Auto Finance Division of Financial Client
  • Achieved 400% growth in data quality check creation process by creating Rule Automaton tool in Python
  • Worked on developing ETL pipelines on S3 parquet files on data lake using AWS Glue
  • Performed data analytics on DataLake of Capital one using Pyspark on databricks platform
  • Responsible for assessing and improving the Quality of Customer Data.
    Worked on end-to-end data quality process setup on AWS for entire Financial Auto Loans division
  • Experience with AWS cloud services: EC2, S3, EMR, RDS, Athena, and Glue
  • Analyzed data quality issues through Exploratory data analysis (EDA) using SQL,Python and Pandas
  • Worked on creating Data Navigator portal to provide overview of data load and data quality using R,Python and Snow pipe improving efficiency of analysis by 200%
  • Regulatory Reporting. Streamlined the process of stress test regulatory reporting by developing a detailed process tracker in Excel and Python
  • Performed data quality issue analysis using Snow SQL by building analytical warehouses on Snowflake
  • Reconciled monthly stress test and quarterly financial reports in order to ensure regulatory compliance
  • Worked on creating automation scripts leveraging various Python libraries to perform accuracy checks from various sources to target databases
  • Worked on building Python scripts to generate heatmaps to perform issue and root cause analysis for data quality report failures
  • Interacting with Financial services stakeholders delivering regulatory reports and to recommend best remediation strategies to ensure pristine quality of high priority usage data elements by building various
    analytical dashboards using excel, python plotting libraries
  • Involved in the code migration of quality monitoring tool from AWS EC2 to AWS lambda and built logical datasets to administer quality monitoring on Snowflake warehouses.
    Environment: Python, Databricks, Spark SQL,PySpark, Snow SQL, Pandas, Numpy, Matplotlib, seaborn, Excel, Tableau, AWS EC2, AWS S3, AWS Lambda , AWS EC2, S3, Glue.
    Lambda, OneNote(Jupyter Notebook),Spyder, Anaconda , Nebula, Linux Shell Scripting
01/2016 to 04/2016
Software Engineering Analyst Wsp Lexington, KY,
  • Worked on logistics application to do shipment and field logistics of Energy and Utilities Client
  • Worked agile in a team of 4 members and contributed to the backend development of application using microservices architecture
  • Contributed to the spring boot based multi tier web application using Java 8, Spring Boot , Microservices
  • Partially contributed to the design of the Spring Data Apis and fully contributed in development of the APis to access data from the Relational datastores
  • Gained expertise in writing SQL queries against MSSQL server with query optimization
  • Performed data analysis using Python and involved in critical problem solving situation and troubleshooting abilities
  • Leveraged various python modules to enhance the data validation, testing and automated daily processes
  • Environment: Java 1.7, Spring 4, Spring Boot, Microservices, Oracle 11g, Junit, Mockito
08/2013 to 01/2016
Project Engineer Constellation Brands Philadelphia, PA,
  • Worked with Fedex client for Fedex Locker Management System (FLMS) to develop RESTful services to the Fedex kiosk( FSNG) in request-response style
  • Extensive Involvement in Requirement Analysis and system implementation, SDLC phases
  • Contributed to a team of 6 in an agile environment in developing new interfaces for the FLMS application
  • Contributed in developing RESTful services and business logic in the backend of the FLMS portal to serve requests of several Fedex Ship and Get Kiosk applications on front end using Spring REST , Spring MVC
  • Worked on Object Relation mapping technologies like JPA(Hibernate) to develop the Data Access Layer and Repositories layer
  • Worked on RabbitMQ Message Queueing System to integrate FLMS application with Cargo applications and other downstream applications to receive parcel data in real time and batches
  • Worked on developing web services to interact with cross platform applications (FSNG) to exchange the data in the form of XML and SOAP protocol for consuming WCF services
  • Implemented OLTP systems to the backend by creating complex SQL Queries, Reusable Triggers, Functions, Stored procedures using PL/SQL
  • Worked in pair programming, Code reviewing and Debugging
  • Involved in unit test development using Mockito and Bug Fixing
  • Involved in UAT and production deployments and support activities
  • Tools & Technologies: Java SE 7, Spring 4.0, Spring REST, Spring JDBC,Hibernate , Oracle 11g, Maven, Eclipse, SVN, Bugzilla

AWS Certified Solutions Architect - Associate(JWS3E0KCHM4EQ69L)

Data Analyst Nano Degree- Udacity


By clicking Customize This Resume, you agree to our Terms of Use and Privacy Policy

Your data is safe with us

Any information uploaded, such as a resume, or input by the user is owned solely by the user, not LiveCareer. For further information, please visit our Terms of Use.

Resume Overview

School Attended

  • New Jersey Institute of Technology
  • Sastra University

Job Titles Held:

  • Data Engineer
  • Software Engineering Analyst
  • Project Engineer


  • Master of Science
  • Bachelor of Technology

By clicking Customize This Resume, you agree to our Terms of Use and Privacy Policy

*As seen in:As seen in: