LiveCareer-Resume

data engineer resume example with 4+ years of experience

Jessica Claire
, , 609 Johnson Ave., 49204, Tulsa, OK 100 Montgomery St. 10th Floor
Home: (555) 432-1000 - Cell: - resumesample@example.com - : - -
Summary

Results-driven data engineering professional offering four-plus years of experience in designing, developing, and maintaining high-performance data pipelines, along with providing advanced analytical solutions. Skilled in harnessing the power of cutting-edge tools and implementing industry best practices to establish and manage scalable data ecosystems. Currently serving as a member of the data engineering team at the New York Stock Exchange.

Skills
  • Programming/ Querying languages: Python, JavaScript, Shell scripting, Java, SQL
  • Databases: SQL Server, Oracle, MySQL, PostgreSQL, Snowflake, Redshift, DynamoDB, MarkLogic, MongoDB
  • Big Data frameworks/ Tools/ Libraries: Spark, Hive, Kafka, Hadoop, NiFi, PySpark, SPARQL, Pandas, DBT, NumPy, Boto3, SQLAlchemy
  • Schedulers and AWS: Airflow, Kronos, S3, EC2, Glue, EMR, Kinesis, Athena, Lambda, SNS, SQS, CloudWatch, QuickSight
  • Data visualization: Qliksense, Power BI, Alteryx
  • Control systems: Git, Jira, Confluence
Experience
04/2022 to Current
Data Engineer Bank Of America Corporation Baton Rouge, LA,
  • Developed and automated multiple data pipelines to fetch data from diverse financial sources and APIs. Leveraged Python's capabilities to efficiently process and transform the data before loading it into Snowflake. Improved data retrieval efficiency and enabled streamlined analysis, empowering data-driven decision-making.
  • Developed Airflow operators using python to interact with services like EMR, Athena, S3 and Snowflake. These are being used across by over 70 teams. Designed and implemented several ETL pipeline with up to 500 distributed concurrent tasks using Apache Airflow.
  • Developed Python-based Lambda functions to facilitate the migration of files from a staging S3 bucket to a formatted S3 Data Lake and from the Data Lake to Snowflake. Employed SQS and SNS for seamless integration and notification. Utilized Terraform to configure and version the infrastructure.
  • Reduced ETL operation cost by 30% on ETL pipeline by processing delta dataset only in Spark.
01/2021 to 04/2022
Data Engineer II Usaa Lavon, TX,
  • Designed and implemented a scalable real-time data pipeline to process semi-structured data by integrating more than 150 million raw records from 50+ data sources using Kafka and Python and stored processed data in Snowflake.
  • Created ETL pipelines using Hive and Spark to obtain data from the different systems, transform it, and create Data Mart for the Surveillance project. The Data Mart has successfully served as a central data repository utilized by over 30 teams for their Power BI reporting needs
  • Built a robust stream processing framework on EMR using pySpark to extract truck GPS data from Kinesis Data Stream and seamlessly transform it into pandas dataframes from spark dataframes. Effectively orchestrated the entire workflow using Airflow, resulting in significant cost savings of $600k through optimized data processing and analysis.
  • Worked with stakeholders on a daily basis to understand pipeline requirements, data quality testing etc.
06/2019 to 12/2020
Data Governance Engineer Republic Services City, STATE,
  • Built ETL pipelines on AWS Glue to move data from S3 to RDS. Replaced IICS with glue for data integration from S3 to ODS which Provided potential savings of $100K on Informatica license.
  • Orchestrated MSSQL to AWS Redshift migration using Airflow and SQL replication. Resulted in 30% performance improvement.
  • Built data pipelines handling millions of records per day using Python, SQL, S3, SQS, Redshift, DynamoDB etc.
  • Implemented a real-time pipeline for daily reports, accelerating refresh rates from 4 hours to an impressive 10-minute interval on AWS QuickSight
09/2018 to 05/2019
Credit Risk Data Analyst Intern Federal Home Loan Bank Of Des Moines City, STATE,
  • Efficiently migrated 1 TB data warehouse from SQL Server to Oracle. Optimized performance by migrating over 50 SQL procedures, resulting in a remarkable 60-70% improvement.
  • Leveraged QlikSense to design and implement interactive financial data dashboards, empowering Risk analysts to effectively monitor and assess a portfolio of major clients worth $108.9 billion.
  • Collaborated with analytics and credit risk team to improve data models that feed business intelligence tools, increasing data accessibility and fostering data-driven decision making across the organization.
Education
Expected in 05/2019 to to
Master of Science: Management Information Systems
Iowa State University - Ames, IA
GPA:
Expected in 05/2017 to to
Bachelor of Science: Computer Science
Gitam University - Visakhapatnam, India,
GPA:
Certifications

AWS Certified Solutions Architect – Associate - 2021 https://www.credly.com/badges/bb106f7b-c451-4e60-8ab8-c21f46906b84/public_url

SnowPro Core Certification - 2022 https://www.credly.com/badges/a6bb96f5-ad35-4bbd-a6bc-2b797f5285f3/public_url

By clicking Customize This Resume, you agree to our Terms of Use and Privacy Policy

Your data is safe with us

Any information uploaded, such as a resume, or input by the user is owned solely by the user, not LiveCareer. For further information, please visit our Terms of Use.

Resume Overview

School Attended

  • Iowa State University
  • Gitam University

Job Titles Held:

  • Data Engineer
  • Data Engineer II
  • Data Governance Engineer
  • Credit Risk Data Analyst Intern

Degrees

  • Master of Science
  • Bachelor of Science

By clicking Customize This Resume, you agree to our Terms of Use and Privacy Policy

*As seen in:As seen in: