Profile
  • Computer Science Individual with problem solving skills and experience in PySpark, SQL, AWS services, Databricks, Delta Lakes etc.
  • Seeking technical responsibilities within an organization as an AWS Data Engineer
Technical Skills
  • Proficient in programming languages such as Python and SQL
  • Experience with AWS Services like AWS Glue, Lambda, CloudWatch, and Athena.
  • Experience with developing Data Pipelines using PySpark and Spark SQL.
  • Familiarity with big data technologies like Hadoop, HDFS, Hive, Sqoop.
  • Familiar with working with orchestration frameworks like Airflow.
  • Experience with version control systems such as Git.
  • Work Experience
    Data Engineer - Programmer Analyst, Cognizant

    Enterprise ETL Migration: Seamless Transition from On-Premise Cloudera to AWS Cloud

    Responsibilities

    06/2023 – presentKolkata, India
  • Developed an alert job which will send an email requesting to stop the glue job which exceeds job execution time period, if exceeded time limits, with automatic termination after 20 minutes if necessary.
  • Participated in agile ceremonies and maintained the sanity of the sprint task/activities
  • Optimized the performance of a long running job therefore reducing the cost by 80%.
  • Conducted KT sessions with the production team, ensuring proficient monitoring
  • Conducted an analysis of the publication code in the current on-premise system, which is being migrated to AWS and developed generic functions for all XML elements and generated XML documents using PySpark and additional Python modules.
  • Successfully migrated an on-premise application to AWS Production environment.
  • Programmer Analyst Trainee, Cognizant

    Responsibilities

  • Developed robust & configurable data ingestion pipelines for multiple on-premise source systems, including SQL Server and REST APIs.
  • 06/2022 – 06/2023Kolkata, India
  • Implemented AWS services such as Apache Hudi, DynamoDB, AWS Glue, and Lambda for seamless data processing and storage.
  • Built dynamic DAGs (Directed Acyclic Graphs) in Airflow to automate and schedule data workflows.
  • Created reusable transformation jobs applicable across multiple systems, ensuring scalability and efficiency.
  • Integrated audit functionality in target tables using DynamoDB to maintain data integrity and traceability.
  • Developed advanced features like partitioning, bulk insert, upsert, and delete operations in Hudi for optimized data management.
  • Intern, Cognizant
  • Successfully completed the Big Data training.
  • Worked on a Covid 19 Data Analytics Project where performed the data cleanup and data transformation using Pyspark and represent the data using DataBricks.
  • 02/2022 – 05/2022Kolkata, India
    Education
    BTECH in Computer Science, Maulana Abul Kalam Azad University of Technology

    CGPA - 9.09

    08/2018 – 06/2022
    Publications
    Smart Attendance Tracking and Performance Monitoring System, IEEE

    Designed and Developed a system that will keep track of the attendance of an individual and also can monitor the performance.

    11/02/2023
    Certificates
    Awards & achievements
    Bronze Award, Cognizant
    15/08/2023
    Module Lead, Cognizant
    10/08/2023
    Silver Award, Cognizant
    24/02/2023
    Always Striving, Never Settling, Cognizant
    14/06/2023
    Runner up in HACKMOL 2.0, NIT JALANDHAR
    09/01/2021
    5 Star In Python and SQL, HackerRank
    02/10/2020