
Detail-oriented Data Engineer / Software Engineer with 6 years of experience in designing and implementing data-driven solutions across the full Software Development Life Cycle (SDLC).
Proficient in SQL, data modeling, and data pipeline development, with hands-on experience in data integration, ETL, and interoperability frameworks.
Adept at converting complex business requirements into scalable, high-performance data architectures that improve system efficiency and reliability.
Developed comprehensive data architecture documentation, including physical, logical, and conceptual models, to optimize data flow, enhance system performance, and ensure data integrity across enterprise platforms.
Developed and executed data movement operational plans to ensure timely, accurate, and secure data transfers.
Design and develop data pipelines to extract, transform, and load (ETL,ELT) data from source systems into the data lake.
Designed and implemented data models based on the star schema to support efficient data warehousing and analytics.
Managed and developed data batch flow processing to automate data ingestion and transformation.
Implemented and maintained data pipelines on Cloudera Data Platform (CDP) using Hadoop, Spark, and Hive for large-scale data integration and analytics.
Tools: Informatica developer,SQL Developer Data Modeler,Toad for Oracle.
Developed NPS forecasting and CX (CSAT, CES) models using Python and Excel to provide actionable insights for improving customer satisfaction and retention
Developed a predictive analysis model using linear regression to identify and categorize high-risk customers based on historical payment data.
Utilized SQL to retrieve and analyze business data, providing actionable insights for stakeholders.
Analyzed business needs and developed ad hoc reports to provide actionable insights.
Tools:Pycharm,Power query,Microsoft SQL Server
Design data integration technical solution
Build and develop ETL/ELT pipeline for data integration across multiple source.
Monitored and managed batch processing jobs, ensuring successful completion and resolving failures within defined SLAs.
Managed data warehouse staging and storage areas to ensure efficient data flow and integrity.
Conducted data migration activities between systems and databases.
Investigated and resolved data pipeline performance bottlenecks.
Experienced in SQL, including DDL (Data Definition Language) and DML (Data Manipulation Language) operations.
tools: Toad for Oracle,Informatica PowerCenter
Implemented machine learning techniques, including Naive Bayes Algorithm, for text classification to detect spam emails using the ,UCI Spambase dataset using python.
Established and maintained Key Risk Indicators (KRIs), IT Key Performance Indicator (KPI) and data quality reports using excel to monitor performance and operational risk.
Oversaw governance and quality review of Change Requests CR to maintain compliance with IT change management standards.
Tools: Jupyter Notebook,Share Point,Power BI
CDMP - Certified Data Management Professionals
DAMA International
The Path to Insights: Data Models and Pipelines
Advanced Data Analytics Certificate