Energetic committed and motivated individual with a passion for innovation, learning and technology.
8+ years of diverse industry experience in the Cloud, Data warehouse and complete Software development lifecycle with proven track record in planning, building, and managing projects in the retail and health insurance sector.
Experience in AWS resources including EMR, EC2, S3 and Lambda
Expertise in ETL tool Talend, Informatica Power Center Client Suite (9.1.0).
Hands-on experience on Snowflake, Postgresql, Teradata 13.10, Oracle10g, PostgreSQL and database tools such as Teradata SQL Assistant, SQL Developer, Teradata Viewpoint and Teradata utilities such as Multiload, Fastload, Fastexport and BTEQ.
Proficient in Python and Java programming.
Knowledge of Informatica Data Quality and Informatica Analyst.
Well acquainted with waterfall and agile development methodologies.
Experience in creating complex dashboards using Microstrategy as a reporting tool.
Skilled and experienced in UNIX Shell Scripting.
Strong analytical, problem solving and communication skills with ability to work as part of a team as well as independently.
Overview
11
11
years of professional experience
Work History
Lead Data Engineer/Data Architect
HCF
Sydney, NSW
03.2022 - Current
Drive deep architectural discussions to ensure solutions are successfully and securely deployed on the cloud.
Designed and Implemented an ELT-based Data Lake and Data Warehouse/Database solution for the big Australian Insurance client.
Implemented a robust metadata-based ingestion framework using Talend Data Fabric and Postgres SQL to orchestrate the data ingestion to Data Lake in Amazon S3 and Snowflake Cloud Warehouse from the existing source systems
Analyze the existing Voice of Customer Application and migrated the entire solution from SPSS to the warehouse, whilst helping the stakeholders in moving the upstream applications to integrate with the new landscape.
Analyzed the development life cycle process, identified the pain points, and laid down the CICD strategy for Datawarehouse deployments using AWS Code Commit, Flyway Docker,AWS Code builder.
Analyzing the current machine learning/Data Science strategy by doing value chain analysis for the department and devise an analytical strategy.
Performance optimization of complex Snowflake SQL to reduce the query run time and appropriate consideration of Snowflake costing parameters to reduce the overall billing cost.
Demonstrated leadership by making improvements to work processes and helping to train others.
Data Engineer
Macquarie Group
Sydney, NSW
10.2021 - 03.2022
Implemented a robust metadata based reference data framework and automation jobs using Talend Data Fabric and MYSQL to orchestrate the data ingestion process to Data Lake in Amazon S3.
Created Scripts to manage inbound as well as out bound files for multiple projects.
Design and Developed big data jobs in talend to load gb's of file into S3.
As part of framework team, created multiple data pipelines to enable the business with their process into Data Harbour.
Creation of orchestration using Airflow and managing the runs through SQL based framework.
Conducted several Knowledge Transfer sessions for the business teams to start with the framework.
Data Consultant
Servian
Sydney, NSW
08.2019 - 10.2021
Analyzed, provided consultation and implemented a robust metadata based AWS cloud-based data integration framework and automation jobs using Talend Data Fabric and Postgres SQL to orchestrate the data ingestion process to Data Lake in Amazon S3 and Snowflake Cloud Warehouse from the existing source systems.
Defined database design layers in RDV100 and RDV200 on specifications based for performing CDC and data fetch.
Cloud infrastructure set up in AWS- VPC, subnets, VPC end points, private link etc
Analyzed the existing data and business scenarios (Nobel Campaign and IVR call routing) and developed Talend Jobs while using advanced SQL in Postgres, to be further used by Business Intelligence teams and Customers.
Identify the current technology and landscape limitations and provide guidance and expertise to architects to build a data strategy design of future customer application and implement it.
Created various Talend job templates to enable data pipelining from various systems such as New health, Ebix etc
Testing of Transformed view layers based on business requirement.
Performance optimization of complex SQL to reduce the query run time in Postgres and Snowflake.
Oversaw various types of data warehouse testing by devising plans, scripts and test files.
Identification of PII and NPII data and appropriately masking the data based on user roles.
Analysed the existing data and business scenarios and developed Semantic layers using advanced SQL in Snowflake, to be further used by Business Intelligence teams for data insights and reporting purposes.
TechLauncher – ANU
02.2018 - 11.2018
Worked on ANU TechLauncher project as a developer to automate the existing CHIIA workflow structure.
Involved in gathering requirements, coordination and monitoring to ensure the timely flow and complete
delivery of the project.
Enhanced technical capabilities by acquiring new skills like Heroku and Sqlalchemy.
Involved in data modelling and database development.
Worked on python backend to support the front end utility.
Bug Fixing and resolving technical issues arising during testing phase.
Creating wireframes and front-end build of software.
Using Git for code promotions and version maintenance.
Software Development Senior Analyst
Accenture – Mumbai
03.2013 - 01.2017
In Accenture for a renowned US based retail client
supporting several facets of their business, helping them derive insights from data by writing complex SQL
queries on Teradata.
Extensively worked on ETL components in Informatica, Analysis and Issue fixing, Unit Testing of the
developed components, deployment and support.
Designing and Developing Error Framework to improve data quality and address critical data issues.
Design, Build, Test and Deployment of mappings,session and workflow.
Development of High Level and Low Level Design documents.
Development of ETL components for Source to Stage and Stage to Target level.
Source to Stage level
components included implementation of insert only logic and Stage to Target level components included
implementation of SCD type 1, type 2 and insert only logic.
Extensively worked on Informatica Power Center Transformations such as Source Qualifier, Lookup, Filter,
Expression, Router, Joiner, Update Strategy, Rank, Aggregator, Sorter and Sequence Generator.
Development and design for the mappings of SLT,Dimensions and Facts.
Creating Profiles in IDQ/ Informatica Analyst and scorecard on top of that and scheduling.
Worked on IDQ transformations like Parser, Standardizer, rule and filter creation.
Developing Microstrategy Objects (Schema and Public) and creation of report Document, Intelligence Cube
and Template on top of that.
Creation of BTEQ scripts to fix and build history data for the EDW tables.
Creation of Fastload,MultiLoad scripts for feeding large volume data into EDW and FastExport script for
generating large volume extracts.
Used Pushdown Optimization to fix the long running processes on the ETL server and minimized the load on
the Integration Service.
Used other Teradata concepts like Collect Statistics, Partitioned Primary Indexes,
and Join indexes to optimize the performance.
Experience in writing complex SQL queries for building SLT’s, SCD Type 1 and Type 2 mappings.
Experience in Unix scripting to implement the archival and purging process of files, handling MFT of files,
splitting of files.
Creation of Unix script to PGP and AES encrypt the source and target files.
Exposure to tools like StarTeam, SourceTree, Github, Jenkins and Anthill for code migration and Unix file
migration.
Experience in working on Mainframe scheduling tool ESP for job and event creation, job status monitoring,
jobdoc and joblib creation.
Performed Unit Testing and involved in Production Support.
EDW-ETL Tools- Talend, Informatica Power Center Client Suite (910):Mapping,Designer,Workflow Manager/Monitor,Repository Manager, DT Studio, Informatica Analyst;IDQ