Summary
Overview
Work History
Education
Skills
Certification
Timeline
Generic
Naga Venkata Lakshmi Baruvu

Naga Venkata Lakshmi Baruvu

Sydney,NSW

Summary

Insightful Data Modeler recognized for high productivity and efficient task completion. Specialize in developing complex data models, optimizing database solutions, and implementing data warehousing techniques. Excel in analytical thinking, problem-solving, and communication skills, ensuring projects meet both technical specifications and business objectives.

Overview

17
17
years of professional experience
1
1
Certification

Work History

Senior Data Modeler | Data Engineer

University of Sydney
Sydney, NSW
01.2024 - Current

Project Description: My Research Sydney

This project is to deal with Human and Animal Ethics application process flow which was dealt by university of Sydney staff and executives and it will be approved by Ethics office team for further steps of processing. Initially the entire process was setup in legacy system called IRMA & It’s been migrated to Omnistar(F1) system.

Modelling Responsibilities:

· Understanding data entities structure in both legacy and new Omnistar(F1 Solutions)Platform.

· Analysed source system views to understand the relationship of the entities for 400 + tables for Omnistar(F1 Solutions) Source system and translated it ER diagram in Oracle Sql Developer modelling tool.

· Implemented various data profiling techniques to understand data relationship in old and new systems.

· Interacting with internal USYD business stakeholders and external vendors for new source systems. Documenting Business information into technical data points.

· Designed Conceptual & Logical Modelling for target snowflake warehouse.

· Translated them into physical diagram in Oracle Sql modelling tool with Dimensional Modelling Techniques and developed the data mapping document rules for target tables.

· Exporting the DDLs from physical design model and publishing them into confluence for data engineer’s use

· Day to Day Support to Data engineers to for building the warehouse and helping them to write the complex queries.

· Followed Industry Naming Conventions across all the models by using Oracle Sql developer tool facility with Import Naming Standards CSV file

· Preparing detailed data mapping sheet from source tables to destination table columns by giving detailed mapping rules for derived columns and published them along with DDL scripts in confluence.

· Created TDA (Transformed Data Analysis) views for reporting requirements on top of the new Data mart dimensional tables.

· Transferred the Data Modelling Diagram from Oracle Sql Developer tool to Erwin Modelling tool through reverse engineering (USYD procured more licenses from Quest).

· Produced. Erwin & .EMF & high-quality resolution of .pdf files & publishing them into Confluence.

· Showcasing the entire end to end model with a wider audience including product owner, project manager and data engineer team along with scrum master.

Engineering Responsibilities:

As part of the Data migration process from old system IRMA to new system F1 solutions. Below are the tasks involved.

Historical Data Migration :

· Closely involved with business stake holders to understand the clear business requirements and migration templates for historical movement of the data.

· Built the Transformation logic in coalesce tool by sourcing snowflake source data for old IRMA system.

· Created the snowflake staging views as per the new business requirements and executed them using coalesce migration tool and pushed the staging views in to snowflake.

· Built the talend studio jobs by sourcing newly created staging views from snowflake and transformed them in to excel templates by following new source system (F1) standards.

· Created the control tables framework in snowflake for dynamic templates in future.

· Built the talend job to push the historical migration template files to shared point location (This location accessed by new Source System only).

· Published the talend studio job in TMC and created task to run on one – off adhoc basis.

On- going Data Migration :

· Built the talend job to source the data from new source system F1 and landing them snowflake staging schema.

· Creating an intermediatory schema in snowflake to implement CDC framework for the source by snowflake streams and tasks.

· And process the data from snowflake Intermediatory schema to data mart schema to accommodate SCD-2 implementation.

· Thorough testing with UAT data to perform CDC and SCD frameworks. And pushing the code in to GIT repo.

· Created Talend Plans in TMC and schedule the jobs on daily basis to flow data from F1 source to USYD snowflake target.

· Monitoring and maintaining the production jobs.

Senior Data Modeler | Data Engineer

Department of Customer Service
Sydney, NSW
09.2023 - 01.2024

Project Description: Night Time Economy (NTE)

The night time economy includes all the commercial activity that takes place outside the traditional business day, across the evening (6pm-9pm), night time (9pm-2am), and late night (2am-6am). The range of industries involved in the Night Time Economy is diverse and broad, including core functions such as live music venues, restaurants, and bars; as well as supporting businesses including transport, retail and accommodation services.

The Data Analytics Centre (DAC) is working with the Office of 24-Hour Economy Commissioner (O24HEC) to deliver NTE Insights. This will help O24HEC, NSW Government Agencies and Local Government Agencies to understand the dynamics of Greater Sydney’s NTE, as well as plan, design, track and measure NTE investments, policies and services.

As part of this project, Will be dealing with many government agencies to procure the data and process them at DCS,

Here are the list of Data Providers:

· Spatial Data to know Statistical Boundaries (SA1, SA2,SA3),Suburb and LGA Boundaries from ABS and NSW Spatial Services for the purpose of Geolocation Identification.

· Expenditure Details from Westpac for the Economic Activity.

· Business Composition ABR(Australia Business Register) and CBR( Custom Business Register Data )from ABR and Illion ( Commercial Risk Data) Providers.

· Licensed Venues data with locations and descriptive information from NSW Liquor and Gaming to know business operations.

· Movement Data (Mobile Phone Tower/Ping data) from Optus DSpark

· Transport Data (Public Transport Timetable, Mode Choices, General Transit Feed Specification and Travel Zones Data) from TfNSW.

· Employment Data from ABS

· Crime data from BOCSAR (NSW Bureau of Crime Statistics and Research)

· Weather Data from BOM (Bureau of Meteorology).

Modelling Responsibilities:

· Understanding the data and profiling on various sources data from various stakeholders like Westpac, TfNSW, BOM & LNG (Liquor and Gaming)

· As part of the Initial modelling Step, Identified the main entities and their relationships for modelling conceptual design by using Oracle Sql Developer Modelling Tool

· Enhancing Conceptual Model to Logical Model in terms of making understanding of business stakeholders and SMEs. And finally translating them into physical design and publishing

· Exporting the DDLs from physical design model and publishing them into confluence for developer’s use

· Followed Industry Naming Conventions across all the models by using Oracle Sql developer tool facility with Import Naming Standards CSV file

· Preparing detailed data mapping sheet from source tables to destination table columns by giving detailed mapping rules for derived columns and published them along with DDL scripts in confluence.

· Showcasing the entire end to end model with a wider audience including product owner, project manager and data engineer team along with scrum master.

Engineering Responsibilities:

· Working with creating different streaming delta live tables in azure databricks for numerous amounts of files from different cloud stage platforms (ADLS & AWS S3)

· Creating the simplified and optimized sql scripts for complex business requirements.

· Configuring the tasks and creating triggered pipelines in azure databricks workflow environment.

· Monitoring the production pipelines and Immediate action on fixing of any production pipelines.

· Creating necessary Change requests for production changes, presenting change to weekly change board meetings and deploying them into production.

· Working with complicated snowflake views to improve performance of the queries.

Principle Consultant Data Modeler

DXC
Sydney, NSW
04.2023 - 09.2023

URBIS:

Description : Urbis is an organisation dedicated to shaping cities and communities for a better future.

We work to create places and spaces to improve the lives of the people who use them.

Project Responsibilities :

● It was short term engagement ( 3 months) with Urbis to address their data issues and chaos.

● I worked with all the following business units stakeholders from Urbis to understand their day to day activities and their key operations.

● Understands their data sources and what kind of data sets is needed for their operational purposes.

● Delivered successful data taxonomy for all the below business units from Level0 to Level 2 ( L0 – Business Unit names , L1 - Major Data sources information , L2- subset of information)

● Delivered all the levels from L0 to L4 for sample property data set. ( L3- TableLevel Information, L4- Attributes level information)

● Delivered successful sample for data dictionary documents for property WH.

● Working with Urbis Data warehouse ( Snowflake ) and their Transformation Logic by exploring their Talend Orchestration Jobs.

● Delivered the Conceptual , Logical and Physical Modeling for Property Dataset by using Sparx EA.

● Business units Covered:

o Economics ad Property

o Planning

o Heritage

o Transport/Traffic Management

o HR ( People and Culture)

o Marketing

o Economics and Social Advisory

o Stakeholder Engagement

o Finance

o Future State

Western Sydney International Airport :

Project Description : Western Sydney Airport is a new international airport currently under construction within the suburb of Badgerys Creek, New South Wales, Australia. The airport is planned to have 24-hour and curfew-free operations and will supplement Kingsford Smith Airport. And the first stage is expected complete and open by December 2026.

Responsibilities: AS WSI requires data from 67 different source systems which was maintained by different vendors. DXC is responsible to get the requirements from them do the modelling for conceptual and logical.

I’m responsible for developing the following system models

1) AOS (Airport Operating System)

2) AMS (Asset Management System)

3) PMS (Parking Management System)

Senior Data Engineer | Data Modeler

NAB
Sydney, NSW
04.2022 - 04.2023

Project Description: CAS (Credit Appetite Strategy) Reporting, an Initiative within the APRA Portfolio Management program, is to provide a CAS metrics exception monitoring and reporting solution both on deal and aggregated portfolio level. CAS metrics are defined to help manage credit risk in an organisation. CAS metrics thresholds, which are used in credit decisioning, are currently set on an annual basis. These thresholds are used as basis for CAS exception reporting whereby any CAS metrics exceeding the corresponding thresholds will be monitored and reported.

Responsibilities:

● Proactively Involved with various business stakeholders on requirement gathering for CAS Project

● Converting the Business information to Data points and documenting them in confluence

● Deep dive into Data Profiling from source system for different databases like Redshift,SqlServer & DB2, Informix . Since the data is coming from multiple databases as this CAS project involving the data from extracting 6 different source systems like ME , ECRS, DRM, Siebel, DST and IAP CRE and 5 different Flat files data provided by business stakeholders for ANZSIC industry classification and threshold values set for CAS measures like EBITDA,ADJ EBITDA, RATE CODE & LVR & ANZSIC 1993 and 2006 Hierarchies and rules defined to calculate CAS outcome variable.

● As part of the Initial modelling Step, Identified the main entities and their relationships for modelling conceptual design by using Oracle Sql Developer Modelling Tool

● Enhancing Conceptual Model to Logical Model in terms of making understanding of business stakeholders and SMEs. And finally translating them into physical design and publishing

● Exporting the DDLs from physical design model and publishing them into confluence for developer’s use

● Followed Industry Naming Conventions across all the models by using Oracle Sql developer tool facility with Import Naming Standards CSV file

● Converted the whole model by using DDL scripts from Oracle to Erwin to follow NAB standards.

● Preparing detailed data mapping sheet from source tables to destination table columns by giving detailed mapping rules for derived columns and published them along with DDL scripts in confluence.

● Showcasing the entire end to end model with a wider audience including product owner, project manager and data engineer team along with scrum master.

● Supporting Data Engineer Team to clarify all the modelling and mapping related clarifications and giving helping hand to deeply involve critical LNK and satellite tables development work to make sure to pass hard rules for referential integrity checks.

● Continues support to offshore development team and testing team as well.

● Successfully implemented the POC for Ad hoc file ingestion into snowflake with different delimiters.

● Strong initiative, innovative thinking skills, and the ability to analyse details and adopt a big-picture view

● Ability to communicate and discuss software components in simple, general terms with business partners and in detail with software development engineers

● Strong analytical skills with the ability to collect, organize and analyse large amounts of information with attention to detail and accuracy

● Ability to write high quality, maintainable, and robust code, often in Snowflake, DBT Scripts.

● Strong data modelling skills and excellent knowledge in Data Vault 2.0

● Building Splunk Dashboards for notifications on critical alerts

● Good understanding and experience on data bricks notebooks.

Data Automation Lead

HCF
Sydney, NSW
02.2018 - 04.2022

Project Description: HCF is one of the Largest Insurance Company in Australia, It provides Health & Life Insurance services to many customers. Massive System Transformation program happened from legacy system to cloud environment with latest technologies to provide more robust solutions and serve our customers in a better way.

Responsibilities:

● I have been involved in both modelling and engineering tasks in HCF .

Modelling Responsibilities:

● Collaborated with stake holders to gather requirements and documenting them in confluence.

● Involved in modeling tasks for Ebix Life system using Data Vault 2.0 methodology

● As part of the System Transformation Project, Translation of the model from Dimensional to Data Vault Model by using Oracle Sql Developer Tool

● Exporting the DDLs , Published and followed HCF naming standards.

● Very Strong Experience and handling different Business Hubs, links and multi active and hierarchical satellite tables

● Designed various PIT & Bridge Tables at Business vault and Access control layer according to the business requirements.

● Certified in Data Vault 2.0.

● Good Experience in Kimball, Inmon methodologies and Dimensional Modelling as well.

● Align data model with enterprise data models and agreed standards.

● Implementation of Data Architecture deliverables such Business & technical metadata,

data security & privacy catalogue, data quality assessments.

● Supporting data quality initiatives. Performed activities such as data profiling, define &

manage business rules for standardisation and cleansing.

Engineering Responsibilities:

● Analysed legacy existing Data Stage & SSIS Jobs to understand existing requirements and Implementation and document them for critical reports.

● Involved in developing new SSIS packages for “HCF Thank You” Loyalty project for providing extract files for card printing for 4 different Tiers (Opal, Emerald, Ruby and Diamond).

● Supporting the BAU activities for Enhancements for HCF thank you loyalty Program

● Very strong experience Data Vault 2.0 Methodology by using Oracle Sql developer tool only in logical and physical designs for entire life system data which is resided in EBIX One Office Source System ( Sql Server)

● Proactively Involved in developing Data Integration (DI) control framework metadata to ingest data from different source systems residing in different servers like health policies in Informix server, Life policies data in EBIX system and claims and ancillary data in sql server and many more .

● Strong Advanced SQL coding skills to provide optimized query in snowflake access layers to make use of Tableau Reporting Team.

● Strong Experience in working with creating talend jobs for orchestration and source versioning and publishing them in to talend cloud.

● Strong experience and debugging skills for talend jobs for the failures of Rawvault and Business Vault Orchestration Jobs and modifying the jobs to fix those issues and publishing them in talend cloud.

● Developed various talend outbound export jobs ( Excel, CSV formats) for different business requirements like cost calculator , health share and flybuys projects.

Senior Software Engineer

Crimson Logic India Pvt Ltd
Bangalore, Karnataka
12.2015 - 06.2017

Project Description: Crimson Logic’s eJudiciary solution is able to eradicate problems (i.e. from case management to delays in the delivery of judgment) that plagued traditional judiciaries.

Allows court documents to be filed, served and extracted electronically. Legal Professionals enjoy easy and convenient submission of court documents and online access to activate case files using our intuitive electronic filing (eFiling), while judicial officers can enjoy improved efficiency and effectiveness in handling court cases and schedules with our court case management system.

Responsibilities:

As a Team Leader, I am involved in the development of following modules and responsibilities towards them.

Backend Modules

**************

Casefile

Case Register

SICC

Operational Reports

Pending Cases Reports within 12 Months and Over 12 Months

Assigned Judge and Case Type Reports

• Implemented the Critical CRs and DB modification user details based on Information Provided by Singapore Colleagues

• Writing Stored Procedures and Functions at DB Level

• Work Item and Snapshot Creation in Report Server Manger

• Designing the Reports in Power BI Environment and Customizing they based on the requirements provided by the client.

• Creating / Sharing the Power BI Dashboards

• Developing Pie Chart, Stackbar and Tabular Reports in SSRS

• Creating and Editing Stored Procedures and Functions for use in the SSRS Reports

• Involved in Logical and Physical Data Modelling Sections.

• Unit Test Plan Preparation and Independent Unit Testing

• Taking ownership of full deployment activities in live application and DB servers

Senior Software Engineer

Wipro Technologies
Bangalore, Karnataka
05.2010 - 01.2013

MS Applications, Cable & Wireless Account (UK)

(Jun 2012 – Jan 2013)

Environment: C#.Net, ASP.Net, SqlServer 2012, SSIS

Description: MS application includes two applications one is Freeze Escalation and another is Service Catalogue

Freeze Escalation: Freeze is a critical period of the year during which Cable & Wireless will stop all delivery activities in a phased and controlled schedule to protect customer’s services. Freeze Escalation application provides online form to raise and approve the freeze ticket for implementing change requests during the freeze. Freeze Team can track all the tickets and can approve or reject the same. Freeze requester will be able to track the status of the ticket raised by him.

Service Catalogue: Objective of this project is to provide a web based tool to store and manage the IT Service Catalogue which will be accessible to all colleagues over the C&W intranet.

Responsibilities:

As a Team Leader, I involved in the following activities:

• Implemented the Critical CRs and DB modification user details based on Information Provided by UK Colleagues

• Writing Triggers and Stored Procedures at DB Level

• Developing SSIS Package for ETL Transactions

• Creating and Editing Stored Procedures for use in the SSIS Packages

• Unit Test Plan Preparation and Independent Unit Testing

• Taking ownership of full deployment activities in live application servers

NB Mailer Application, Wipro Technologies

(March 2012 – May 2012)

Environment: C#.Net, ASP.Net, SQL Server 2012, SSIS, SSRS and Tableau

Description: NB Mailer Application is a Web based Application. The main purpose of this application is to send the e-mails to the Wipro Employees who are non-billed for more than some number of specific days based on the information provided by the Business.

The Content of the Mail will be varied based on the Number of Non billing days of the Individual Employee. And Admin can choose Exclusion elements in admin tab. So, that before sending mails to the employee's system checks for those exclusion elements.

Apart from the Exclusions, admin can do some extra tasks like Setting the NB Mail Frequencies, Editing the Mail Content and Subject Info, Giving access to the Other User's for this app and Uploading the excel data to the oracle table.

This Application contains Mail Report on Date basis. It is used as logged purpose. This report contains the information of the Wipro employees who got the NB mails from this application and contains Export to Excel feature for this report.

Responsibilities:

• Requirement Analysis and Designing the Pages and Database Design

• Data Importing by Using SSIS and Developed SSIS Packages for Implementing ETL Logics

• Generating Reports through SSRS and Visualizing in Tableau

• Coding the Application

• Unit testing of above tasks.

• Prepared Use case Document and Traceability Matrix document

• Build and deployment of the entire application and Conducting Code reviews & Knowledge transfer sessions.

IRIS (Interactive Research Information Systems), Capital One, US. ( WIPRO’s Client)

(August 2010 – December 2011)

Environment: C#.Net, ASP.Net, SSRS, Tableau

Description

IRIS is a Web Application, This Application is used by a call centre agents for the Capital One Bank. This Application mainly serves the Capital One Customer related Tasks or Issues. If Customer has any queries related to their Credit Card Transactions or Personal Data, They will call to Capital One Customer Care Centres. So, the Agents in Customer Cares will use this web application based on the basic input data provided by the Customers.

Responsibilities:

I have involved in the following workflows

• Credit Limit Decrease

• Credit limit Increase

• Request Statement Copy

• Change address and phone

As a Module Lead, I involved in the following activities:

• Designed the Whole Workflow from scratch and Developed the Workflow as per RPD(Requirement Package Document)

• Implemented Design Patterns

• Designing the reports through SSRS

• Visualizing the reports in Tableau

• Writing ATCs (Automated Test Cases) for Presenter and Controller Files

• Running Code Analysis Once code is completed and fixed all code related issues

Software Engineer

Mahindra Satyam
Chennai, Tamilnadu
10.2007 - 05.2010

I have worked for following multiple projects for the Client called Chevron (Oil and Gas Domain)

(I) OE (Operational Excellence) Forum 2010 - Nov 09 to May 10- (C#.Net, Sql Server, SSRS, Followed Agile - SCRUM)

Desc: This Project OE Forum will be happening every year in the month of October at Sinnamon and Houston Places. It explains about Some Users or Employees in the associates in Chevron are going to submit Presentation kind of thing called abstracts through this site across different categories like Safety, Health, Environment, and Leadeship Accountability. All these Abstracts will be scrutinized by the admin people and then they will send mails to the Users whose abstracts are selected. These Users should go to their respective places to present their abstracts.

(II)DPAR (Data Privacy, Assessment and Remediation) – Dec 08 to May 10- (C#.Net, Sql Server, SSRS, Followed Agile - SCRUM)

Desc: This Project Provides the Privacy for the data (It Maintains different types of data like Sensitive Personal Data, Inventory Data).Assessment phase conducts the audit for the data process. If any Risks are there in that data, it maintains all the risks in the finding phase. For each and every risk there is remedy, and all the remediations are maintained in remediation’s phase. And more over this project generates different types of Reports. All these Reports are generated by using Sql Server.

(III)EIM (Early Injury Management) - Apr 09 to May 10 (Jasper Reports Technology)

Description :

Early Injury Management (EIM) process supports their Global Downstream employees by providing systematic medical management of minor work-related injuries. It requires a global, robust, system which allows for logging and managing all of their process work flows, from within one source

• (IV)GCAS (Global Consulting Application Solutions)-Sep 08 to Dec 08 (Cold Fusion, Ms Access, Adobe Dreamweaver CS4)

Description:

GCAS Site Contains the Team members involved in GCAS Unit, Personal and Professional Details of Team Members, Various Projects Handled in GCAS,Document Library (Different Documents that are used for GCAS Team Members),Value Adds, Best Practices, Documents Related to Knowledge Sharing, Learning Materials. And all other information related to GCAS Team Members.

Responsibilities:

• Worked for Report Generation. And Writing SQL Stored Procedures that are required for reports and Designing the .rdl files

• Deploying the .rdl files in Report Server ad Integrating the .rdl files with .aspx pages

• Generating .Jasper files using Tomcat Web Server

• Involved in development and Unit Testing of above tasks.

• Prepared Use case Document.

• Prepared Document for Test Plan and Test Cases.

• Building and deployment of the entire application

Education

Master of Science - Artificial Intelligence And Machine Learning

University of Texas At Austin
United States
10-2024

Bachelor of Science - Computer Science And Engineering

Jawaharlal Nehru Technological University
India
05-2007

Skills

  • Data Vault 20 Modelling
  • Dimensional Modelling
  • Relational Modelling
  • Modelling Tools are Erwin, Oracle Sql developer, Sparx EA and TOAD Data Modeler
  • End to End building of data warehouse
  • ELT / ETL methodologies
  • Talend
  • Data bricks
  • SSIS
  • AWS
  • Azure Delta Lake
  • GIT HUB
  • Tableau
  • Power BI
  • SSRS
  • Snowflake
  • SQL/Oracle/Postgres/Informix/AWS Redshift dbs

Certification

  • Databricks Certified Data Engineer – Associate
  • CDVP2 – Certified Data Vault 2.0 Practitioner
  • Splunk Core Certified User
  • Splunk Core Certified Power User

Timeline

Senior Data Modeler | Data Engineer

University of Sydney
01.2024 - Current

Senior Data Modeler | Data Engineer

Department of Customer Service
09.2023 - 01.2024

Principle Consultant Data Modeler

DXC
04.2023 - 09.2023

Senior Data Engineer | Data Modeler

NAB
04.2022 - 04.2023

Data Automation Lead

HCF
02.2018 - 04.2022

Senior Software Engineer

Crimson Logic India Pvt Ltd
12.2015 - 06.2017

Senior Software Engineer

Wipro Technologies
05.2010 - 01.2013

Software Engineer

Mahindra Satyam
10.2007 - 05.2010

Master of Science - Artificial Intelligence And Machine Learning

University of Texas At Austin

Bachelor of Science - Computer Science And Engineering

Jawaharlal Nehru Technological University
Naga Venkata Lakshmi Baruvu