Chiranjib - Updated Resume
Chiranjib - Updated Resume
Chiranjib - Updated Resume
As a Data & Analytics Cloud data Solutions Architect/Senior Data Engineer, I have over 17 years of
experience in data projects, business intelligence, cloud migration, AI/ML projects in the Healthcare and
Retail sectors. Have extensive experience with data insights of HC Clinical analytics, CPG Supply chain &
Sales domain. Holding GCP and AWS certifications and have worked onsite in the USA and Australia.
PROFESSIONAL SUMMARY
Extensive experience with working as cloud data migration, architect/senior consultant from on
premises data warehouse into different cloud platforms
Develop comprehensive migration strategy, including the use of Lift and Shift and Transformation
methods, to ensure a seamless transition to cloud-based platforms like aws/gcp/azure
Reverse engineering of the existing systems like on premises databases and Hadoop systems and build
strategy to migrate historical and on -going (delta) data into new cloud platform.
Extensive experience with design and implementation of GCP cloud migration using GCP native
services like dataproc, cloud spanner, app engine, BigQuery , Compute engine and devops services
like GKE, Composer, Terraform, Jenkins
Extensive experience on design and implementation Azure cloud solutions using medallion
architecture (bronze-silver-gold) pattern using Databricks, Data factory, ADF, Synapse, ADLS Gen2,
blobs, Delta Lake
Experienced with design and implementation of AWS cloud solutions using Glue ETL,S3 (lake),
Lambda, Cloud functions, Redshift(warehouse) and scheduling through airflow, control M, Autosys
for batch processing
Experienced with working accelerator tools to migrate existing application codes to equivalent
Pyspark code for large scale application/data migration.
Extensive experienced defining IT Roadmap, cloud strategy, Enterprise/solution architecture,
Architecture assessment
Worked in Azure platform validation for the newly acquired Azure lake and Synapse
Expertise in DevOps, Release Engineering, Configuration Management, Cloud Infrastructure,
Automation. It includes ADF, Amazon Web Services (AWS), GCP(Google Cloud platform),Ant,
Maven, Jenkins, Chef, GitHub,
Worked with cloud security team to make compliance with all the security protocol used enterprise
wise. Worked in all kinds of cost optimization techniques used in cloud platforms
Extensive previous experience in building data warehouse on Teradata
Demonstrated experience on working with USA Healthcare Data & Analytics (major in Care
management area) , experienced handling Membership/Claims/Provider data with EDI formats like
834.274.837.HL7 etc
Demonstrated experience in CPG(Consumer product Goods) business model implementation in retail
projects, while migrating Master data, Sales data with domain granularization
Excellent in building strong relationships with clients and providing support to all stakeholders
including senior leadership, helping with different metric generation
Deft in providing guidance to business unit partners regarding issues or opportunities.
Wrote technical white paper on HealthCare (Clinical Analytics)
Successfully executing Cloud migration data projects in cloud platforms by transforming traditional
Data warehouse, whereas reducing 40% cost
Managing team of 15+ in onshore offshore delivery model , typically using Agile scrum method
SKILLS
Cloud architecture Data lake Applications Big Data Tools
Project Management Data Migration Technical Support
Software Development Data architecture Client Relationship
Product Development Business Development Team Management
EDUCATION
Bachelor of Computer Science & Technology from West Bengal University
of Technology, India, 2006
CERTIFICATIONS:
No Name of Professional Certificates Year of
Institution Completio
n
Google Cloud Google Cloud, Google Certified 2024
Professional Machine Learning Engineer
Certificate ID-uZ4kLh
Databricks Accredited Generative AI Fundamentals 2024
Projects Handled
Project Title CVS Health (Aetna) GCP cloud transformations
Duration Jul 2023-Current
Location Hartford CT ,USA
Role Cloud Solutions Architect/Senior Data Engineer
Technologies used GCP services (Big Query, Spanner, Dataproc, Cloud SQL,
Terraform, GKE, Jenkins etc) , Hadoop , Sql Server
databases, Autosys,
PROJECT DESCRIPTION:
CVS health (Aetna) is undergoing tech modernization while transforming on
premises systems into Google cloud. As part of migration 40+ applications need to
move in GCP BigQuery (BQ), while existing reporting systems will repoint to BQ for
KPI’s. Goal is to migrate the On premises databases and Hadoop hosted applications
into BQ in phase wise
Key Deliverables:
Actively involved in a program geared towards migrating existing mission and
business applications into a cloud-based environment. Activities required to
re-host an application into the cloud may include architecture modifications,
database and/or application server re-hosting, and potentially recoding of
existing capabilities to take advantage of cloud platform services.
Reverse engineering of 40+ applications from existing Hadoop system and
build roadmap for GCP- BQ migration. Planning for pilot phase data migration
for 3 applications
Collecting data sets that meets functional / non-functional business
requirements. Working with data and analytics experts to strive for greater
functionality in data system.
Guiding team of 20+ members on 3 tracks for reverse engineering, mapping
creation and data pipeline creation. Building strategy for historical data
migration (using existing data in Hadoop and SQL server) with minimum
changes to accumulate exiting KPI’s and new data pipelines for on – going data
ingestion
Creating new data pipelines from source to target for incremental ingestion
and existing data import in GCP BQ as historical load for KPI and analytics use
Worked in creating new chatbot using LLM, integrating ServiceNow and
Github for internal user use for faster response on knowlegebase
Working closely with customer Directors and VP’s assisting them for
continuous betterment of the program delivery
Creating Metrics/dashboards using Rally for all levels of burndowns for
successful migration
Precedence Experience:
Project Title PepsiCo Cloud Transformations
Duration Aug 2022- Jun 2023
Location Kolkata, India
Role Cloud Solutions Architect
Technologies used Azure Data factory, Synapse, Databricks,
Teradata,Unix
PROJECT DESCRIPTION:
‘PepsiCo Inc’ has IT strategic movement to modernize the platform from Legacy
systems to Cloud. ‘PepsiCo cloud transformations’ is the migration’ program from
Teradata to Azure Synapse platform. The project needs to migrate all the existing
Teradata analytical applications into Azure Synapse. Reporting tools will repoint to
Synapse for KPI’s. This project is planned by phases starting from pilot followed by
Phase 1 , Phase 2 , Phase 3.Objective is to onboard business users to provide benefits
of modern cloud technologies with cost affective robust solutions and finally retire
Teradata from PepsiCo landscape
Key Deliverables
Actively involved in a program geared towards migrating existing mission and
business applications into a cloud-based environment. Activities required to
re-host an application into the cloud may include architecture modifications,
database and/or application server re-hosting, and potentially recoding of
existing capabilities to take advantage of cloud platform services.
Worked as Azure architect/Data management techno-functional lead from
offshore
Approach finalization for historical and ongoing incremental data load
Worked in reverse engineering of the Teradata existing jobs to build the
inventory for app migration.
Worked in creating detailed level design document for pilot apps
Develop solutions for Pilot phase, to migrate ‘Power Of One’- Sales and
Marketing KPI application from Teradata to Azure Synapse, by converting
existing TD Bteq logics into ADB PySpark codes. Move Teradata ACQ layer
data using S&T framework into Bronze and then build the silver and gold layer
inside Delta lake. Finally move the data into Synapse using store procedures.
Worked on production go-live/cutover for pilot apps
Responsible for mentoring the team of 10+ members in offshore for seamless
delivery
Collecting data sets that meets functional / non-functional business
requirements
Working with data and analytics experts to strive for greater functionality in
data system
Prepare for future applications discovery and analysis for upcoming sprints
Following Agile scrum model to adhere to PepsiCo’s standard agile practices.
PROJECT DESCRIPTION:
‘Bayer Corp Science’ has IT strategic movement to modernize the platform from
Legacy systems to Cloud. ‘TD-GCP migration’ project is the platform modernization
from Teradata to Google BigQuery platform for CS (Corp science) division. The
project needs to migrate all the existing data to migrate in Google Cloud along with
that new data pipeline creation which will bring the data from SAP and Non- SAP
sources into GCP (Google cloud platform) on regular basis. Existing e applications and
reporting platforms will repoint from BigQuery. Teradata will be decommissioned
after the migration is completed
Key Deliverables:
Actively involved in a program geared towards migrating existing mission and
business applications into a cloud-based environment. Activities required to
re-host an application into the cloud may include architecture modifications,
database and/or application server re-hosting, and potentially recoding of
existing capabilities to take advantage of cloud platform services.
Create and maintain project related documentation (Statement of Work
Document, Proof of concept document, technical design document etc.)
Develop overall solutions including Statements of Work and Bills of Materials
Responsible for creating and maintaining optimal data pipeline architecture.
Create end to end design for applications to host on GCP platform.
Collecting data sets that meets functional / non-functional business
requirements.
Working with data and analytics experts to strive for greater functionality in
data system
Build data warehouse solutions on BQ platform, to make sure all the existing
applications perform as-is or better to meet client business SLA
Working with different API's on extraction of data
Managing team of 10+ members, in the POD model for project execution
Gathering customer requirements on regular basis and working on multiple
RFP’s to meet Bayer’s IT strategic and architectural goals
PROJECT DESCRIPTION:
Kellogg Company’s strategic movement towards migrating the on Premises databases
and Big data Hadoop into AWS cloud This project includes the movement of data
from on prem SQL server and Hadoop systems to re-platform into AWS cloud by
migrating the data and existing applications into AWS and decommissioning the
Hadoop and SQL server boxes
Key Deliverables
Strategist and AWS data engineer track lead on re-engineering of applications
to migrate from on premises and external file systems to AWS cloud services,
which involved data warehouse rebuilding over AWS which will be accessed
through all the reporting services and analytics team
Worked on classification of the existing tables in domain group (Master Data
and Sales data major) in the activity of re-engineering and grouping of the
existing tables for building domain granular model
Involved in Kellogg specific new data model activities which was restructuring
the tables on domain and region basis
Architecting, developing and prototyping of different frameworks to bring
data from on premises to cloud using Spark job (written in Python ) with AWS
Glue services for non-SAP sources and confluence file type
Create end to end design for applications to host on AWS which involved s3
RAW-> S3 CleansedProcessed->Redshift Target Db for domain creating
domain granular layer
Created STTM document for Cleansed to Redshift target tables using current
transformation logics which were in on premises sql
Worked extensively with Cognizant accelerator framework (DIF tool) to move
data raw to cleansed layer while developing all kind of file formats like
csv/txt/xls/xlsb/json/dat to accumulate in cleansed layer
Written Glue ETL scripts to move the data from S3 Raw to cleansed bucket
using data quality rules
Worked on moving data from cleansed to processed layer -> Redshift DW
layer to make the data available for reporting and analytics
Written transformations for target tables in Redshift using Glue ETL with
python programming language
Involved in developing DevOps activities for applications using Github.
Worked extensively on analysis of SAP global tables ECC/SCM (major)
modules for analysis and mapping for building domain granular data model
Successfully migrated Master data and Sales data in Domain Granular
model(Material/Customer/location /Foundation reference/Sales-
Execution/Strategy/Performance/Order etc) into AWS environment.
Working as advisory to all the different tracks for data related issue
fix/debugging in all around the tracks in Kellogg
Worked with AWS support for fixing different issues while executing the Glue
codes which were not able to execute due to VPC/subnet mask.
Managed a team of 15+ members in the Waterfall and POD model of
execution.
PROJECT DESCRIPTION:
Amgen is one of the largest life sciences company in the world. It manufactures medicines with
innovations and research. This is the EDW Teradata Admin project for Amgen. It has total 9
Teradata systems for data warehouse which includes EDW systems and Radar EDW systems.
Developed Bteq, Multiload, Fastload, TPT jobs in Teradata and using Shell scripts
Involved in writing and tuning of critical views in different layers of the business application
Identifying and tuning bad performing batch queries
Responsible for tuning all kind of performance issues handling(L3 level)in Teradata applications
PROJECT DESCRIPTION:
Walgreens is the one of the largest pharmaceutical retail in USA. Walgreens'
Datawahouse system is build in Teradata database. This is a warehouse for Walgreen’s
store data
Key Deliverables
Resolved data issues for reports such as missing data in the BO (business object)
report, failure analysis for Informatica job flows: Example: If there is any data missing
for any report we need to do RCA for the same such as analysis of query running in
backend/environmental issue etc.
Provided RCA for any regular job failure in Informatica, missing file in
servers/environmental issue/long running job issue etc.
PROJECT DESCRIPTION:
Westpac is 3rd largest bank in Australia. This project is retail banking reporting
system. Retail banking is related to personal loan, home loan, credit card etc. The
reporting system is basically related to scorecards generation regarding the retail data
of the bank. This includes the growth of both the retail sales of the bank and the
performance of the bank employees.
Key Deliverables
Write codes on JCL, Teradata, store procedure and Sql server as per required
enhancements
Solve L3 level production issues
Project Title ATT Data Analyst DBA
Duration Jan 2010 – Dec 2010
Location Pune, India
Technologies used Teradata, Unix
PROJECT DESCRIPTION:
AT&T is world’s one of the most well-known and biggest Telecom company. There
are 33 markets in USA in which their telecom services is going on. This project is
support and enhancement of existing and upcoming applications in data warehouse
Key Deliverables
Write codes on JCL, Teradata, store procedure and Sql server as per required
enhancements
Solve L3 level production issues
24/7 On-Call Teradata DBA support
L3 application support for 3 applications
PROJECT DESCRIPTION:
This is the project of Vodafone data warehouse. There are five source system from
which data is
coming. The total data is loaded in the data warehousing production box. The final
reporting’s are
done from this system. Total analytical process for the months are going here.
Key Deliverables:
Write codes on JCL, Teradata, store procedure and Sql server as per required
enhancements
Solve L3 level production issues
Accomplishments
Presented with Best performance recognition from Kellogg and PepsiCo clients for AWS & Azure
Delta Lake implementation
Completed the PepsiCo Cloud migration Project on time and $1.2M under budget.
Supervised team of 20+ staff members.
Designed and launched the Kellogg Kotex Cloud Transformation Project, resulting in a 43%
decrease in Total cost Of Ownership (TCO)for the company.
Achieved 60% time optimization, integrating internal github and Servicenow by creating ChatBot
using chatgpt LLM model for internal users in Aetna
Achieved 38% monthly billing cost of GCP introducing Compute projects overall for apps and ML
platforms. Also fine tuned bottlenecks by analyzing current graphs on GCP billing costs which
resultant huge success on TCO