Chiranjib - Updated Resume

Download as doc, pdf, or txt
Download as doc, pdf, or txt
You are on page 1of 12

CHIRANJIB

Cloud Solution Architect

As a Data & Analytics Cloud data Solutions Architect/Senior Data Engineer, I have over 17 years of
experience in data projects, business intelligence, cloud migration, AI/ML projects in the Healthcare and
Retail sectors. Have extensive experience with data insights of HC Clinical analytics, CPG Supply chain &
Sales domain. Holding GCP and AWS certifications and have worked onsite in the USA and Australia.

My core competencies include developing comprehensive migration strategies, reverse engineering


existing systems, designing, and implementing cloud solutions using various cloud platforms (Experienced
in GCP, AWS and Azure) and tools, defining IT roadmap and architecture, and working with cloud
security and DevOps teams. I am passionate about delivering seamless and scalable solutions that meet the
business needs and goals of clients. Also enthusiastic about learning and applying new technologies such
as AI and ML to enhance data and analytics capabilities.

PROFESSIONAL SUMMARY
 Extensive experience with working as cloud data migration, architect/senior consultant from on
premises data warehouse into different cloud platforms
 Develop comprehensive migration strategy, including the use of Lift and Shift and Transformation
methods, to ensure a seamless transition to cloud-based platforms like aws/gcp/azure
 Reverse engineering of the existing systems like on premises databases and Hadoop systems and build
strategy to migrate historical and on -going (delta) data into new cloud platform.
 Extensive experience with design and implementation of GCP cloud migration using GCP native
services like dataproc, cloud spanner, app engine, BigQuery , Compute engine and devops services
like GKE, Composer, Terraform, Jenkins
 Extensive experience on design and implementation Azure cloud solutions using medallion
architecture (bronze-silver-gold) pattern using Databricks, Data factory, ADF, Synapse, ADLS Gen2,
blobs, Delta Lake
 Experienced with design and implementation of AWS cloud solutions using Glue ETL,S3 (lake),
Lambda, Cloud functions, Redshift(warehouse) and scheduling through airflow, control M, Autosys
for batch processing
 Experienced with working accelerator tools to migrate existing application codes to equivalent
Pyspark code for large scale application/data migration.
 Extensive experienced defining IT Roadmap, cloud strategy, Enterprise/solution architecture,
Architecture assessment
 Worked in Azure platform validation for the newly acquired Azure lake and Synapse
 Expertise in DevOps, Release Engineering, Configuration Management, Cloud Infrastructure,
Automation. It includes ADF, Amazon Web Services (AWS), GCP(Google Cloud platform),Ant,
Maven, Jenkins, Chef, GitHub,
 Worked with cloud security team to make compliance with all the security protocol used enterprise
wise. Worked in all kinds of cost optimization techniques used in cloud platforms
 Extensive previous experience in building data warehouse on Teradata
 Demonstrated experience on working with USA Healthcare Data & Analytics (major in Care
management area) , experienced handling Membership/Claims/Provider data with EDI formats like
834.274.837.HL7 etc
 Demonstrated experience in CPG(Consumer product Goods) business model implementation in retail
projects, while migrating Master data, Sales data with domain granularization

 Excellent in building strong relationships with clients and providing support to all stakeholders
including senior leadership, helping with different metric generation
 Deft in providing guidance to business unit partners regarding issues or opportunities.
 Wrote technical white paper on HealthCare (Clinical Analytics)
 Successfully executing Cloud migration data projects in cloud platforms by transforming traditional
Data warehouse, whereas reducing 40% cost
 Managing team of 15+ in onshore offshore delivery model , typically using Agile scrum method

SKILLS
 Cloud architecture  Data lake Applications  Big Data Tools
 Project Management  Data Migration  Technical Support
 Software Development  Data architecture  Client Relationship
 Product Development  Business Development  Team Management

EDUCATION
 Bachelor of Computer Science & Technology from West Bengal University
of Technology, India, 2006

CERTIFICATIONS:
No Name of Professional Certificates Year of
Institution Completio
n
Google Cloud Google Cloud, Google Certified 2024
Professional Machine Learning Engineer
Certificate ID-uZ4kLh
Databricks Accredited Generative AI Fundamentals 2024

1 Google Cloud Google Cloud Architect 2022


Certificate ID-uZ4kLh
2 Google Cloud Associate Cloud Engineer 2022
Certificate ID- N1D2vu
3 Amazon Web AWS Certified BigData Specialty– 2020
Services Certificate ID : “1PNEGGKC3FEQ1CW4”
4 Teradata Teradata certified Master(V2R5) 2010

Projects Handled
Project Title CVS Health (Aetna) GCP cloud transformations
Duration Jul 2023-Current
Location Hartford CT ,USA
Role Cloud Solutions Architect/Senior Data Engineer
Technologies used GCP services (Big Query, Spanner, Dataproc, Cloud SQL,
Terraform, GKE, Jenkins etc) , Hadoop , Sql Server
databases, Autosys,

PROJECT DESCRIPTION:
CVS health (Aetna) is undergoing tech modernization while transforming on
premises systems into Google cloud. As part of migration 40+ applications need to
move in GCP BigQuery (BQ), while existing reporting systems will repoint to BQ for
KPI’s. Goal is to migrate the On premises databases and Hadoop hosted applications
into BQ in phase wise
Key Deliverables:
 Actively involved in a program geared towards migrating existing mission and
business applications into a cloud-based environment. Activities required to
re-host an application into the cloud may include architecture modifications,
database and/or application server re-hosting, and potentially recoding of
existing capabilities to take advantage of cloud platform services.
 Reverse engineering of 40+ applications from existing Hadoop system and
build roadmap for GCP- BQ migration. Planning for pilot phase data migration
for 3 applications
 Collecting data sets that meets functional / non-functional business
requirements. Working with data and analytics experts to strive for greater
functionality in data system.
 Guiding team of 20+ members on 3 tracks for reverse engineering, mapping
creation and data pipeline creation. Building strategy for historical data
migration (using existing data in Hadoop and SQL server) with minimum
changes to accumulate exiting KPI’s and new data pipelines for on – going data
ingestion
 Creating new data pipelines from source to target for incremental ingestion
and existing data import in GCP BQ as historical load for KPI and analytics use
 Worked in creating new chatbot using LLM, integrating ServiceNow and
Github for internal user use for faster response on knowlegebase
 Working closely with customer Directors and VP’s assisting them for
continuous betterment of the program delivery
 Creating Metrics/dashboards using Rally for all levels of burndowns for
successful migration

Precedence Experience:
Project Title PepsiCo Cloud Transformations
Duration Aug 2022- Jun 2023
Location Kolkata, India
Role Cloud Solutions Architect
Technologies used Azure Data factory, Synapse, Databricks,
Teradata,Unix

PROJECT DESCRIPTION:
‘PepsiCo Inc’ has IT strategic movement to modernize the platform from Legacy
systems to Cloud. ‘PepsiCo cloud transformations’ is the migration’ program from
Teradata to Azure Synapse platform. The project needs to migrate all the existing
Teradata analytical applications into Azure Synapse. Reporting tools will repoint to
Synapse for KPI’s. This project is planned by phases starting from pilot followed by
Phase 1 , Phase 2 , Phase 3.Objective is to onboard business users to provide benefits
of modern cloud technologies with cost affective robust solutions and finally retire
Teradata from PepsiCo landscape
Key Deliverables
 Actively involved in a program geared towards migrating existing mission and
business applications into a cloud-based environment. Activities required to
re-host an application into the cloud may include architecture modifications,
database and/or application server re-hosting, and potentially recoding of
existing capabilities to take advantage of cloud platform services.
 Worked as Azure architect/Data management techno-functional lead from
offshore
 Approach finalization for historical and ongoing incremental data load
 Worked in reverse engineering of the Teradata existing jobs to build the
inventory for app migration.
 Worked in creating detailed level design document for pilot apps
 Develop solutions for Pilot phase, to migrate ‘Power Of One’- Sales and
Marketing KPI application from Teradata to Azure Synapse, by converting
existing TD Bteq logics into ADB PySpark codes. Move Teradata ACQ layer
data using S&T framework into Bronze and then build the silver and gold layer
inside Delta lake. Finally move the data into Synapse using store procedures.
 Worked on production go-live/cutover for pilot apps
 Responsible for mentoring the team of 10+ members in offshore for seamless
delivery
 Collecting data sets that meets functional / non-functional business
requirements
 Working with data and analytics experts to strive for greater functionality in
data system
 Prepare for future applications discovery and analysis for upcoming sprints
 Following Agile scrum model to adhere to PepsiCo’s standard agile practices.

Project Title Bayer’s TD-GCP migration


Duration Mar 2022- Jul 2022
Location Kolkata, India
Technologies used GCP services, SAP, Teradata, PySpark,Unix

PROJECT DESCRIPTION:
‘Bayer Corp Science’ has IT strategic movement to modernize the platform from
Legacy systems to Cloud. ‘TD-GCP migration’ project is the platform modernization
from Teradata to Google BigQuery platform for CS (Corp science) division. The
project needs to migrate all the existing data to migrate in Google Cloud along with
that new data pipeline creation which will bring the data from SAP and Non- SAP
sources into GCP (Google cloud platform) on regular basis. Existing e applications and
reporting platforms will repoint from BigQuery. Teradata will be decommissioned
after the migration is completed
Key Deliverables:
 Actively involved in a program geared towards migrating existing mission and
business applications into a cloud-based environment. Activities required to
re-host an application into the cloud may include architecture modifications,
database and/or application server re-hosting, and potentially recoding of
existing capabilities to take advantage of cloud platform services.
 Create and maintain project related documentation (Statement of Work
Document, Proof of concept document, technical design document etc.)
 Develop overall solutions including Statements of Work and Bills of Materials
 Responsible for creating and maintaining optimal data pipeline architecture.
 Create end to end design for applications to host on GCP platform.
 Collecting data sets that meets functional / non-functional business
requirements.
 Working with data and analytics experts to strive for greater functionality in
data system
 Build data warehouse solutions on BQ platform, to make sure all the existing
applications perform as-is or better to meet client business SLA
 Working with different API's on extraction of data
 Managing team of 10+ members, in the POD model for project execution
 Gathering customer requirements on regular basis and working on multiple
RFP’s to meet Bayer’s IT strategic and architectural goals

Project Title Kellogg Kortex Migration


Duration Jan 2021-Feb 2022
Location Kolkata, India
Technologies used AWS services, Sql Server, Hadoop, SAP,
PySpark

PROJECT DESCRIPTION:
Kellogg Company’s strategic movement towards migrating the on Premises databases
and Big data Hadoop into AWS cloud This project includes the movement of data
from on prem SQL server and Hadoop systems to re-platform into AWS cloud by
migrating the data and existing applications into AWS and decommissioning the
Hadoop and SQL server boxes
Key Deliverables
 Strategist and AWS data engineer track lead on re-engineering of applications
to migrate from on premises and external file systems to AWS cloud services,
which involved data warehouse rebuilding over AWS which will be accessed
through all the reporting services and analytics team
 Worked on classification of the existing tables in domain group (Master Data
and Sales data major) in the activity of re-engineering and grouping of the
existing tables for building domain granular model
 Involved in Kellogg specific new data model activities which was restructuring
the tables on domain and region basis
 Architecting, developing and prototyping of different frameworks to bring
data from on premises to cloud using Spark job (written in Python ) with AWS
Glue services for non-SAP sources and confluence file type
 Create end to end design for applications to host on AWS which involved s3
RAW-> S3 CleansedProcessed->Redshift Target Db for domain creating
domain granular layer
 Created STTM document for Cleansed to Redshift target tables using current
transformation logics which were in on premises sql
 Worked extensively with Cognizant accelerator framework (DIF tool) to move
data raw to cleansed layer while developing all kind of file formats like
csv/txt/xls/xlsb/json/dat to accumulate in cleansed layer
 Written Glue ETL scripts to move the data from S3 Raw to cleansed bucket
using data quality rules
 Worked on moving data from cleansed to processed layer -> Redshift DW
layer to make the data available for reporting and analytics
 Written transformations for target tables in Redshift using Glue ETL with
python programming language
 Involved in developing DevOps activities for applications using Github.
 Worked extensively on analysis of SAP global tables ECC/SCM (major)
modules for analysis and mapping for building domain granular data model
 Successfully migrated Master data and Sales data in Domain Granular
model(Material/Customer/location /Foundation reference/Sales-
Execution/Strategy/Performance/Order etc) into AWS environment.
 Working as advisory to all the different tracks for data related issue
fix/debugging in all around the tracks in Kellogg
 Worked with AWS support for fixing different issues while executing the Glue
codes which were not able to execute due to VPC/subnet mask.
 Managed a team of 15+ members in the Waterfall and POD model of
execution.

Project Title WALMART Tech modernization


Role Big Data Engineer, Architect, Developer
Location Kolkata , India
Duration Jul 2019 – Dec 2020
Technologies used Big Data-Hadoop, Kafka , Spark, DB2, Teradata, SQL Server,
Cassandra, Azure Cosmos DB
Key Deliverables:
 Responsible for creating and maintaining optimal data pipeline architecture
 Collecting data sets that meets functional / non-functional business
requirements
 Working with data and analytics experts to strive for greater functionality in
data system
 Creating and maintaining Java Spingboot applications using Kafka Producer,
Consumer, Kstream
 Writing aggregation through Kstream, Ktable, GlobalKTable
 Working with different API's on extraction of data
 Loading/extracting Json files by using Azure Cosmos DB and Cassandra as per
requirement
 Developing Datalake applications using Hive, PIG, Sqoop process with
complex queries for regular extraction of jobs and send to respective vendors
for analytics
 Designing and creating Hive external tables using shared meta store instead of
derby with partitioning, dynamic partitioning and buckets
 Actively involving in HBASE setup and storing data in HBASE, which will be
used for analytics
 Using Spark -SQL to load JSON data and create schema RDD and loaded into
Hive tables and handled structured data using SparkSQL
 Used Teradata ,DB2 ,SQL server for SQL scripting to extract data from data
warehouse and loaded into Datalake using Sqoop for consumption/analytics
team

Project Title HCSC (HEALTH CARE SERVICE CORPORATION), Chicago,


IL
Role Application Development Lead, Systems Analyst, App Arch,
Data Engineer
Location Chicago, IL , USA
Duration Oct 2016 – Jun 2019
Key Deliverables:
 Worked as data delivery application lead for 20+ work streams in Care
management area which includes sensitive data like Membership, Claims,
provider, Lab, Biometrics, Authorization, ASO
 Developed Datalake applications using Hive, PIG, Sqoop process with complex
queries for regular extraction of jobs
 Worked closely with Business Analyst and understand the requirement
properly to create the source to target technical mapping spec. Worked on HC
specific file formats like :834,837,274,HL7,ncpdp for mapping creation
 Analyzed different data sources like Data lake (Hadoop systems), Databases
(Teradata, DB2), XML, Flat file, etc. to create correct dataset for analytics
 Extract and match data from data warehouse using Data stage and
Teradata,Db2 for member match and used Zena tool for automating the jobs
 Transformed Dw jobs in Datalake for cost optimization
 Data architecture roadmap creation for AWS cloud

Project Title Amgen – Teradata Application


development
Duration 07/15/2014-08/01/2016
Location Thousand Oaks, CA, USA
Technologies used Teradata, Hadoop, Informatica

PROJECT DESCRIPTION:
Amgen is one of the largest life sciences company in the world. It manufactures medicines with
innovations and research. This is the EDW Teradata Admin project for Amgen. It has total 9
Teradata systems for data warehouse which includes EDW systems and Radar EDW systems.

 Developed Bteq, Multiload, Fastload, TPT jobs in Teradata and using Shell scripts
 Involved in writing and tuning of critical views in different layers of the business application
 Identifying and tuning bad performing batch queries
 Responsible for tuning all kind of performance issues handling(L3 level)in Teradata applications

Project Title Walgreens – Data Factory


Duration Jan 2012 – Nov 2013
Location Chennai India, Chicago IL USA
Technologies used Teradata, Unix

PROJECT DESCRIPTION:
Walgreens is the one of the largest pharmaceutical retail in USA. Walgreens'
Datawahouse system is build in Teradata database. This is a warehouse for Walgreen’s
store data
Key Deliverables
Resolved data issues for reports such as missing data in the BO (business object)
report, failure analysis for Informatica job flows: Example: If there is any data missing
for any report we need to do RCA for the same such as analysis of query running in
backend/environmental issue etc.
Provided RCA for any regular job failure in Informatica, missing file in
servers/environmental issue/long running job issue etc.

Project Title Westpac – Data management


Duration Mar 2011– Dec 2011
Location Mumbai India, Sydney Australia
Role Application Development Lead, Systems Analyst
Technologies used Teradata, Unix, , C++

PROJECT DESCRIPTION:
Westpac is 3rd largest bank in Australia. This project is retail banking reporting
system. Retail banking is related to personal loan, home loan, credit card etc. The
reporting system is basically related to scorecards generation regarding the retail data
of the bank. This includes the growth of both the retail sales of the bank and the
performance of the bank employees.
Key Deliverables
 Write codes on JCL, Teradata, store procedure and Sql server as per required
enhancements
 Solve L3 level production issues
Project Title ATT Data Analyst DBA
Duration Jan 2010 – Dec 2010
Location Pune, India
Technologies used Teradata, Unix
PROJECT DESCRIPTION:
AT&T is world’s one of the most well-known and biggest Telecom company. There
are 33 markets in USA in which their telecom services is going on. This project is
support and enhancement of existing and upcoming applications in data warehouse
Key Deliverables
 Write codes on JCL, Teradata, store procedure and Sql server as per required
enhancements
 Solve L3 level production issues
 24/7 On-Call Teradata DBA support
 L3 application support for 3 applications

Project Title Vodafone – Teradata application DBA


Duration Nov 2006 – Jan 2010
Location Hyderabad , Pune India
Technologies used Teradata, Unix

PROJECT DESCRIPTION:
This is the project of Vodafone data warehouse. There are five source system from
which data is
coming. The total data is loaded in the data warehousing production box. The final
reporting’s are
done from this system. Total analytical process for the months are going here.
Key Deliverables:
 Write codes on JCL, Teradata, store procedure and Sql server as per required
enhancements
 Solve L3 level production issues

Accomplishments
 Presented with Best performance recognition from Kellogg and PepsiCo clients for AWS & Azure
Delta Lake implementation
 Completed the PepsiCo Cloud migration Project on time and $1.2M under budget.
 Supervised team of 20+ staff members.
 Designed and launched the Kellogg Kotex Cloud Transformation Project, resulting in a 43%
decrease in Total cost Of Ownership (TCO)for the company.
 Achieved 60% time optimization, integrating internal github and Servicenow by creating ChatBot
using chatgpt LLM model for internal users in Aetna
 Achieved 38% monthly billing cost of GCP introducing Compute projects overall for apps and ML
platforms. Also fine tuned bottlenecks by analyzing current graphs on GCP billing costs which
resultant huge success on TCO

End of the Document

You might also like