Srilakshi M Resume
Srilakshi M Resume
Srilakshi M Resume
Professional Summary:
Experienced AWS Data Engineer with expertise in deploying and managing AWS services including EC2, Lambda,
Beanstalk, Batch, ECS, S3, EFS, Glacier, CloudTrail, CloudWatch, Athena, EMR, Glue, Redshift, SNS, SQS, Step
Functions, Code Deploy, and Code Pipeline.
Strong experience in designing, building, and maintaining data pipelines, data lakes, and data warehouses on AWS
platform using various services such as Glue, Athena, EMR, Redshift, and Kinesis.
Proficient in writing complex SQL queries and creating ETL scripts using Python or Spark to process and transform
large datasets stored in S3 or other data sources.
Skilled in data modelling, schema design, and performance optimization for big data applications.
Experience in configuring and monitoring data workflows using AWS services such as CloudTrail, CloudWatch, SNS,
and SQS.
Designined, implemented, and tested machine learning models to extract valuable insights from large and complex
datasets.
Developed data models using Star and Snowflake schemas, ensuring optimal performance and flexibility for
analytical queries.
Familiar with best practices for security, compliance, and data governance on AWS platform.
Demonstrated ability to work in a fast-paced and collaborative environment, and to communicate effectively with
cross-functional teams.
AWS certified in one or more relevant areas such as AWS Certified Big Data - Specialty, AWS Certified Solutions
Architect - Associate, or AWS Certified Developer - Associate.
Experienced AWS Data Engineer with expertise in Python object-oriented programming, PySpark, and data storage
technologies such as RDS and DynamoDB.
Experience building out machine learning algorithms for fraud detection and risk analysis.
Developed data models and implemented ETL processes using Ab Initio and Snowflake, ensuring efficient data
integration, transformation, and loading from various sources.
Developed ETL pipelines in and out of data warehouse using a combination of Python and Snowflakes Snow SQL
and SQL queries against Snowflake.
Proficienct in programming languages such as Python, R, or Java, and experience with machine learning libraries
(e.g., TensorFlow, PyTorch, scikit-learn).
Expertise in transforming business requirements into analytical models, designing algorithms, building models,
developing Data Mining, Data Acquisition, Data Preparation, Data Manipulation, Feature Engineering, Machine
Learning Algorithms, Validation and Visualization and reporting solutions that scales across massive volume of
structured and unstructured Data.ml
Skilled in designing and implementing robust data pipelines using AWS services such as S3, Glue, and Lambda
functions.
Proficient in developing RESTful APIs and integrating with third-party services, utilizing JSON as the primary data
exchange format.
Proficient in GCP core services , including Compute Engine, Storage, and Networking.
Proven track record of delivering end-to-end data solutions that meet business requirements, from data ingestion
to visualization and reporting.
Strong analytical and problem-solving skills, with the ability to identify and resolve data-related issues quickly and
efficiently.
Excellent communication and collaboration skills, able to work effectively with cross-functional teams and
stakeholders to ensure project success.
Experienced AWS Data Engineer with a strong background in SQL, Airflow, Avro, Parquet, Sequence, JSON, ORC,
Kafka, Spark, HDFS, Hadoop, HiveQL, and HBASE.
Skilled in data modeling, ETL development, data warehousing, and data integration using various tools and
technologies such as Python, PySpark, Scala, and Java.
Used apache airflow in GCP composer environment to build data pipelines used various airflow operations
like bash operator.
Demonstrated ability to work with cross-functional teams to understand business requirements and translate them
into technical solutions that deliver value to the business.
Strong understanding of data security, compliance, and governance frameworks, including GDPR, CCPA, HIPAA, and
PCI.
Experience in performance tuning, optimization, and troubleshooting of data pipelines and workflows to ensure
timely and accurate data delivery.
Excellent communication, collaboration, and leadership skills with a track record of delivering high-quality projects
on time and within budget.
Experienced AWS data engineer with expertise in designing, building, and maintaining scalable data solutions in the
cloud.
Skilled in data warehousing, data lakes, data marts, and big data technologies such as Hadoop, Spark, and EMR.
Proficient in data migration, cloud migration, and ETL processes using tools such as AWS Glue, AWS Data Pipeline,
and Apache Nifi.
Utilized Snowflake for data warehousing, implementing SnowSQL, SnowPipe, Streams, Tasks, Shares, Snowpipe,
Data Sharing, zero copy Clone, Materialized views, Time Travel, and advanced SnowSQL for optimized data
processing and analysis.
Familiar with SCD1 and SCD2 techniques for handling slowly changing dimensions in data warehouses.
Skilled in GCP integration, data infrastructure design, and ETL processes. Strong collaborator and problem-solver
with a focus on delivering high-quality solutions. Continuously updated on cloud technologies.
Knowledgeable in testing frameworks such as JUnit and JMeter for ensuring data quality and performance.
Expertise in ELT (Extract, Load, Transform) processes. Skilled in designing and implementing efficient data pipelines
for optimal data extraction, loading, and transformation. Proficient in AWS technologies and Collaborative team
player focused on delivering high-quality data solutions. Up-to-date with the latest advancements in ELT
methodologies and cloud technologies.
Proficient in project management tools such as Jira for agile development methodologies.
Experienced in version control tools such as GitHub and Bitbucket for managing code repositories and collaborating
with team members.
Experienced professional proficient in leveraging Alation to streamline data management, enhance metadata
organization, and facilitate data-driven decision-making
Analysed complex business challenges and devised innovative solutions to improve efficiency, productivity, and
profitability.
Proficient in leveraging PostgreSQL and Alation to streamline data organization, enhance metadata management,
and facilitate data-driven decision-making for optimal business outcomes.
Education:
Technical skills:
AWS Services AWS s3, redshift, EMR, SNS, SQS, Athena, glue, CloudWatch, IAM
Big Data Technologies HDFS, SQOOP, PySpark, hive, MapReduce, spark, spark streaming, HBASE
Hadoop Distribution Cloudera, Horton Works
Languages SQL, PL/SQL, Python, HiveQL, Scala.
Operating Systems Windows (XP/7/8/10), UNIX, LINUX, UBUNTU, CENTOS.
Database Teradata, oracle, SQL server,
Scheduling Control-m, oozie, airflow
Version Control GIT, GitHub, VSS
Methodology Agile, Scrum,Jira.
IDE &Build Tools, Design Eclipse, Visual Studio.
Cloud Computing Tools AWS, Snowflake.
Work Experience:
Environment: Sqoop, MYSQL, HDFS, Apache Spark Scala, Hive Hadoop, Cloudera, Kafka, MapReduce, Zookeeper, Oozie, Data
Pipelines, RDBMS, Python, PySpark, Ambari, JIRA.
Environment: SQL Server 2008/2012 Enterprise Edition, SSRS, SSIS, T-SQL, Windows Server 2003, Performance Point Server
2007, Oracle 10g, visual Studio 2010.