Divya Sri

Divya Sri Email and Phone Number

Looking for C2C opportunities!AWS Certified I’m seeking C2C (Corp-to-Corp) Senior Data Engineer roles specializing in ETL, AWS, Azure, and Big Data (Hadoop, Spark, Kafka). @ Fifth Third Bank
cincinnati, ohio, united states
Divya Sri's Location
United States, United States
About Divya Sri

I’ve developed strong expertise in building, optimizing, and maintaining data solutions on AWS. My work has centered on designing scalable ETL pipelines, data lakes, and data warehouses to support analytics and business intelligence. I use AWS services like Glue for data transformations, S3 for data lake storage, and Redshift for warehousing, ensuring that data flows smoothly from ingestion through to analysis.I’m proficient in setting up and managing data ingestion using Kinesis for real-time streaming data and Data Pipeline for batch processing. I’ve also implemented complex workflows using Step Functions to orchestrate multi-step ETL processes and Lambda for serverless data transformations, reducing operational overhead while maintaining scalability and flexibility. Security and compliance are critical in my projects, so I leverage IAM for access management, along with encryption strategies to ensure data security.Data processing frameworks: Apache Spark, Hadoop, and DatabricksData streaming tools: Kafka and Kinesis Skilled in ETL and data integration tools: AWS Glue, SSIS, and Airflow.Programming languages: Python, SQL, and Scala.Cloud environments: AWS (S3, Lambda, Redshift), Azure (Data Factory, Synapse), and Snowflake.Familiar with data modeling and query optimization using SQL and NoSQL databases, including MySQL, PostgreSQL, MongoDB, and DynamoDB.For data processing, I’ve built Spark jobs on EMR, optimized for performance and cost efficiency, using both PySpark and Scala. My experience extends to designing and deploying Redshift clusters, where I optimize performance through partitioning, sort keys, and distribution strategies, as well as using Spectrum for querying data stored in S3 directly.I also work with stakeholders to understand data requirements, translating them into technical specifications. I use Python and SQL extensively for ETL scripting, data manipulation, and performance tuning, ensuring the pipelines are efficient and cost-effective. In terms of monitoring and alerting, I utilize CloudWatch to track data pipeline performance and set up alerts, allowing proactive resolution of issues before they impact downstream systems.

Divya Sri's Current Company Details
Fifth Third Bank

Fifth Third Bank

View
Looking for C2C opportunities!AWS Certified I’m seeking C2C (Corp-to-Corp) Senior Data Engineer roles specializing in ETL, AWS, Azure, and Big Data (Hadoop, Spark, Kafka).
cincinnati, ohio, united states
Website:
53.com
Employees:
19739
Divya Sri Work Experience Details
  • Fifth Third Bank
    Senior Data Engineer
    Fifth Third Bank Mar 2023 - Present
    Columbus, Ohio Metropolitan Area
    Implemented distributed ETL solutions using AWS (S3, EC2, Glue, EMR, PySpark) for alerting and reporting enhancements.Developed CI/CD pipelines with AWS CloudFormation and GitHub Actions, reducing deployment errors and downtime.Collaborated with data science teams to operationalize ML models using SageMaker and EMR for scalable training.Designed real-time streaming data pipelines using Amazon Kinesis, Lambda, and Kafka, processing over 1M events/day.Automated data migration from… Show more Implemented distributed ETL solutions using AWS (S3, EC2, Glue, EMR, PySpark) for alerting and reporting enhancements.Developed CI/CD pipelines with AWS CloudFormation and GitHub Actions, reducing deployment errors and downtime.Collaborated with data science teams to operationalize ML models using SageMaker and EMR for scalable training.Designed real-time streaming data pipelines using Amazon Kinesis, Lambda, and Kafka, processing over 1M events/day.Automated data migration from legacy systems using AWS Data Pipeline and DMS with zero data loss.Led a team in adopting DevOps practices, using IaC, automated testing, and monitoring via AWS CloudWatch and X-Ray.Utilized Databricks for collaborative data processing, analytics, and ML model development.Developed and optimized NLP algorithms for text analysis at scale.Built real-time fraud detection models using Spark MLlib in Databricks.Processed real-time data with Amazon Kinesis Data Analytics, identifying patterns in transactional data.Automated CI/CD pipelines with GitHub Actions, Docker, and Terraform, deploying on Kubernetes.Managed infrastructure with Terraform, ensuring secure data access via AWS IAM and KMS.Designed and developed Power BI reports, established data relationships, and performed data validation. Show less
  • At&T
    Senior Data Engineer
    At&T Dec 2021 - Sep 2022
    Ramona, California, United States
    Developed data integration and analytics solutions using Hadoop (Hortonworks) and big data technologies.Participated in Agile methodologies, including daily Scrum and Sprint planning.Performed data transformations in Hive, optimizing performance with partitions and bucketing.Created Hive external tables based on MapReduce outputs, enhancing query performance.Monitored system health using Amazon CloudWatch, set alarms, and logged system events.Designed AWS cloud… Show more Developed data integration and analytics solutions using Hadoop (Hortonworks) and big data technologies.Participated in Agile methodologies, including daily Scrum and Sprint planning.Performed data transformations in Hive, optimizing performance with partitions and bucketing.Created Hive external tables based on MapReduce outputs, enhancing query performance.Monitored system health using Amazon CloudWatch, set alarms, and logged system events.Designed AWS cloud infrastructure blueprints, including EC2, S3, Route53, Load Balancers, and VPC.Developed custom UDFs in Hive and Pig tailored to business needs.Led Spark performance optimization and COE initiatives at Cisco for data simplification.Implemented Spark with Scala/Java, leveraging DataFrames and Spark SQL for efficient processing.Spearheaded multi-terabyte AWS Redshift data warehouse design, improving query performance by 50%.Built a secure data lake in S3 integrated with Lake Formation and Glue, supporting ML workloads.Developed Spark-SQL and Spark Streaming code for fast data processing and testing.Built data pipelines using Kafka, HBase, Spark, and Hive for customer behavioral data analysis.Managed metadata and migration of applications to Hive and Spark.Extracted data from Oracle and MongoDB using Sqoop, transformed it with Spark/Hive, and loaded it into HDFS.Installed and managed Hadoop ecosystem components, including Spark, Hive, and HBase.Ensured data security in AWS Data Lake with IAM policies, adhering to regulatory standards.Developed daily reports on BigQuery tables using SQL and matplotlib, including failure detection.Collaborated with stakeholders to architect and deliver Big Data analytics solutions. Show less
  • Lowe'S Companies, Inc.
    Senior Data Engineer
    Lowe'S Companies, Inc. Jul 2020 - Nov 2021
    Orlando, Florida, United States
    Evaluated business requirements and prepared specifications for program development.Loaded and transformed structured, semi-structured, and unstructured data using Hadoop/Big Data concepts.Created automated build and deployment processes, implementing continuous integration systems.Developed and optimized MapReduce programs for unstructured data and HDFS utilization.Designed Big Data analytics platforms using Hadoop, Hive, Pig, and Cloudera for customer insights.Exported… Show more Evaluated business requirements and prepared specifications for program development.Loaded and transformed structured, semi-structured, and unstructured data using Hadoop/Big Data concepts.Created automated build and deployment processes, implementing continuous integration systems.Developed and optimized MapReduce programs for unstructured data and HDFS utilization.Designed Big Data analytics platforms using Hadoop, Hive, Pig, and Cloudera for customer insights.Exported data to relational databases via Sqoop for BI team reporting.Designed and developed Azure relational databases and created tabular models on Azure Analysis Services.Used Azure services (ADLS, Synapse Analytics) for data processing and transformations with HDInsight and Hive.Developed Spark and SparkSQL code on Azure HDInsight for data transformations.Created ETL pipelines in Azure Data Factory (ADF) from various sources like Azure SQL and Blob storage.Developed Spark scripts using Python and Scala for faster data processing.Designed dimensional data models using Star and Snowflake schemas.Configured monitoring systems with Nagios, MCS, and CloudWatch for performance tracking.Worked with relational databases (Oracle, SQL, PostgreSQL) and Snowflake for parallel data processing.Implemented DevOps pipelines using Openshift and Kubernetes for microservices architecture. Show less
  • Cybage Software
    Big Data Developer
    Cybage Software Jun 2016 - Aug 2019
    Hyderabad, Telangana, India
    • Implemented data ingestion and Data-at-Rest processing solutions using Hadoop, MapReduce Frameworks, HBase, and Hive.• Utilized Sqoop to transfer data efficiently between databases and HDFS, and Flume for streaming log data from servers.• Loaded and transformed large sets of structured, semi-structured, and unstructured data from relational databases into HDFS using Sqoop imports.• Implemented an enterprise-grade platform (MarkLogic) for ETL from mainframe to NoSQL… Show more • Implemented data ingestion and Data-at-Rest processing solutions using Hadoop, MapReduce Frameworks, HBase, and Hive.• Utilized Sqoop to transfer data efficiently between databases and HDFS, and Flume for streaming log data from servers.• Loaded and transformed large sets of structured, semi-structured, and unstructured data from relational databases into HDFS using Sqoop imports.• Implemented an enterprise-grade platform (MarkLogic) for ETL from mainframe to NoSQL (Cassandra).• Responsible for importing log files into HDFS from various sources using Flume.• Analyzed data using HiveQL to generate payer reports for payment summaries transmission.• Imported millions of structured data from relational databases using Sqoop import, processed it using Spark, and stored it in HDFS in CSV format.• Leveraged Scala's Data Frame API for converting distributed collections of data organized into named columns.• Conducted data profiling and transformation on raw data using Pig, Python, and Java.• Developed Shell, Perl, and Python scripts for automating and controlling flow in Pig scripts.• Created prototypes for Big Data analysis using Spark, RDD, Data Frames, and the Hadoop ecosystem, working with CSV, JSON, Parquet, and HDFS files.• Developed Hive SQL scripts for transformation logic and data loading from staging to landing and semantic zones.• Managed and led development efforts with a diverse team, including both internal and overseas members Show less
  • Amigos Software Solutions Private Limited
    Hadoop Developer
    Amigos Software Solutions Private Limited Apr 2014 - May 2017
    Hyderabad, Telangana, India
    • Involved in end-to-end process of setting up Hadoop clusters, including installation, configuration, and monitoring.• Automated the setup of Hadoop clusters and implemented Kerberos security for various Hadoop services using Hortonworks.• Responsible for cluster maintenance, including commissioning and decommissioning of data nodes, cluster monitoring, troubleshooting, data backups, and managing Hadoop log files.• Installed and configured Hive, Pig, HBase, and Sqoop on the Hadoop… Show more • Involved in end-to-end process of setting up Hadoop clusters, including installation, configuration, and monitoring.• Automated the setup of Hadoop clusters and implemented Kerberos security for various Hadoop services using Hortonworks.• Responsible for cluster maintenance, including commissioning and decommissioning of data nodes, cluster monitoring, troubleshooting, data backups, and managing Hadoop log files.• Installed and configured Hive, Pig, HBase, and Sqoop on the Hadoop cluster.• Configured property files like core-site.xml, hdfs-site.xml, mapred-site.xml based on job requirements.• Performed data analysis, feature selection, and feature extraction using Apache Spark Machine Learning streaming libraries in Python.• Analyzed system failures, identified root causes, and recommended actions. Documented processes and procedures for future reference.• Collaborated with systems engineering team to plan and deploy new Hadoop environments and expand existing Hadoop clusters. Installed and configured Kerberos for user authentication and Hadoop daemons. Show less

Frequently Asked Questions about Divya Sri

What company does Divya Sri work for?

Divya Sri works for Fifth Third Bank

What is Divya Sri's role at the current company?

Divya Sri's current role is Looking for C2C opportunities!AWS Certified I’m seeking C2C (Corp-to-Corp) Senior Data Engineer roles specializing in ETL, AWS, Azure, and Big Data (Hadoop, Spark, Kafka)..

Who are Divya Sri's colleagues?

Divya Sri's colleagues are Christy Anderson, Bhavana P, James Charek, Jean Koerner, Christopher Wesley, Cortney Sheldon, Erin Smith.

Not the Divya Sri you were looking for?

  • Divya Sri

    Power Bi Developer | Data Analyst | Business Intelligence Analyst
    United States
  • Divya Sri

    Senior Data Engineer At Amex Group | Actively Looking For C2C/C2H Position As Data Engineer| Pl/Sql | Sql | Nosql | Aws | Bigdata | Azure | Hadoop | Etl | Snowflake| Databricks | Bi | Python
    Phoenix, Az
  • DIVYA S.

    Sr. Java J2Ee Developer
    Harrison, Nj
  • Divya Sri

    Boston, Ma

Free Chrome Extension

Find emails, phones & company data instantly

Find verified emails from LinkedIn profiles
Get direct phone numbers & mobile contacts
Access company data & employee information
Works directly on LinkedIn - no copy/paste needed
Get Chrome Extension - Free

Aero Online

Your AI prospecting assistant

Download 750 million emails and 100 million phone numbers

Access emails and phone numbers of over 750 million business users. Instantly download verified profiles using 20+ filters, including location, job title, company, function, and industry.