Sai Ganesh

Sai Ganesh Email and Phone Number

Senior Data Engineer @ CVS Health
United States
Sai Ganesh's Location
United States, United States
About Sai Ganesh

I am a results-driven Senior Data Engineer with over a decade of experience specializing in designing and implementing scalable data infrastructure solutions. Throughout my career, I have successfully led cross-functional teams in delivering robust ETL pipelines, data warehousing solutions, and cloud-based platforms such as AWS, GCP, and Azure. My proficiency extends to programming languages like Python, Scala, and SQL, enabling me to effectively manipulate and visualize data for actionable insights. I possess a deep understanding of Big Data frameworks including Hadoop and Spark, and have a proven track record in deploying machine learning pipelines to drive predictive analytics. With a strong commitment to data integrity, security, and compliance, I excel in optimizing data processes to meet business objectives efficiently. I am passionate about leveraging my skills and experience to contribute to innovative projects and look forward to connecting with professionals in the field.

Sai Ganesh's Current Company Details
CVS Health

Cvs Health

View
Senior Data Engineer
United States
Sai Ganesh Work Experience Details
  • Cvs Health
    Senior Data Engineer
    Cvs Health
    United States
  • Cvs Health
    Senior Data Engineer
    Cvs Health Mar 2023 - Present
    Providence, Rhode Island, United States
    • Enhanced Data Processing Efficiency: Managed intricate ETL systems utilizing Python, SQL, and AWS services such as Step Functions, EMR, Lambda, S3, and Redshift, achieving a 30% increase in data processing efficiency. • Rapid Issue Resolution: Engineered comprehensive platform issue resolutions using MongoDB, AWS Redshift, and Sumo Logic; ensuring stable system performance slashed downtime by 40%, and fortified system dependability.• Executed Scalable Data Warehousing Solution: Engineered a high-performance data warehousing solution using AWS Glue and Step Functions, enhancing data storage and retrieval processes resulting in a 25% decrease in data processing time.• Managed Complex Data Migration Projects: Led the successful execution of data migration projects, ensuring smooth integration of new data sources with minimal disruption; expanded data analytics capabilities by 60%, enabling enhanced insights.• Utilized MongoDB and implemented SCD Type 2 in Snowflake for personalized healthcare communications and historical tracking and leveraged CDC and transformed OLTP data into OLAP format for real-time updates and efficient data analysis• Implemented Airflow for Workflow Orchestration: Utilized Airflow to orchestrate complex data workflows, ensuring efficient execution and monitoring of data pipelines. Developed custom Airflow operators using Python to handle specific data processing tasks.• Developed and optimized complex SQL queries and Scala scripts for data transformation and analysis, collaborating with data scientists and business stakeholders in an Agile environment to deliver actionable insights and drive data-driven strategies.• Implemented CI/CD pipelines using GitHub and Jenkins, automating the build, test, and deployment processes for data pipeline code, and ensuring code quality and reliability through rigorous testing and code reviews.
  • Fannie Mae
    Senior Cloud Data Engineer
    Fannie Mae Oct 2020 - Feb 2023
    Plano, Texas, United States
    •Worked on a project that uses AWS lambda functions, DynamoDB, S3 Bucket, SQS, Secrets Manager, IAM, and Cloud Watch to automate the process of synchronizing all files from storage to AWS S3 bucket on a nightly basis.•Python was used to extract, transform, and aggregate data from multiple source systems for business use cases.•Sqoop was designed and implemented for an incremental job to read data from DB2 and load it into Hive tables, and it was linked to Tableau to generate interactive reports using Hive server2.•ETL/ELT processes were designed and developed to handle data migration from multiple business units and sources, including Oracle, Postgres, and Informix.•Enterprise Data Lake was designed and set up to support a variety of use cases, including analytics, processing, storing, and reporting on large amounts of rapidly changing data.•Created an API for using AWS Lambda to manage servers and run code in AWS.•ADF Pipelines were created to load data from on-premises storage to Azure cloud storage and databases.•On the AWS cloud, S3 buckets were used for file storage and backup.•Used cloud shell SDK in GCP to configure the services Data Proc, Storage, BigQuery.•Created end-to-end data processing pipelines that begin with data receiving via distributed messaging systems Kafka and end with data persisting into Cassandra.•Data Integration Processes were demonstrated/presented to business stakeholders, and work was done on developing standards and processes for integration initiatives.•Pipelines and analytical workloads were implemented using big data technologies such as Hadoop, Spark, Hive, and HDFS.•Integrated Tableau with a Hadoop data source to create a dashboard that provides various insights into the organization's sales.•Importing data to Data Lake Raw Zone using Sqoop from various source systems such as mainframes, Oracle, MySQL, DB2, and so on.
  • Csi
    Senior Data Engineer
    Csi May 2018 - Sep 2020
    Atlanta, Georgia, United States
    • Designed, developed, and implemented Hadoop applications using Sqoop, Spark, Hive and MapReduce• Responsible for loading structured, unstructured, and semi-structured data into Hadoop by creating static and dynamic partitions.• Strong knowledge of various Data warehousing methodologies and Data modeling concepts.• Developed Hadoop streaming Map/Reduce works using Python.• Improved query performance and space utilization by 40% by creating Hive partitioned tables using Parquet & Avro format.• Responsibilities include Database Design and Creation of User Database. Moving ETL pipelines from SQL server to Hadoop Environment.• Extensively worked on NGAP for managing clusters and accessing them.• Scheduling mechanism in Airflow for developing jobs for scheduling DAG’s.• Developed complex SQL queries and managed roles in Snowflake.• Worked on ingestion data from multiple relational databases including Snowflake into S3 buckets.• Designed, developed and implemented Hadoop applications using Sqoop, Spark, Hive and MapReduce• Developed Spark applications using Pyspark and Spark-SQL for data extraction, transformation, and aggregation from multiple file formats for analyzing & transforming the data to uncover insights into the customer usage patterns.• Implemented a CI/CD pipeline using Jenkins, Airflow for Containers from Docker, and Kubernetes.• Used Python scripts, Spark Applications for ETL Operations to create data flow pipelines and involved in transforming data from legacy tables to Hive, HBase tables, and S3 buckets for handoff to business and Data scientists to create analytics over the data.• Support current and new services that leverage AWS cloud computing architecture including EMR S3, and other managed service offerings.• Implemented data quality checks using Spark Streaming and arranged bad and passable flags on the data.• Used advanced SQL methods to code, test, debug, and document complex database queries.
  • Albertsons
    Data Engineer
    Albertsons Jul 2016 - Apr 2018
    Pleasanton, California, United States
    • Proficient in writing efficient SQL queries, developing stored procedures, and optimizing Snowflake SQL for enhanced performance and ELT/ETL processes.• Expertise in Snowflake data modeling, user management, and Role-Based Access Control (RBAC).• Developed and scheduled automated processes, including creating jobs, using SQL Mail Agent, and setting up alerts and DTS/SSIS packages. • Provided technical guidance and best practices for access controls, data security, and Snowflake resource optimization.• Implemented and optimized ETL processes using SQL, PL/SQL, and stored procedures for data integrity and quality.• Developed automated processes and managed databases, ensuring optimized performance and reliable deployments.• Documented technical and management processes and tracked and resolved defects for effective issue management.• Maintained users, roles, and permissions within the SQL Server.
  • Intellect Design Arena Ltd
    Data Warehouse Developer
    Intellect Design Arena Ltd Aug 2013 - Feb 2016
    Hyderabad, Telangana, India
    • Demonstrated experience in Agile Scrum methodology, actively participating in daily stand-up meetings and collaborating with cross-functional teams for successful project delivery.• Proficient in utilizing Visual SourceSafe for version control and code management in Visual Studio 2010, coupled with adept usage of project tracking tools like Trello for monitoring and facilitating project progress.• Expertise in designing and optimizing data warehouse solutions, leveraging relational databases (Oracle, SQL Server) to efficiently store and process large-scale datasets for analytics and reporting.• Created interactive reports in Power BI with drill-through and drill-down capabilities, drop-down menus, sorting, and subtotals, enabling comprehensive data exploration.• Leveraged the data warehouse to develop a Data Mart, feeding downstream reports and designing a user access tool for ad-hoc reporting and query analysis, empowering self-service data insights.• Deployed SSIS packages and orchestrated efficient job scheduling for streamlined data integration processes.• Experienced in architecting and constructing Cubes and Dimensions with diverse data sources and architectures, utilizing MDX scripting to enhance Business Intelligence capabilities.• Developed SSIS jobs for report automation and cube refresh packages.• Proficient in deploying SSIS packages to production and utilizing package configurations for environment independence.• Experienced with SQL Server Reporting Services (SSRS) for authoring, managing, and delivering paper-based and interactive web-based reports.• Developed stored procedures and triggers to ensure consistent data entry into the database.

Frequently Asked Questions about Sai Ganesh

What company does Sai Ganesh work for?

Sai Ganesh works for Cvs Health

What is Sai Ganesh's role at the current company?

Sai Ganesh's current role is Senior Data Engineer.

Not the Sai Ganesh you were looking for?

Free Chrome Extension

Find emails, phones & company data instantly

Find verified emails from LinkedIn profiles
Get direct phone numbers & mobile contacts
Access company data & employee information
Works directly on LinkedIn - no copy/paste needed
Get Chrome Extension - Free

Download 750 million emails and 100 million phone numbers

Access emails and phone numbers of over 750 million business users. Instantly download verified profiles using 20+ filters, including location, job title, company, function, and industry.