Sailaja Reddy

Sailaja Reddy Email and Phone Number

Senior Data Engineer @ First Citizens Bank
raleigh, north carolina, united states
Sailaja Reddy's Location
Houston, Texas, United States, United States
About Sailaja Reddy

● Data Engineer professional with 9+ years of combined experience in the fields of Data Engineering, Big Data implementations, and Spark technologies.● Experience in Big Data ecosystems using Hadoop, Pig, Hive, HDFS, MapReduce, Sqoop, Storm, Spark, Airflow, Snowflake, Teradata, Flume, Kafka, Yarn, Oozie, and Zookeeper.● High Exposure to Big Data technologies and the Hadoop ecosystem-depth depth understanding of Map Reduce and Hadoop Infrastructure.● Hands-on experience in Unified Data Analytics with Databricks, Databricks Workspace User Interface, Managing Databricks Notebooks, Delta Lake with Python, and Delta Lake with Spark SQL. ● Provide guidance to the development team working on PySpark as an ETL platform. Makes sure that quality standards are defined and met. Optimize the Pyspark jobs to run on Kubernetes Cluster for faster data processing.● Good understanding of Spark Architecture with Databricks, Structured Streaming. Setting Up AWS and Microsoft Azure With Databricks, Databricks Workspace for Business Analytics, Manage Clusters In Databricks, Managing the Machine Learning Lifecycle● Expertise in writing end-to-end Data processing Jobs to analyze data using MapReduce, Spark, and Hive.● Experience with Apache Spark ecosystem using Spark-Core, SQL, Data Frames, RDD, and knowledge of Infopark MLLib.● Experienced in data manipulation using python for loading and extraction as well as with python libraries such as NumPy, SciPy, and Pandas for data analysis and numerical computations.● A solid experience and understanding of designing and operationalization large-scale data and analytics solutions on Snowflake Data Warehouse.● Developing ETL pipelines in and out of the data warehouse using a combination of Python and SnowSQL.● Experience in extracting files from MongoDB through Sqoop and placed in HDFS and processing. ● Worked with NoSQL databases like HBase in creating HBase tables to load large sets of semi-structured data coming from various sources. ● Implemented Cluster for NoSQL tool HBase as a part of POC to address HBase limitations. ● Strong Knowledge of architecture and components of Spark, and efficient in working with Spark Core.● Strong knowledge of Hive analytical functions, extending Hive functionality by writing custom UDFs.● Expertise in writing Map-Reduce Jobs in Python for processing large sets of structured, semi-structured, and unstructured data sets and storing them in HDFS.

Sailaja Reddy's Current Company Details
First Citizens Bank

First Citizens Bank

View
Senior Data Engineer
raleigh, north carolina, united states
Employees:
6304
Sailaja Reddy Work Experience Details
  • First Citizens Bank
    Data Engineer
    First Citizens Bank Feb 2024 - Present
    Pasadena, California, United States
  • Blue Cross Blue Shield Association
    Senior Data Engineer At Bcbs
    Blue Cross Blue Shield Association Apr 2020 - Present
    Chicago, Illinois, United States
    ->Designed and executed batch ETL processes with precision, leveraging Azure Synapse Analytics.->Balanced workloads effectively by utilizing both SQL and Spark pools for diverse data processing needs.->Employed Spark Structured Streaming for real-time data pipelines, enhancing agility in data analytics.->Applied sophisticated physical data modeling constructs, optimizing schemas, tables, views, and indexes.->Implemented star schema and snowflake schema techniques, ensuring a robust foundation for effective data representation.->Developed intricate ETL jobs, specializing in incremental loads and managing change data capture scenarios.->Leveraged PowerShell scripting for end-to-end task automation, encompassing CI/CD pipelines and resource management.->Demonstrated proficiency in decision-making constructs, enhancing automation efficiency in complex workflows.->Implemented Docker for packaging and deploying applications, showcasing mastery in containerization concepts.->Utilized Kubernetes for seamless orchestration, demonstrating adeptness in pod, service, and config map management.->Submitted Hadoop jobs programmatically, showcasing knowledge in HDFS architecture, block storage, and HiveQL.->Implemented robust security measures, encompassing row-level and object-level access controls for data integrity.->Utilized Azure Synapse Studio for proactive monitoring, ensuring pipeline and trigger efficiency.->Integrated Azure Synapse Analytics seamlessly with Power BI, creating visually compelling reports and dynamic dashboards.->Applied visualization tools within Synapse for real-time monitoring, enhancing decision-making capabilities.->Collaborated seamlessly with cross-functional teams, integrating Azure Synapse Analytics into broader data ecosystems.->Implemented cost-effective strategies, optimizing resource utilization, and right-sizing workloads based on demand.
  • Capital One
    Data Engineer
    Capital One Jul 2018 - Mar 2020
    Mclean, Virginia, United States
    • Involved in Requirement gathering, Business Analysis, Design and Development, testing, and implementation of business rules. • Experience in developing Spark applications using Spark-SQL in Databricks for data extraction, transformation, and aggregation from multiple file formats for Analyzing& transforming the data to uncover insights into the customer usage patterns. • Extract Transform and Load data from sources Systems to Azure Data Storage services using a combination of Azure Data factory, T-SQL, Spark SQL, and U-SQL Azure Data Lake Analytics. Data ingestion to one or more Azure services (Azure Data Lake, Azure Storage, Azure SQL, Azure DW) and processing the data in Azure Databricks• Understand business use cases, integration business, write business & technical requirements documents, logic diagrams, process flow charts, and other application-related documents. • Design and develop ETL integration patterns using Python on Spark.• Develop framework for converting existing PowerCenter mappings and to PySpark(Python and Spark) Jobs. Create Pyspark frame to bring data from DB2 to Amazon 53. • Translate business requirements into maintainable software components and understand impact (Technical and Business)• Used Pandas in Python for Data Cleansing and validating the source data. • Designed and developed ETL pipeline in Azure cloud which gets customer data from API and processes it to Azure SQL DB. • Orchestrated all Data pipelines using Azure Data Factory and built a custom alerts platform for monitoring. • Created custom alert queries in Log Analytics and used Webhook actions to automate custom alerts. • Created Databricks Job workflows which extract data from SQL server and upload the files to sftp using pyspark and python. • Used Azure Key vault as a central repository for maintaining secrets and referenced the secrets in Azure Data Factory and also in Databricks notebooks.
  • Black Knight
    Big Data Engineer
    Black Knight Feb 2016 - Jun 2018
    Jacksonville, Florida, United States
    ● Extensively involved in the Installation and configuration of Cloudera Hadoop Distribution.● Implemented advanced procedures like text analytics and processing using in-memory computing capabilities like Apache Spark written in Scala● Developed spark applications for performing large-scale transformations and denormalization of relational datasets.● Have real-time experience with Kafka-Storm on the HDP 2.2 platform for real-time analysis.● Loaded data into the cluster from dynamically generated files using Flume and from relational database management systems using Sqoop.● Created reports for the BI team using Sqoop to export data into HDFS and Hive.● Performed analysis on the unused user navigation data by loading it into HDFS and writing MapReduce jobs. The analysis provided inputs to the new APM front-end developers and the lucent team.● Loading the data from multiple data sources like (SQL, DB2, and Oracle) into HDFS using Sqoop and loading into Hive tables.
  • Mindtree
    Data Analyst
    Mindtree May 2014 - Aug 2015
    Hyderabad, Telangana, India
    ● Understand the data visualization requirements of the Business Users. ● Writing SQL queries to extract data from the Sales data marts as per the requirements. ● Developed Tableau data visualization using Scatter Plots, Geographic maps, Pie Charts, Bar Charts, and Density charts. ● Designed and deploy rich Graphic visualizations with Drill Down and Drop-down menu options and Parameterized using Tableau. ● Created action filters, parameters, and calculated sets for preparing dashboards and worksheets in Tableau. ● Explored traffic data from databases connecting them with transaction data, and presenting as well as writing reports for every campaign, providing suggestions for future promotions. ● Extracted data using SQL queries and transferred it to Microsoft Excel and Python for further analysis. ● Data Cleaning, merging, and exporting the dataset was done in Tableau Prep. ● Data processing and cleaning techniques carried out to reduce text noise, reduce dimensionality to improve the analysis.
  • Seneca Global
    Data Analyst
    Seneca Global Jan 2013 - Apr 2014
    Hyderabad, Telangana, India
    ● Processed data received from vendors and loaded them into the database. The process was carried out every week and reports were delivered on a bi-weekly basis. The extracted data had to be checked for integrity.● Documented requirements and obtained signoffs.● Coordinated between the Business users and development team in resolving issues.● Documented data cleansing and data profiling.● Wrote SQL scripts to meet business requirements.● Analyzed views and produced reports.● Tested cleansed data for integrity and uniqueness.● Automated the existing system to achieve faster and more accurate data loading.● Generated weekly, and bi-weekly reports to be sent to the client business team using business objects and documented them too.● Learned to create Business Process Models.● Ability to manage multiple projects simultaneously tracking them towards varying timelines effectively through a combination of business and technical skills.● Good Understanding of clinical practice management, medical and laboratory billing, and insurance claims with processing with process flow diagrams.● Assisted the QA team in creating test scenarios that cover a day in the life of the patient for Inpatient and Ambulatory workflows.

Sailaja Reddy Education Details

Frequently Asked Questions about Sailaja Reddy

What company does Sailaja Reddy work for?

Sailaja Reddy works for First Citizens Bank

What is Sailaja Reddy's role at the current company?

Sailaja Reddy's current role is Senior Data Engineer.

What schools did Sailaja Reddy attend?

Sailaja Reddy attended Andhra University.

Who are Sailaja Reddy's colleagues?

Sailaja Reddy's colleagues are Md. Sahab Uddin Shaifu, Glenda Zamot, Phil Camden, Rebecca Tibbetts, Virginia Gill, Carl Simmons, Brigitte Kamel.

Not the Sailaja Reddy you were looking for?

Free Chrome Extension

Find emails, phones & company data instantly

Find verified emails from LinkedIn profiles
Get direct phone numbers & mobile contacts
Access company data & employee information
Works directly on LinkedIn - no copy/paste needed
Get Chrome Extension - Free

Aero Online

Your AI prospecting assistant

Download 750 million emails and 100 million phone numbers

Access emails and phone numbers of over 750 million business users. Instantly download verified profiles using 20+ filters, including location, job title, company, function, and industry.