Prashanth Reddy

Prashanth Reddy Email and Phone Number

Sr Data Engineer, Python, ETL, Spark, Scala, Big Data, DataBricks, AWS Services, Machine learning models, Data Migration, Kafka, Jenkins. @ T-Mobile
bellevue, washington, united states
Prashanth Reddy's Location
Westborough, Massachusetts, United States, United States
About Prashanth Reddy

Technology professional with more 10+ Years of experience in solving complex business problems with data-driven, extremely accurate and goal-oriented solutions.Experience in collecting, analyzing, and interpreting large data sets, developing, and forecasting new models.Hands-on experience in analyzing data using Python, SQL, Microsoft Excel, PySpark, Spark SQL for Data Mining, Data Cleansing and Machine Learning.skilled in statistical programming languages like R, Python, Apache Spark including Big Data technologies such as Hadoop, Spark, and Kafka.Hands-on experience in developing and deploying enterprise-based applications using major Hadoop ecosystem components like MapReduce, YARN, Hive, HBase, Flume, Sqoop, Spark MLlib, Spark GraphX, Spark SQL, Kafka.Hands-on experience in working with various Python IDE’s using PyCharm, PyScripter, Spyder, PyStudio and IDLE.Working knowledge of Hadoop technologies in a variety of contexts for the creation and design of scalable systems. extensive experience with Hadoop ecosystems, such as Spark and Hive for data analysis.Expertise in working with SQL/ NoSQL like MongoDB, MySQL, and PostgreSQL and Snowflake.Used Airflow for scheduling the Hive, Spark, and MapReduce jobsExpertise in various Big Data application phases like Data Ingestion, Data Analytics, and Data Visualization. Hands-on experience in developing ETL data pipelines using PySpark on AWS EMR or AWS Glue. Performing data cleaning, wrangling, munging, and merging from various sources to make the data more meaningful. Converting SQL queries into RDD/Data Frame transformations using Apache Spark with Python and Java.Scheduled Airflow DAGs to run multiple Hive and Pig jobs, which independently run with time and data availabilityExperience developing Airflow workflows for scheduling and orchestrating the ETL processWorked with the Python libraries like NumPy, Matplotlib and Pandas throughout the development lifecycle.Experience in converting Hive/SQL queries into Spark transformations using Spark RDD and Pyspark concepts.hands-on experience in installing and configuring various Python packages to handle database problems and connections with SQL and NoSQL databases, such as MongoDB, My SQL, SQL server.Experienced with version control systems like Git, GitHub, and visual studio to keep the versions and configurations of the code organized.Experienced in writing SQL Queries, Stored procedures, functions, packages, tables, views, triggers using relational database like MySQL, PostgreSQL, and MS SQL server.

Prashanth Reddy's Current Company Details
T-Mobile

T-Mobile

View
Sr Data Engineer, Python, ETL, Spark, Scala, Big Data, DataBricks, AWS Services, Machine learning models, Data Migration, Kafka, Jenkins.
bellevue, washington, united states
Website:
tmobile.com
Employees:
77533
Prashanth Reddy Work Experience Details
  • T-Mobile
    Senior Data Engineer
    T-Mobile Nov 2022 - Present
    Westborough, Massachusetts, United States
    Developed custom Python scripts and libraries to automate data processing, data validation, and data transformation tasks in an AWS environment.Used various AWS services like s3, lambda, step functions to automate whole process and reduce daily manual work.Performed Data Cleaning, features scaling, features engineering using pandas and NumPy libraries in python. Proficient with Big Data technologies including Hadoop, Spark/Pyspark, Hive, and YARN.Specialized in optimizing PySpark data pipelines for efficient ELT operations, enhancing data extraction, transformation, and loading capabilities.Proficient with Airflow for orchestrating complex workflows, ensuring efficient scheduling, and managing data pipeline dependencies.Implemented Hadoop automation system that fetches data from Hadoop cluster and inform platform support team regarding health of Hadoop cluster and status or running jobsto users viamail.Developed a pre-processing job to flatten JSON documents into flat files using Spark Data Frames.Participated in a team that developed NoSQL databases, such as MongoDB, for document POC (proof of concept) storage.Developed and designed an API (RESTful Web Service) for the company’s website.Worked on large sets of Structured and Unstructured data.Consumed XML messages using Kafka and processed the xml file using Spark Streaming to capture UI updates.Adapt in the full development lifecycle, including coding, debugging, performance tuning, and deploying applications in production environments.Used various AWS services like s3, lambda, step functions to automate whole process and reduce daily manual work.Familiar with Agile Scrum methodologies and tools like Jira, Confluence, GitHub, along with Jenkins for automated deployments.Train machine learning models on large datasets using Python, optimizing hyperparameters and evaluating model performance through techniques like cross-validation.Convert SQL queries into Spark SQL code for ETL pipelines.
  • Jpmorganchase
    Senior Data Engineer
    Jpmorganchase May 2020 - Oct 2022
    Boston, Massachusetts, United States
    Created Data Pipelines in AWS Data Pipeline to extract, transform, and load data from different sources like Amazon S3, Amazon RDS, Amazon Redshift, AWS Glue, and Databricks.Processed data in Databricks and utilized various Hive optimization techniques like partitioning, bucketing, and Map join for efficient data processing.Used Databricks to write PySpark code for performing transformations on data in Amazon S3.Scheduled and monitored the scripts using AWS Data Pipeline and created AWS API Gateway to publish the code and call it in AWS Data Pipeline.Used GitHub to push Python scripts from Databricks to AWS CodeCommit repositories and wrote Python scripts to parse XML documents and load data in the database.Performed Hive/SQL queries performing Spark transformations using Spark RDDs and Python and PySpark.Designed custom-built input adapters using Spark, Hive, and AWS Glue to ingest and analyze data (Amazon Redshift, Amazon RDS, MongoDB) into Amazon S3.Designed and executed ETL processes for SAP data, integrating various data sources and transforming data into usable formats using AWS Glue and AWS Lambda.Managed and monitored data pipelines and jobs using AWS Data Pipeline, AWS Glue, and other monitoring tools in the AWS Management Console.Developed Spark applications using PySpark and Spark-SQL for data extraction, transformation, and aggregation from multiple file formats for analyzing and transforming data to uncover insights into customer usage patterns.Created several Databricks Spark jobs with PySpark to perform several tables-to-table operations.Extracted, transformed, and loaded data from source systems to Amazon S3 using a combination of AWS Data Pipeline, SQL, Spark SQL, and U-SQL in AWS Data Lake Analytics.Ingested data in mini-batches and performed data frame transformations on those mini-batches of data using Spark Streaming to perform streaming analytics in Databricks.
  • Value Health
    Data Engineer
    Value Health Nov 2018 - Apr 2020
    Houston, Texas, United States
    Involved in entire lifecycle of the projects including Design, Development, and Deployment, Testing and Implementation and support.Create and maintain optimal data pipeline architecture and build the infrastructure required for optimal extraction, transformation, and loading (ETL) of data from a wide variety of data sources like Salesforce, SQL Server, Oracle using AWS, Spark, Python, Hive, Kafka and other Bigdata technologiesBuild Self-service data pipelines using AWS Services like, SNS, StepFunction, Lambda, Glue, EMR, EC2, Athena, Sage Maker, QuickSight, Redshift etc.Design and implement end-to-end data solutions (storage, integration, processing, and visualization) in AWS.Building models in spark using python in Databricks and Sage Maker Notebooks.Worked on migrating SQL scripts from Redshift and Athena.Python is used to extract data for web scraping.Proficient in writing complex SQL transformations in DBT to clean, aggregate, and prepare data for analysis.Proficient in designing and developing Hive databases using various techniques like bucketing and partitioning to organize data.Proficiency in designing and managing data warehouses using Amazon Redshift.Experience in building CICD pipelines for testing and production environments using Terraform.Experienced in writing live Real-time Processing and core jobs using Spark Streaming with Kafka as a data pipe-line system.Coordinated daily, weekly, bi-weekly, monthly, and quarterly oversight monitoring activities using AWS QuickSight.Worked on POC’s with Apache Spark using Scala to implement spark in project.Implemented Spark using Scala and Spark SQL for faster testing and processing of data.Consumed the data from Kafka using Apache spark.Worked on writing, testing, and debugging SQL code for transformations using Data Generation Tool (dbt).Worked on creating visualizations and dashboards using Power BI reporting tools.
  • Toyota Motor Corporation
    Data Engineer
    Toyota Motor Corporation Apr 2016 - Oct 2018
    Dallas, Texas, United States
    Developed entire frontend and backend modules using Python on Django Web Framework.Used Django framework for application development.Designed and developed the UI of the website using HTML, AJAX, CSS and JavaScriptUsed update strategy to effectively migrate data from source to target.Designed ETL Process using Informatica to load data from Flat Files, and Excel Files to target Oracle Data Warehouse database.Created Informatica mappings using various Transformations like Joiner, Aggregate, Expression, Filter and Update Strategy.Worked on AWS Data pipeline to configure data loads from S3 to into Redshift.Perform Maintenance, including managing Space, Remove Bad Files, Remove Cache Files and monitoring services.Designed and developed Web services using XML and jQuery.Experienced in Agile Methodologies and SCRUM Process.Built various graphs for business decision making using Python matplotlib library.Worked in development of applications especially in UNIX environment and familiar with all its commands.Used NumPy for Numerical analysis for Insurance premium.
  • Nrg
    Data Engineer
    Nrg Sep 2014 - Mar 2016
    Dallas, Texas, United States
    Orchestrated ETL processes using Python and SQL through SSIS, facilitating the seamless extraction, transformation, and loading of diverse datasets into Azure data storage.Employed Pandas and NumPy for comprehensive data manipulation, cleaning, and analysis, enhancing the accuracy and reliability of analytical insights derived from the data.Demonstrated proficiency in big data processing by implementing Hadoop and Spark, showcasing expertise in distributed computing for large-scale datasets.Executed data cleaning techniques in Python and SQL Server, ensuring data integrity by handling missing values and outliers.Developed and maintained ETL pipelines to extract, transform, and load data from various source systems into the Data Warehouse, using tools like AWS Glue, Python, and Spark.Designed and optimized Snowflake data models for the Enterprise Data Warehouse, ensuring high performance and scalability for analytics and reporting needs.Optimized data workflow by efficiently transferring data between Hadoop and SQL Server using Hive and Sqoop, improving overall system performance.Applied DAX functions for advanced data modeling and analysis in Power BI, enhancing the visualization and interpretation of complex business data.Managed version control using Git within VS Code, fostering a collaborative and organized development environment for code tracking and collaboration.Proactively monitored system health using Nagios, identifying and addressing potential data pipeline issues to maintain a smooth and reliable data flow.Integrated Jenkins into the development process, automating data-related tasks and ensuring efficient and continuous integration of code changes.Contributed effectively to an Agile team environment, participating in daily stand-ups and sprints to deliver data solutions aligned with evolving business needs.Tracked and managed project tasks, timelines, and priorities using JIRA, ensuring alignment with Agile methodologies and project goals.
  • Kpit Technology
    Data Engineer
    Kpit Technology May 2013 - Jul 2014
    Worked on both legacy data and new data mostly built around the user experience and grocery inventory available.Performed Data Analysis on target data after transfer to Data Warehouse.Developed and maintained sophisticated data warehousing systems by understanding overall business requirements, system architecture and data flow.Strong technical experience in using Oracle database design and development.Created design documents using Erwin Data modeler (ER diagrams) and MS Visio(Flowcharts).Created mappings to load data from source and target to staging, staging to reporting tables by applying business requirements using Informatica Power Center.Assisted and worked with ETL developers and data modelers to make sure that model being developed is according to the business requirements.Created Multi-Dimensional cubes using Impromptu Query Definition (IQD) files in Power Play Transformer.Worked with ETL team for developing various mappings and workflows in Informatica as required by the design specifications.Created Projects, Models (layers, relations, cardinality, Query subjects, star schema and packages) and Published Packages using Framework manager.Developed unit test cases for the code developed using complex SQL queries.Displayed and explained results using Excel and Power point presentations.Involved in planning and execution of Unit testing and UAT. Involved in end client discussions to make sure UAT go through successfully.Documented all phases of project implementation for future reference and conducting KT sessions.

Prashanth Reddy Education Details

Frequently Asked Questions about Prashanth Reddy

What company does Prashanth Reddy work for?

Prashanth Reddy works for T-Mobile

What is Prashanth Reddy's role at the current company?

Prashanth Reddy's current role is Sr Data Engineer, Python, ETL, Spark, Scala, Big Data, DataBricks, AWS Services, Machine learning models, Data Migration, Kafka, Jenkins..

What schools did Prashanth Reddy attend?

Prashanth Reddy attended Jawaharlal Nehru Technological University Hyderabad (Jntuh).

Who are Prashanth Reddy's colleagues?

Prashanth Reddy's colleagues are Mekete Begashaw, Beau Stephens, Stan Adams, Jose Carlos, Sandra Lacroix, Paul Lapasota, Maritiz Cerbas.

Not the Prashanth Reddy you were looking for?

  • Prashanth Reddy

    Talent Acquisition Lead At Tektree Inc..., Source 🔍 | Screen 🎥 | Hire ✅ | Repeat 🔄 -- We Help It Professionals Find Work. Connect With Us! 👋
    Newark, De
    4
    my3tech.com, gmail.com, 9thnetworks.com, blueera.com
  • Prashanth R.

    Aws Admin | Devops Engineer | Certified Aws Solutions Architect | Terraform Associate Certified | Docker | Kubernetes | Jenkins
    United States
  • Prashanth Reddy

    Actively Looking For Us It Recruiter Roles
    Metuchen, Nj
  • 2
    peoplecloudlabs.com, gmail.com

    2 +141588XXXXX

Free Chrome Extension

Find emails, phones & company data instantly

Find verified emails from LinkedIn profiles
Get direct phone numbers & mobile contacts
Access company data & employee information
Works directly on LinkedIn - no copy/paste needed
Get Chrome Extension - Free

Download 750 million emails and 100 million phone numbers

Access emails and phone numbers of over 750 million business users. Instantly download verified profiles using 20+ filters, including location, job title, company, function, and industry.