Pradeep V Email and Phone Number
I bring over 8 years of IT experience with a strong focus on big data technologies. Proficient in Spark (streaming, SQL) with Scala and Kafka, I excel in data processing on HDFS. My expertise extends to various reporting objects in Tableau, and I have hands-on experience with Flume and NiFi for log file loading into Hadoop.I possess extensive knowledge in troubleshooting HBase Shell/API, Pig, Hive, and MapReduce. A seasoned professional in importing/exporting data between HDFS and RDBMS using Sqoop, I have worked with Cloudera, AWS, Azure, and Hortonworks, managing Hadoop clusters with Cloudera Manager.My skills include a deep understanding of Spark architecture with Databricks, utilization of various Hadoop infrastructures, and the implementation of algorithms for analytics using Cassandra with Spark and Scala. I am well-versed in NoSQL databases like Cassandra and MongoDB and have experience in developing custom UDFs for Pig and Hive.In addition to my expertise in Azure big data technologies, I have a strong background in AWS cloud services, including EC2, S3, Auto Scaling, Glue, Lambda, CloudWatch, CloudFormation, Athena, DynamoDB, and RedShift.My ability to work effectively in cross-functional teams, excellent communication skills, and proficiency in converting Hive/SQL queries into Spark transformations using Spark DataFrames and Scala contribute to my success. I have a creative approach to solving pipeline engineering challenges, coupled with strong experience in core Java, Scala, SQL, PL/SQL, and Restful web services.
Charter Communications
View- Website:
- charter.com
- Employees:
- 33307
-
Lead Data Engineer At Wells FargoCharter CommunicationsBallwin, Mo, Us -
Lead Data EngineerCharter Communications Apr 2023 - PresentDenver, Colorado, United StatesResponsible for maintaining quality reference data in source by performing operations such as cleaning,transformation and ensuring Integrity in a relational environment by working closely with the stakeholders& solution architect. Designed and developed Security Framework to provide fine grained access to objects AWS S3 usingAWS Lambda, DynamoDB. Performed end- to-end Architecture & implementation assessment of various AWS services like AmazonEMR Redshift, S3. Used AWS EMR to transform and move large amounts of data into and out of other AWS data stores anddatabases, such as Amazon Simple Storage Service (Amazon S3) and Amazon DynamoDB. Install and configure Apache Airflow for AWS S3 bucket and created DAGs to run the Airflow. Prepared scripts to automate the ingestion process using Pyspark and Scala as needed through varioussources such as API, AWS S3, Teradata and Redshift. Created multiple scripts to automate ETL/ ELT process using Pyspark from multiple sources. Developed Pyspark scripts utilizing SQL and RDD in spark for data analysis and storing back into S3. Developed Pyspark code to load from stg to hub implementing the business logic. Developed code in Spark SQL for implementing Business logic with python as programming language. Worked on Sequence files, Map side joins, bucketing, partitioning for hive performance enhancement andstorage improvement. Wrote, compiled, and executed programs as necessary using Apache Spark in Scala to perform ETL jobsingested data. -
Senior Azure Data EngineerOptum Labs Inc Feb 2021 - Mar 2023Minnesota, United StatesDeveloped a deep understanding of the data sources, implemented data standards, maintained data quality,and mastered data management. Build Complex distributed systems involving huge amounts of data handling, collecting metrics buildingdata pipeline, and Analytics. Extract Transform and Load data from Sources Systems to Azure Data Storage services using acombination of Azure Data Factory, T-SQL, Spark SQL, and U-SQL Azure Data Lake Analytics. Involved in Data Ingestion to one or more Azure Services - (Azure Data Lake, Azure Storage, Azure SQL,Azure DW) and processing the data in Azure Databricks. Analyze, design, and build Modern data solutions using Azure PaaS service to support visualization ofdata. Understand the current Production state of the application and determine the impact of newimplementation on existing business processes. Implemented data governance practices on Azure, including data security, privacy, and compliance withregulations such as GDPR and HIPAA. Created Pipelines in ADF using Linked Services/Datasets/Pipeline/ to Extract, Transform, and load datafrom different sources like Azure SQL, Blob storage, Azure SQL Data warehouse, write-back tool, andbackward. Involved in developing Spark applications using PySpark and Spark-SQL for data extraction,transformation, and aggregation from multiple file formats for analyzing & transforming the data touncover insights into customer usage patterns. Created several Databricks Spark jobs with PySpark to perform several tables-to-table operations. Create and maintain optimal data pipeline architecture in cloud Microsoft Azure using Data Factory and Azure Databricks -
Senior Big Data EngineerWalmart Apr 2018 - Jan 2021Bentonville, Arkansas, United StatesActively Participated in all phases of the Software Development Life Cycle (SDLC) from implementationto deployment. Responsible for building scalable distributed data solutions using Hadoop. Responsible for Cluster maintenance, adding and removing cluster nodes, Cluster Monitoring andTroubleshooting, Managing, and reviewing data backups & log files. Responsible for managing the test data coming from different sources. Analyzed data using Hadoop components Hive and Pig. Load and transform large sets of structured, semi structured, and unstructured data using Hadoop/Big Data concepts. Involved in importing and exporting the data from RDBMS to HDFS and vice versa using Sqoop. Worked on developing ETL processes (Data Stage Open Studio) to load data from multiple data sourcesto HDFS using FLUME and SQOOP, and performed structural modifications using Map Reduce, HIVE. Developing Spark scripts, UDFS using both Spark DSL and Spark SQL query for data aggregation,querying, and writing data back into RDBMS through Sqoop. Strong understanding of Partitioning, bucketing concepts in Hive and designed both Managed andExternal tables in Hive to optimize performance. Interacted with business partners, Business Analysts, and product owner to understand requirements andbuild scalable distributed data solutions using the Hadoop ecosystem. Developed Spark Streaming programs to process near real time data from Kafka, and process data withboth stateless and stateful transformations. Experience in report writing using SQL Server Reporting Services (SSRS) and creating various types ofreports like drill down, Parameterized, Cascading, Conditional, Table, Matrix, Chart and Sub Reports. Wrote Oozie scripts and set up workflow using Apache Oozie workflow engine for managing andscheduling Hadoop jobs. Used Hive to analyze data ingested into HBase by using Hive-HBase integration and compute variousmetrics for reporting on the dashboard. -
Data Base DeveloperWells Fargo Nov 2016 - May 2018San Francisco Bay AreaGather Requirements that are to be incorporated into the system. Extensively worked on the analysis of Tables in both Legacy Data store and new data store. Extensively worked on the analysis of Columns in mapping tables for both Legacy Data store and newdata store. Initialized utilization of Data Warehouse ETL software during conversion of data to Oracle DB. Developed the complete documentation of the project based on the analysis of tables and Columns. Created DDL scripts to create, alter, drop tables, views, synonyms and sequences. Worked on SQL Tables, Records and Collections. Wrote SQL Procedures, Functions, and Triggers for Insert, Update and Delete transactions and optimizedfor maximum performance. Extensively worked on the Database Triggers, Stored Procedures, Functions and Database Constraints. Developed SQL queries to fetch complex data from different tables in remote databases using databaselinks. Used ETL process to identify the new or the changed data in order to make better decisions in the project. Participated in Performance Tuning of SQL queries using Explain Plan to improve the performance of theapplication. Source data residing in Excel formats are exported to flat files and then accessed via Oracle external tablesin order to load into the staging schema, at which point all source data can be efficiently transformed andmigrated to the target schema. Extracted data from Flat files using SQL*LOADER. Developed Unix Shell Scripts for loading data into the database using SQLCreated partitions on the tables to improve the performance. Participated in application planning, design activities by interacting and collecting requirements from the end users. -
Sql DeveloperHigate Info Systems Private Limited Dec 2013 - Sep 2016Hyderabad, Telangana, IndiaGenerated database monitoring and data validation reports in SQL Server Reporting Service (SSRS). Created partition table for very large database for good performance. Designed dynamic SSIS Packages to transfer data crossing different platforms, validate data duringtransferring, and archived data files for different DBMS. Responsible for writing SQL queries, stored procedures, views, triggers, T-SQL and DTS/SSIS Deployed SSIS packages and Reports to Production Servers. Worked on Varian EMR Integration to Track Data overtime and improve overall quality of care in apractice. Involved in creation of Data Warehouse Cubes in SQL Server Analysis Service (SSAS). Migrated data using ETL Tool in SSI. Maintained system monitoring jobs and provide high availability support. Worked on Migration of packages from DTS using SQL Server Integration Service (SSIS). Participated Databases Architecture and data modeling design. Reported all events and requirements through established reporting mechanisms in SSRS. Generated test data and tested database to meet the functionalities deliverables in the projectdocumentation and specifications. Designed and developed OLAP using MS SQL Server Analysis Services (SSAS). Designed and developed MS SQL Server Reporting Services (SSRS) under SQL Server 2008. Generated periodic reports based on the statistical analysis of the data using SQL Server ReportingServices (SSRS). Created Drill-down, Parameterized and Sub-Reports using RDL and promoted RDLs to Reporting8Service Server (SSRS).
Pradeep V Education Details
-
Mechanical Engineering
Frequently Asked Questions about Pradeep V
What company does Pradeep V work for?
Pradeep V works for Charter Communications
What is Pradeep V's role at the current company?
Pradeep V's current role is Lead Data Engineer at Wells Fargo.
What schools did Pradeep V attend?
Pradeep V attended Gitam Deemed University.
Who are Pradeep V's colleagues?
Pradeep V's colleagues are Jeffrey Tollison, Johnise Mcclure, Andrea Renee Hayes, Jeff Slavin, Zachary Heaster, Delroy Murphy, Theodore Magnus.
Not the Pradeep V you were looking for?
-
1gmail.com
-
Free Chrome Extension
Find emails, phones & company data instantly
Aero Online
Your AI prospecting assistant
Select data to include:
0 records × $0.02 per record
Download 750 million emails and 100 million phone numbers
Access emails and phone numbers of over 750 million business users. Instantly download verified profiles using 20+ filters, including location, job title, company, function, and industry.
Start your free trial