Kavya P

Kavya P Email and Phone Number

Passionate Professional Actively Seeking Exciting New Data Engineer Opportunities @ Charles Schwab
San Francisco
Kavya P's Location
New Market, Maryland, United States, United States
About Kavya P

 Experience in developing and deploying enterprise-based applications using major Hadoop ecosystem components like Map Reduce, YARN, Hive, Pig, HBase, Flume, Sqoop, SparkStreaming, SparkSQL, Storm, Kafka, Oozie, Zookeeper and Cassandra.  Experience of multiple distributions like Cloudera, Hortonworks and Mapr.  Experience in managing Hadoop clusters using Cloudera manager tool.  Capable of processing large sets of structured, semi-structured data and supporting systems application architecture.  Experience in using MapReduce programming model for Batch processing of data stored in HDFS.  Very good experience in Python and Shell scripting.  Extensive experience in developing PIG Latin Scripts and using Hive Query Language for data analytics.

Kavya P's Current Company Details
Charles Schwab

Charles Schwab

View
Passionate Professional Actively Seeking Exciting New Data Engineer Opportunities
San Francisco
Website:
schwab.com
Company phone:
800-435-4000
Kavya P Work Experience Details
  • Charles Schwab
    Data Engineer
    Charles Schwab Sep 2022 - Present
    Westlake, Texas, Us
     Built data APIs and delivery services to support critical operational process, analytical models, and machine learning applications Skilled in Python programming for automation, scripting, and data manipulation tasks. Implemented and maintained PostgreSQL databases, including performance tuning and data replication. Assisted in the development and maintenance of CI/CD pipelines using Bamboo. Experience in building efficient pipelines for moving data between GCP and Azure using Azure Data Factory Worked on migrating data from Teradata to AWS using Python and BI tools like Alteryx Automate the data flow process in the Alteryx from data sources(flat files, Postgress database) to S3 bucket using Python, SQL and Alteryx tool inbuild capabilities. Created BigQuery authorized views for row level security or exposing the data to other teams Built data pipelines in airflow in GCP for ETL related jobs using different airflow operators both old and new operators Writing and optimizing complex SQL queries for data extraction, transformation, and loading, ensuring high performance and data accuracy
  • Truist Bank
    Data Engineer
    Truist Bank Jun 2021 - Aug 2022
     Involved in building data pipelines that extract, classify, merge and deliver new insights on the data.  Worked on python and shell scripting to automate and schedule the workflows to run on Azure.  Experienced in working with UNIX/LINUX environments, writing UNIX shell scripts, Python.  Designed and build Modern data solutions using Azure PaaS service to support visualization of data.  Designed and built ETL pipelines to automate ingestion of structured and unstructured data. Developed Spark jobs using Pyspark and Spark-SQL for data extraction, transformation & aggregation from multiple file formats.  Optimization of Hive queries using best practices and right parameters and using technologies like Hadoop, Python, PySpark.  Developed framework for converting existing PowerCenter mappings and to PySpark(Python and Spark) Jobs.  Worked with Enterprise data support teams to install Hadoop updates, patches, version upgrades as required and fixed problems, which raised after the upgrades.
  • First Republic Bank
    Data Engineer
    First Republic Bank Aug 2019 - May 2021
    San Francisco, Ca, Us
     Deployed the HBase cluster in cloud (AWS) environment with scalable nodes as per the incremental business requirement.  Implemented AWS IAM for managing the user permissions of applications that runs on EC2 instances.  Deployed applications onto AWS lambda with HTTP triggers and integrated them with API Gateway  Developed multiple ETL Hive scripts for data cleansing and transformations for data.  Developed spark applications in python (PySpark) on distributed environment to load huge number of CSV files with different schema in to Hive ORC tables.
  • Abbvie
    Data Engineer
    Abbvie Feb 2017 - Jul 2019
    North Chicago, Illinois, Us
     Perform data validation and reconciliation on raw data in the Data Lake using Spark.  Maintain data quality and completeness as well as check user credentials so that right people are accessing the right data.  Responsible for validation of transactional and profile data from RDBMS which are transformed and loaded to Data Lake using Hadoop Bigdata technologies.  Perform tasks such as writing scripts, calling APIs, write SQL queries, etc.  Schedule the Spark jobs in cluster using Airflow.  Involved in loading data from UNIX file system to HDFS.  Responsible for creating Jenkins pipeline for deployment using Ansible.  Identified opportunities to improve the quality of data, through process and system improvements.
  • Ibing Software Solutions Private Limited
    Data Engineer
    Ibing Software Solutions Private Limited Mar 2015 - Nov 2016
    Hyderabad, Telangana, In
     Created Source to Target Mapping documents, documented business and transformation rules and participated in working sessions and ensured full business participation throughout the process. Created DFD Data Functional Design artifacts that incorporated the process flow visios, S-T Mapping document and all the specifications for proper ETL implementation. Involved in creating the test strategy. Prepared Test plan, test case according to the Source to target mapping document Tested the FSDs and ETL mapping which were developed to load from different source systems in the Teradata Staging /Target areas. Validated the data between the Source to Staging and Staging to Target using the Source to target mapping document as reference. Performing own Data profiling on source to compare with list of valid values while mapping source to target and writing exception cases if necessary. Helped the Infrastructure team in loading the XML's generated by the Informatica into the ALGO Limit management system.
  • Dhruvsoft Services Private Limited
    Business Analyst
    Dhruvsoft Services Private Limited Jun 2013 - Feb 2015
    Hyderabad, Telangana, In
     Prepared interview questionnaires for the users and gathered requirements. Performed Functional and GUI testing to ensure that the user acceptance criteria are met. Examined, analyzed and modeled Test plans Use Case Model cases Test procedures based on RUP methodology. Identified Use cases from the Functional requirements and wrote Use Case Specifications and created business process workflow diagrams Activity diagrams, Sequence diagrams, Collaboration diagrams in VISIO Actively participated in performing testing for thoroughly understanding of the engineering process Worked with SQL and Data Warehousing for arrangement of customer data. Created Use Case scenarios and work flow diagrams after requirements gathering.

Kavya P Education Details

  • Jntuh College Of Engineering Hyderabad
    Jntuh College Of Engineering Hyderabad
    Computer Science

Frequently Asked Questions about Kavya P

What company does Kavya P work for?

Kavya P works for Charles Schwab

What is Kavya P's role at the current company?

Kavya P's current role is Passionate Professional Actively Seeking Exciting New Data Engineer Opportunities.

What schools did Kavya P attend?

Kavya P attended Jntuh College Of Engineering Hyderabad.

Who are Kavya P's colleagues?

Kavya P's colleagues are Rick Mihm, Pamela Smith, Queenie May, Shrikkanth Raghunathan, Ralph Moreno, Blazy Aldana, Chris Elpers.

Free Chrome Extension

Find emails, phones & company data instantly

Find verified emails from LinkedIn profiles
Get direct phone numbers & mobile contacts
Access company data & employee information
Works directly on LinkedIn - no copy/paste needed
Get Chrome Extension - Free

Aero Online

Your AI prospecting assistant

Download 750 million emails and 100 million phone numbers

Access emails and phone numbers of over 750 million business users. Instantly download verified profiles using 20+ filters, including location, job title, company, function, and industry.