Aishwarya S Email and Phone Number
- Over 9 years in Information Technology: Specialized in Big Data using Hadoop, covering Analysis, Design, Development, Testing, Documentation, Deployment, and Integration via SQL and Big Data technologies.- SDLC Expertise: Extensive involvement across system analysis, design, development, testing, and project implementation within SDLC frameworks.-Strong SQL Proficiency: Skilled in Subqueries, CTEs, and complex joins across multiple tables. Proficient in SQL Server Constraints, T-SQL Queries, and Dynamic SQL.-Advanced Excel and Reporting: Hands-on experience with Power View, Power BI, Power Pivot, Excel Data Connection, and SSRS for generating varied reports, including drill-down reports, parameterized reports, and stylish layouts.-VBA Knowledge: Proficient in VBA for data export/import within Microsoft products.-Database Development: Developed complex database objects using PL/SQL, including Stored Procedures, Functions, Packages, and Triggers. Proficient in SQL Meta Tracker.-Business Intelligence and Data Modeling: Experience in Cube designing (SSAS), logical modeling, and designing reports for different forms.-SDLC Involvement: Engaged in end-to-end SDLC phases: requirement gathering, analysis, design, development, and implementation.-RDBMS Understanding: Strong grasp of RDBMS concepts and proficient in writing complex stored procedures, functions, and triggers.-ETL and Data Integration: Expertise in SSIS solutions for ETL, data integration, and migration services. Skilled in merging data from diverse sources and data cleaning for OLTP and OLAP databases.-Big Data and Cloud Proficiency: Understanding of Spark Architecture with Databricks, AWS, Microsoft Azure, and their integration for data engineering purposes.-SSIS Management and Performance Tuning: Experience in managing SSIS packages, performance tuning, and optimization techniques.-Agile Implementation: Practiced Agile principles and tools (e.g., JIRA) for project management, enhancing collaboration, and ensuring continuous improvement.-Team Player: Proactive team member with problem-solving and analytical skills.
Ge Power
View- Website:
- gepower.com
- Employees:
- 29534
-
Senior Data EngineerGe Power May 2021 - PresentSchenectady, New York, United States-Designed, developed, and deployed reports using MS SQL Server (SSRS2008) and SSIS in Visual Studio, creating various report types for State and Local Agency users. Developed parameterized stored procedures for report data retrieval.-Participated in Normalization and De-Normalization of databases, managed clustered/non-clustered indexes to enhance performance. Monitored SQL query performance, executed backups for disaster recovery.-Configured, supported, and maintained AWS EC2 instances, integrated Cloudera distribution, and employed Apache Kafka with Spark Streaming on AWS. Utilized GCP's Cloud Functions, Pub/Sub, Dataflow for serverless data processing and real-time data streaming solutions.-Designed SSIS packages with various configurations, XML, and SQL Server files, performed ETL operations, and supported change requests.-Leveraged GCP's Pub/Sub, Dataflow for real-time data streaming, and processing. Integrated external REST APIs with Spark Streaming.-Proficient in high-performance query building, exporting/importing data using VBA. Wrote conversion scripts (SQL, PL/SQL) for data migration. Deployed reports using Reporting Services, Excel Power Pivot for ad-hoc reporting.-Proficient in querying and manipulating large-scale datasets in BigQuery, understanding data modeling principles, optimizing table structures, partitioning, clustering for efficient query performance. Built end-to-end data pipelines using GCP services for BigQuery data analysis. -
Data EngineerDte Energy Oct 2019 - Apr 2021Detroit, Michigan, United States-Designed AWS architecture, migration, and implemented AWS services like EMR, DynamoDB, Redshift, and Lambda for event processing.-Utilized Amazon EMR for processing Big Data across Hadoop Clusters, focusing on Amazon EC2 and S3. Integrated and transformed diverse data sources using GCP's integration tools.-Worked on Snowflake schemas, batch/streaming data loads via Snow Pipe, and imported/exported data to HDFS/Hive using Sqoop. Contributed to Cloudera Hadoop environment development.-Extensively involved in migrating on-premises data pipelines to GCP and AWS cloud environments. Employed Terraform and Ansible for automation, used Lambda functions, and set up Cloud Watch Events-Validated Sqoop jobs, Shell scripts, and performed data validation. Conducted migration and testing of static/transactional data between core systems.-Created and managed Docker images with multiple microservices, employing ECS, ALB, and Lambda for Docker container orchestration.-Developed Spark scripts in Scala for data transformations, utilized Spark-SQL/streaming, and created custom RDDs for actions on RDDs. Implemented Kafka producers and consumers for the data platform.-Developed migration plans, defined risks, milestones, quality aspects, and collaborated across projects to understand data usage implications. Established best practices for data migration activities.-Conducted advanced procedures including text analytics and processing using Spark's in-memory computing capabilities in Scala. Migrated MapReduce programs to Spark transformations.-Deployed projects on AWS EMR with S3 storage, created CI/CD pipelines using Jenkins, Maven, and Bash scripting. Built regression test suites with Cucumber-Gherkin, Java, Spring DAO, and PostgreSQL.-Connected Redshift to Tableau for dynamic dashboard creation, facilitating analytics teams. Conducted ETL integration, cleansing, and transformations using AWS Glue Spark scripts. -
Data Engineer/Hadoop Spark EngineerKeybank Sep 2016 - Sep 2019Cleveland, Ohio, United States-Extensive experience in Big Data Analytics and design using MapReduce Programming, Spark, Hive, Pig, Sqoop, HBase, Oozie, Impala, and Kafka.-Implemented Hive tuning techniques including partitioning, bucketing, and memory optimization. Worked with various file formats such as Parquet, ORC, JSON, and text files.-Successfully migrated MapReduce programs into Spark transformations using Spark, Scala, and Python (PySpark). Employed Spark SQL for data loading, schema RDD creation, and structured data handling.-Developed solutions for data ingestion and processing utilizing Hadoop, MapReduce, MongoDB, Hive, Oozie, Flume, Sqoop, and Talend, catering to Data-at-Rest.-Explored Spark for performance improvement and optimization in Hadoop, employing various Spark components like Spark Context, Spark SQL, Data Frames, Pair RDDs, and YARN.-Utilized Oozie for managing interdependent Hadoop jobs. Architect and implement BI solutions on Azure using various services like Azure Data Lake, Data Factory, Data Lake Analytics, and HDInsight/Databricks.-Proficient in Azure environment for Hadoop app development, creating Databricks clusters, and handling Azure Blob containers. Designed and implemented database solutions in Azure SQL Data Warehouse and Azure SQL.-Extracted and loaded data from CSV, JSON files in AWS S3 to Snowflake Cloud Data Warehouse and Data Lake environment (MS Azure) using Sqoop. Leveraged Azure Data Factory, T-SQL, Spark SQL, and U-SQL for ETL processes.-Followed best practices for securing network access to cloud applications on Databricks platform. Conducted data validation by comparing record-wise counts between source and destination.-Provided support in bug fixes, memory tuning, schema changes, and historical data loading. Implemented checkpoints and worked with Agile and Kanban methodologies. -
Data EngineerAditya Birla Group May 2014 - May 2016India Extensively worked with Spark-SQL context to create data frames and datasets to preprocess the model data. Migrated the ETL pipelines that run on Informatica to AWS cloud and run the pipelines on AWS EMR clusters. Used Hive to implement data warehouse and stored data into HDFS. Stored data into Hadoop clusters which are set up in AWS EMR. Experienced in loading and transforming large sets of Structured, Semi-Structured and Unstructured data and analyzed them by running Hive queries and Pig scripts. Responsible for data extraction and data ingestion from different data sources into Hadoop Data Lake by creating ETL pipelines using Pig, and Hive. Built pipelines to move hashed and un-hashed data from XML files to Data Lake. Developed and maintained batch data flow using HiveQL and Unix scripting. Specified the cluster size, allocating Resource pool, Distribution of Hadoop by writing the specification texts in JSON File format. Involved in designing the row key in HBase to store Text and JSON as key values in HBase table and designed row key in such a way to get/scan it in a sorted order. Wrote Hive Queries for analyzing data in Hive warehouse using Hive Query Language (HQL).
Aishwarya S Education Details
Frequently Asked Questions about Aishwarya S
What company does Aishwarya S work for?
Aishwarya S works for Ge Power
What is Aishwarya S's role at the current company?
Aishwarya S's current role is Experienced in senior data engineer | Open to Global Opportunities | Cloud Specialist: AWS, Azure | Databricks | Python | SQL | Hadoop | Spark | Kafka.
What schools did Aishwarya S attend?
Aishwarya S attended Jntuh College Of Engineering Hyderabad.
Who are Aishwarya S's colleagues?
Aishwarya S's colleagues are Grzegorz Poltorak, Norbert Ugi-Rácz, Edriss Afli, Chris Cox, Dhananajay Singh, Ray Chupp, Ádám András Fodor.
Not the Aishwarya S you were looking for?
-
-
Aishwarya S
North Brunswick, Nj -
Aishwarya S
San Francisco Bay Area -
Free Chrome Extension
Find emails, phones & company data instantly
Aero Online
Your AI prospecting assistant
Select data to include:
0 records × $0.02 per record
Download 750 million emails and 100 million phone numbers
Access emails and phone numbers of over 750 million business users. Instantly download verified profiles using 20+ filters, including location, job title, company, function, and industry.
Start your free trial