Mahi K

Mahi K Email and Phone Number

Actively looking for Big Data Engineer roles. Cloudera,Hadoop,HDFS,AWS,,PIGO,Hive,Impala,Drill,SparkSql,MapReduce,Flume,Sqoop,Oozie,Storm,Docker,Java,Kafka.Spark,Scala,Hbase,ZooKeeper,MySQL, Tableau, Shell Scripting. @ Credit Acceptance
Southfield, MI, US
Mahi K's Location
Southfield, Michigan, United States, United States
About Mahi K

• Around 7 years of experience in IT industry with specialization Analysis, Design, Building applications, providing ETL solutions and Implementation as a Data Engineer• More than 4+ years experience with the Hadoop and related Big Data tools, including Spark, Hive, Kafka, Apache Mesos, Cascading and Hadoop MapReduce using Scala, Python, and Java.Can reach me out at mahita595@gmail.comProgramming Languages: C, C++, SQL, PL/SQL, UML, Python, Scala, R, JavaHadoop Eco System: HDFS, MapReduce, Yarn, Pig, Hive, HBase, Sqoop, Flume, Oozie and TWSFrameworks: Apache Spark, Spring, Hadoop, Databricks, Hive, Sqoop, Ambari, HBase Spring MVC, Hibernate, Struts, Junit, log4JDesign Patterns: Singleton, MVC, DAO, VO, Service locator Java/J2EE Technologies: Servlets, JSP, JSF, Apache, EJB 2.0/3.0, JDBC, RMI, JMS, JNDI.Tools: Maven, log4j, SVN, CVS, GIT, DB Visualizer, Team city, TWS (Tivoli Workload Scheduler), Putty, RStudio, WinSCP, Hortonworks,SQL Server Integration, Analysis Services and Reporting Services (SSIS/SSAS/SSRS) IDE’s: Eclipse, Net Beans, My Eclipse, IntelliJ Web/Application Servers: Apache Tomcat, IBM WebSphere, WebLogicDatabase Technologies: Oracle, MySQL, DB2, Mongo DB, MS Access, IBM Big SQLMethodologies: Agile Scrum, Waterfall modelOperating Systems: Windows, Linux/Unix.

Mahi K's Current Company Details
Credit Acceptance

Credit Acceptance

View
Actively looking for Big Data Engineer roles. Cloudera,Hadoop,HDFS,AWS,,PIGO,Hive,Impala,Drill,SparkSql,MapReduce,Flume,Sqoop,Oozie,Storm,Docker,Java,Kafka.Spark,Scala,Hbase,ZooKeeper,MySQL, Tableau, Shell Scripting.
Southfield, MI, US
Employees:
2230
Mahi K Work Experience Details
  • Credit Acceptance
    Credit Acceptance
    Southfield, Mi, Us
  • Credit Acceptance
    Big Data Engineer
    Credit Acceptance Feb 2020 - Present
    Southfield, Mi, Us
    • Research and recommend suitable technology stack for Hadoop migration considering current enterprise architecture. • Responsible for building scalable distributed data solutions using Hadoop. • Experienced in loading and transforming of large sets of structured, semi structured and unstructured data. • Developed Spark jobs and HiveJobs to summarize and transform data. • Involved in converting Hive/SQL queries into Spark transformations using Sparkdataframes, Scala and Python. • Expertise in implementing SparkScala application using higher order functions for both batch and interactive analysis requirement. • Experienced in developing Spark scripts for data analysis in both python and scala. • Wrote Scala scripts to make spark streaming work with Kafka as part of sparkKafka integration efforts. • Built on-premise data pipelines using kafka and spark for real time data analysis. • Created reports in TABLEAU for visualization of the data sets created and tested native Drill, Impala and Spark connectors. • Implemented Hive complex UDF's to execute business logic with Hive Queries. • Responsible for loading bulk amount of data in HBase using MapReduce by directly creating H-files and loading them. • Developed different kind of custom filters and handled pre-defined filters on HBase data using API. • Evaluated performance of SparkSQL vs IMPALA vs DRILL on offline data as a part of poc. • Worked on solr configuration and customizations based on requirements. • Implemented Spark using Scala and utilizing Data frames and SparkSQLAPI for faster processing of data. • Handled importing data from different data sources into HDFS using Sqoop and performing transformations using Hive, MapReduce and then loading data into HDFS. • Exporting of result set from HIVE to MySQL using Sqoop export tool for further processing. • Collecting and aggregating large amounts of log data using Flume and staging data in HDFS for further analysis.
  • Volkswagen Group Of America
    Data Engineer / Big Data Developer
    Volkswagen Group Of America Aug 2018 - Jan 2020
    • Involve in all phases of SDLC (Software Development Life Cycle) which includes requirement collection, design and analysis, development and deployment of the application.• Architecture and design of business requirements and to make Visio Diagrams for the design and to develop the application and deploy the application in various environments. • Develop Spark 2.1/2.4 Scala component to process the business logic and store the computation results of 10 TB data into HBase database to access the downstream web apps using Big SQL db2 database.• Uploaded and processed more than 10 terabytes of data from various structured and unstructured sources into HDFS using Sqoop and Flume.• Test the developed modules in the application using Junit Library and Junit testing Framework• Analyze structured, unstructured data, and file system data and loading the data to HBase tables based on the project requirement using IBM Big SQL with Sqoop mechanism and processing the data using Spark SQL in-memory computation &processing results to Hive, HBase• Handle importing other enterprise data from different data sources into HDFS using JDBC and Load Hadoop in Big SQL and perform transformations using Spark API to perform necessary transformations and actions on the fly for building the common learner data model which gets the data from upstream in near real time and persists into HBase.• Working with different file structures with different Hive file formats like Text file, Sequence file, ORC file, Parquet and Avro to analyze the data to build data model and reading them from HDFS and processing through parquet files and loading into HBASE tables.
  • Tsys
    Data Engineer / Big Data Developer
    Tsys Jan 2017 - Jul 2018
    Columbus, Ga, Us
    • Used Spark-Streaming APIs to perform necessary transformations and actions on the fly for building the common learner data model which gets the data from Kafka in near real time and Persists into MongoDB. • Developed Spark scripts by using Scala shell commands as per the requirement. • Used Spark API over Hortenworks to perform analytics on data in Hive. • Developed Scala scripts using both Data frames/SQL and RDD/MapReduce in Spark 1.6 for Data Aggregation, queries and writing data back into OLTP system through Sqoop. • Experienced in performance tuning of Spark Applications for setting right Batch Interval time, correct level of Parallelism and memory tuning. • Loaded the data into Spark RDD and do in memory data Computation to generate the Output response. • Optimizing of existing algorithms in Hadoop using SparkContext, Spark-SQL, DataFrames and PairRDD's. • Performed advanced procedures like text analytics and processing, using the in-memory computing capabilities of Spark using Scala.
  • People Tech Group
    Hadoop Developer
    People Tech Group Jan 2015 - Dec 2016
    • Worked on analyzing Hadoop cluster and different Big Dataanalytic tools including Pig, Hive HBase database and SQOOP. Involved in Unit testing and delivered Unit test plans and results documents. • Collected and aggregated large amount of web log data from various sources such as webservers, mobile and network devices using ApacheFlume and stored the data into HDFS for analysis. • Exported data from HDFS environment into RDBMS using Sqoop for report generation and visualization purpose. • Installed Hadoop, Map Reduce, HDFS, and Developed multiple map reduce jobs in PIG and Hive for data cleaning and pre-processing. • Installed, monitored and maintained hardware/software related issues on Linux/Unixsystems. • Investigated, installed and configured software fail-over system for production Linux servers. • Designed, developed, debug, tested and promoted Java/ETL code into various environments from DEV through to PROD.
  • Reliance Communications
    Sql Developer
    Reliance Communications Apr 2013 - Dec 2014
    Navi Mumbai, Maharashtra, In
    • Installed and configured SQL server 2005 and set-up Various Databases.• Installed and configured Oracle databases on Unix and Linux platform.• Created all administration tasks of SQL Server and designed database• Worked with development team in relation to application tuning.• Created tables, views, and indexes.• Developed various stored procedures, cursors, functions, and triggers. • Created tables and provided constraints and user defined rules and defaults.• Used system stored procedures and DBCC commands.• Created Database Maintenance Planner for the Performance of SQL Server which covers Database Integrity Checks, Update Database Statistics and Re-indexing. • Scheduled backups and performed restore operations for system test data depends on testers requirement.

Frequently Asked Questions about Mahi K

What company does Mahi K work for?

Mahi K works for Credit Acceptance

What is Mahi K's role at the current company?

Mahi K's current role is Actively looking for Big Data Engineer roles. Cloudera,Hadoop,HDFS,AWS,,PIGO,Hive,Impala,Drill,SparkSql,MapReduce,Flume,Sqoop,Oozie,Storm,Docker,Java,Kafka.Spark,Scala,Hbase,ZooKeeper,MySQL, Tableau, Shell Scripting..

Who are Mahi K's colleagues?

Mahi K's colleagues are Eric Smith, Scott Barnett, Joshua Price, Aaron Marsh, Praneeth Vellore, Jared Wilburn, Seigner Olympe.

Free Chrome Extension

Find emails, phones & company data instantly

Find verified emails from LinkedIn profiles
Get direct phone numbers & mobile contacts
Access company data & employee information
Works directly on LinkedIn - no copy/paste needed
Get Chrome Extension - Free

Aero Online

Your AI prospecting assistant

Download 750 million emails and 100 million phone numbers

Access emails and phone numbers of over 750 million business users. Instantly download verified profiles using 20+ filters, including location, job title, company, function, and industry.