Ankita Agrawal

Ankita Agrawal Email and Phone Number

Senior Hadoop Developer at Juniper Networks @ GSK
Ankita Agrawal's Location
Sunnyvale, California, United States, United States
Ankita Agrawal's Contact Details

Ankita Agrawal work email

Ankita Agrawal personal email

n/a
About Ankita Agrawal

Ankita Agrawal is a Senior Hadoop Developer at Juniper Networks at GSK. She possess expertise in hadoop, hdfs, mapreduce, hbase, apache pig and 34 more skills.

Ankita Agrawal's Current Company Details
GSK

Gsk

View
Senior Hadoop Developer at Juniper Networks
Ankita Agrawal Work Experience Details
  • Gsk
    Senior Hadoop Developer
    Gsk Nov 2013 - Present
    San Francisco Bay Area
    • Worked on the RDBMS system using PL/SQL to create packages, procedures, functions and triggers as per the business requirements.• Involved in ETL, Data Integration and Migration.• Ran ETL jobs on the data in HDFS.• Imported data using Sqoop to load data from HBase to HDFS on a regular basis.• Developing scripts and batch jobs to schedule various Hadoop Programs.• Developed Pig Scripts, Pig UDFs and Hive Scripts, Hive UDFs to load data files into Hadoop• Have written… Show more • Worked on the RDBMS system using PL/SQL to create packages, procedures, functions and triggers as per the business requirements.• Involved in ETL, Data Integration and Migration.• Ran ETL jobs on the data in HDFS.• Imported data using Sqoop to load data from HBase to HDFS on a regular basis.• Developing scripts and batch jobs to schedule various Hadoop Programs.• Developed Pig Scripts, Pig UDFs and Hive Scripts, Hive UDFs to load data files into Hadoop• Have written Hive Queries for data analysis to meet the business requirements.• Creating Hive Tables and working on them using Hive QL.• Importing and exporting data into HDFS from HBase, and vice versa using Sqoop.• Experienced in defining job flows.• Experience with NoSQL database HBase. • Wrote and modified stored procedures to load and modifying of data according to business rule changes.• Involved in creating Hive Tables, loading the data and writing Hive Queries that will run internally in a map reduce way.• Developed a custom file system plugin for Hadoop to access files on data platform. • The custom file system plugin allows Hadoop Map Reduce programs, HBase, Pig, and Hive to access files directly.• Extracted feeds from social media using Python scripts.• Organized and benchmarked Hadoop/HBase Clusters for internal use. Show less
  • Juniper Networks
    Senior Hadoop/Software Developer
    Juniper Networks Jun 2012 - Oct 2013
    Sunnyvale
    • Installed and configured Apache Hadoop 1.0.1 to test the maintenance of log files in Hadoop cluster• Responsible for all aspects for product development.• Focus on design & coding of new algorithms.• Designed and developed automated, complex, and efficient ETL processes to match multiple large-scale datasets together.• Implemented algorithms tailored to particular business needs and tested on large data sets.• Implemented and deployed large-scale machine learning… Show more • Installed and configured Apache Hadoop 1.0.1 to test the maintenance of log files in Hadoop cluster• Responsible for all aspects for product development.• Focus on design & coding of new algorithms.• Designed and developed automated, complex, and efficient ETL processes to match multiple large-scale datasets together.• Implemented algorithms tailored to particular business needs and tested on large data sets.• Implemented and deployed large-scale machine learning applications and tools.• Developed Java MapReduce programs for the analysis of sample log file stored in cluster.• Developed Map Reduce Programs for data analysis and data cleaning.• Developed PIG Latin scripts for the analysis of semi structured data.• Used Hive and created Hive tables and involved in data loading and writing Hive UDFs.• Troubleshoot cluster and query issues, evaluate query plans, and optimize schemas and queries.• Migration of ETL processes from MySQL to Hive to test the easy data manipulation.• Developed Hive queries to process the data for visualizing. Show less
  • Big Fish Games
    Hadoop/Data Engineer
    Big Fish Games Jan 2011 - Dec 2011
    Greater Seattle Area
    • Installed and configured Hadoop and responsible for maintaining cluster and managing and reviewing Hadoop log files• Involved in creating Hive tables, loading with data and writing hive queries which will run internally in map reduce way.• Built data solutions using open source and proprietary software that enables product and business teams to make data driven decisions.• Built tools and processes to handle performance, scale, availability, usability, accuracy and… Show more • Installed and configured Hadoop and responsible for maintaining cluster and managing and reviewing Hadoop log files• Involved in creating Hive tables, loading with data and writing hive queries which will run internally in map reduce way.• Built data solutions using open source and proprietary software that enables product and business teams to make data driven decisions.• Built tools and processes to handle performance, scale, availability, usability, accuracy and monitoring.• Influence strategy and roadmaps for building scalable data solutions and a scalable data warehouse environment.• Collaborated effectively with other technology teams and architects to solve complex problems spanning their respective areas.• Distributed machine learning and statistical algorithms across multiple platforms including AWS.• Identified emerging trends and opportunities for company growth.• Analyzed complex data sets representing the behavior of online users for strategic and operational business questions.• Acquired and synthesized the data from different external sources into usable and consistent format.• Design, develop, QA and maintain application code. Show less
  • Magellan Health
    Hadoop Developer
    Magellan Health Jun 2010 - May 2011
    Richmond, Virginia Area
    • Resposible for Java Batch, MapReduce programs, Hive queries, testing, debugging, Peer code Review, troubleshooting and maintaining status reports.• Involved in identifying possible ways to improve the efficiency of the system.• Wrote extensive MapReduce Jobs in Java to train the classifier. • Used Scoop to dump data from relational database into HDFS for processing. • Configured flume to capture the news from various sources for testing the classifier. • Wrote MR jobs… Show more • Resposible for Java Batch, MapReduce programs, Hive queries, testing, debugging, Peer code Review, troubleshooting and maintaining status reports.• Involved in identifying possible ways to improve the efficiency of the system.• Wrote extensive MapReduce Jobs in Java to train the classifier. • Used Scoop to dump data from relational database into HDFS for processing. • Configured flume to capture the news from various sources for testing the classifier. • Wrote MR jobs using various Input and Output formats. Also used custom formats whenever necessary. • Developed workflow in Oozie to automate the tasks of loading the data into HDFS and pre-processing, analyzing and training the classifier using MapReduce jobs, Pig jobs and Hive jobs. • Used OpenNLP in the removal of stop words and Stemming of the words. • Used Pig and Hive in the analysis of data. • Created/modified UDF and UDAFs for Hive whenever necessary. • Worked on Hive for further analysis and for generating transforming files from different analytical formats to text files. • Recently started using Mahout for machine learning in identifying a more subtle classifier Show less
  • Amadeus Reseach Lab
    Java/Hadoop Developer
    Amadeus Reseach Lab Aug 2009 - Dec 2009
    Greater Boston Area
    • Installed and configured Apache Hadoop 0.20.1 and responsible for maintaining cluster and managing and reviewing Hadoop log files• Developed MapReduce programs in Java for Data Analysis• Worked on Cloudera to analyze data present on top of HDFS• Worked on large sets of structured, semi-structured and unstructured data• Built tools and processes to handle performance, scale, availability, usability, accuracy and monitoring.• Analyzed complex data sets representing the… Show more • Installed and configured Apache Hadoop 0.20.1 and responsible for maintaining cluster and managing and reviewing Hadoop log files• Developed MapReduce programs in Java for Data Analysis• Worked on Cloudera to analyze data present on top of HDFS• Worked on large sets of structured, semi-structured and unstructured data• Built tools and processes to handle performance, scale, availability, usability, accuracy and monitoring.• Analyzed complex data sets representing the behavior of online users (both buy and sell sides) for strategic and operational business questions.• Developed PIG Latin scripts to play with the data• Involved in creating Hive tables, loading with data and writing hive queries, which will run internally in map, reduce way.• Developed efficient PL/SQL packages for data migration and involved in bulk loads, testing and reports generation• Developed UML diagrams like use cases, class diagrams, interaction diagrams, and activity diagrams were used.• Extensively worked on User Interface for few modules using JSPs and JavaScript.• Built and deployed Java applications into UNIX based environment and produced both unit and functional test results along with release notes. Show less
  • Bank Of The West
    Java Developer
    Bank Of The West Jul 2008 - Aug 2009
    San Francisco Bay Area
    • Involved in requirements analysis and prepared Requirements Specifications document.• Designed implementation logic for core functionalities.• Developed service layer logic for core modules using JSPs and Servlets and involved in integration with presentation layer.• Involved in implementation of presentation layer logic using HTML, CSS, JavaScript and XHTML.• Design of MySQL database to store customer's general and billing details.• Used JDBC connections to store and… Show more • Involved in requirements analysis and prepared Requirements Specifications document.• Designed implementation logic for core functionalities.• Developed service layer logic for core modules using JSPs and Servlets and involved in integration with presentation layer.• Involved in implementation of presentation layer logic using HTML, CSS, JavaScript and XHTML.• Design of MySQL database to store customer's general and billing details.• Used JDBC connections to store and retrieve data from the database.• Development of complex SQL queries and stored procedures to process and store the data.• Used ANT, a build tool to configure application.• Used Web services for sending and getting data from different applications using WSDL and SOAP for submitting the order to downstream, getting the status on the same, etc. Used Tom Cat for creating the Web Services. • Prepared use-case diagrams, class diagrams and sequence diagrams as part of requirement specification documentation.• Developed business services to utilize hibernate service classes that connect to the database and perform the required action.• Involved in unit testing and bug fixing.• Prepared design documents for code developed and defect tracker maintenance. Show less
  • Aviva India
    Java Developer
    Aviva India Apr 2007 - Jun 2008
    India
    • Developed coding standards project life-cycles and software development life-cycles to increase quality and reduce maintenance costs while minimizing schedule impact.• Involved in the analysis, design, and implementation and testing of the project. • Implemented the presentation layer with HTML, XHTML, CSS and JavaScript. • Developed web components using JSPs and Servlets. • Involved in implementation of presentation layer logic using HTML, CSS, JavaScript and… Show more • Developed coding standards project life-cycles and software development life-cycles to increase quality and reduce maintenance costs while minimizing schedule impact.• Involved in the analysis, design, and implementation and testing of the project. • Implemented the presentation layer with HTML, XHTML, CSS and JavaScript. • Developed web components using JSPs and Servlets. • Involved in implementation of presentation layer logic using HTML, CSS, JavaScript and XHTML• Established JDBC connection between database and application. • Designed tables and indexes and optimized them for best performance. • Created SQL tables and indexes and also wrote queries to read/manipulate data. • Involved in bug fixing and unit testing with Junit. • Developed Java programs to process huge JSON files received from marketing team to convert into format standardized for the application• Prepared use-case diagrams, class diagrams and sequence diagrams as part of requirement specification documentation.• Developed business services to utilize hibernate service classes that connect to the database and perform the required action.• Designed, developed and implemented the user interface using HTML, CSS and JQuery. • Prepared design documents for code developed and defect tracker maintenance. Show less

Ankita Agrawal Skills

Hadoop Hdfs Mapreduce Hbase Apache Pig Hive Sqoop Flume Cassandra Apache Zookeeper Oozie Yarn Mahout Java C Python Sql Pl/sql Shell Scripting Html Xml Ajax Soap Amazon Web Services Microsoft Sql Server Mysql Linux Unix Windows Mac Os R Weka Oop Data Structures Algorithms Software Development Etl Data Analysis Machine Learning

Frequently Asked Questions about Ankita Agrawal

What company does Ankita Agrawal work for?

Ankita Agrawal works for Gsk

What is Ankita Agrawal's role at the current company?

Ankita Agrawal's current role is Senior Hadoop Developer at Juniper Networks.

What is Ankita Agrawal's email address?

Ankita Agrawal's email address is an****@****gsk.com

What skills is Ankita Agrawal known for?

Ankita Agrawal has skills like Hadoop, Hdfs, Mapreduce, Hbase, Apache Pig, Hive, Sqoop, Flume, Cassandra, Apache Zookeeper, Oozie, Yarn.

Not the Ankita Agrawal you were looking for?

Free Chrome Extension

Find emails, phones & company data instantly

Find verified emails from LinkedIn profiles
Get direct phone numbers & mobile contacts
Access company data & employee information
Works directly on LinkedIn - no copy/paste needed
Get Chrome Extension - Free

Aero Online

Your AI prospecting assistant

Download 750 million emails and 100 million phone numbers

Access emails and phone numbers of over 750 million business users. Instantly download verified profiles using 20+ filters, including location, job title, company, function, and industry.