Kumar Vaibhav work email
- Valid
- Valid
- Valid
- Valid
- Valid
Kumar Vaibhav personal email
- Valid
Engineering Lead and ML enthusiast with about 13 years of experience in developing and leading products, big data/cloud data pipelines/platforms and ML models in fast paced environments and experience ranging from Big Data stack (hybrid cloud - AWS/Horton Works/Cloudera) to .Net.Extensive experience with Cloud, Big Data and Machine Learning makes me suitable for self driven Cloud/Big data based Data Science, Machine Learning or Platform teams. I've been fortunate to work in small, high velocity teams that basically act as a 'SWAT' team called in to save business millions of dollars on multiple occasions (Xede/OADW/CDR FE/CDR BE/ODDW).In my current role as Staff Software Engineer on NextRoll's Data Platform team I mostly work on optimization projects rewriting some of their legacy, but critical pipelines.In my previous capacity at Optum I lead Optum Life Science's most critical and ambitious data platform team for cloud offering as Senior Engineering Lead enabling $200m - $350m business.Eligible to work in US.Some important links - Github - https://github.com/vaibkvResume in Pdf(outdated) - http://resume.panghal.com/resume2022.pdf Techblog - http://techblog.panghal.comGeneral Blog - http://blog.panghal.comHomepage - http://panghal.comAt The Advisory Board Company (now Optum) - worked on .Net based products, created performance benchmarking frameworks from scratch, worked on prediction problems like Student Graduation Prediction and made Text Matching engines based on statistical heuristics at R&D. Also, lead Java based data acquisition and R&D teams. At Austin office, worked as Senior Software Engineer on the small but agile Big Data Platform team architecting data lakes, cloud based tools for other teams. Well versed with Spark, Scala and AWS, my work involved creating Spark based modules, JDBC based tooling, schema parsers, pseudo trans-compilers for Oracle to Spark code, etc. to enable the platform for all potential consumers. Also worked as Senior Engineer (Big Data Platform team) on hybrid cloud (AWS plus Horton Works) enabling Optum's big data ambitions to move product pipelines to big data and AWS.Worked extensively in small teams as Full Stack engineer for creating .Net based products from scratch at Indus Valley Partners, dealing in Capital Markets and making software for the Hedge Funds industry.
-
Staff Software EngineerR1 Rcm Jan 2024 - PresentMurray, Utah, Us -
Senior Manager Software Engineering / Lead Software EngineerOptum Nov 2022 - Nov 2023Eden Prairie , Mn, Us• Leading design, development, creation and deployment of Platform capabilities for EIP Core team.• Envisioned and supervised our in-house developed Spark Data Loader runtime to Azure, leading to optimal cluster usage and cost optimizations across business lines – EIP Core, Custom and Commercial teams.• Creator, admin and maintainer of main EIP repositories that enable the Platform.• Leading a small and fast Platform & Data team, which was created at my request to solve the Platform gaps across EIP.• Architect and creator of EIP configuration-based framework enabling secure population of data per client for multiple dashboards, secure population of small data per client, central client configuration, persona-based capabilities for schema as well as data deployment enabling isolated testing across business functions. Architecture of EIP also protects Optum’s interests by making sure wrong data can never be delivered to a client by way of mapping registries to client config.• Architect for event driven ADF pipelines for EIP Core using Event Grid and Azure functions.• More than 20 hours of recorded video training sessions for engineers ranging from topics like Unit testing to EIP Architecture. Brought about software development best practices to EIP – Unit testing, strong versioning of artifacts, post and pre-release steps, and many more.• Main interviewer for Platform and EIP Core teams.• Pursuing a patentable idea in the area of NLP. -
Staff Software EngineerNextroll Oct 2021 - Oct 2022San Francisco, California, Us• Leading the Rollworks Data Platform team optimization efforts for it’s next generation platform capabilities including User Attributes Pipeline to run cookie segmentation in 4 hours as opposed to 30• Optimized the domain de-anonymization rate for all customers across Rollworks by 35%• Optimized the Company Attributes Domain Healthcheck job time efficiency by more than 50%. Also, stabilized it so that it never fails• Bringing best-in-class technology upgrades on Cloud capabilities like AWS, EMR, Scala and Spark. Laying Software-first best practices in place and creating the foundations of a unified schema and persistence layer• Leading the efforts to optimize the existing workflows and Spark/MapReduce jobs in terms of time efficiency, uniform metrics, stability for production runs via optimized Spark, EMR configurations, guidelines for setting up cluster hardware and code changes• Mentoring other engineers on how to produce optimized code with right configurations. Part of interview panel for System Design and Algorithms knowledge interviews -
Senior Software Engineering ManagerOptum (Formerly Alere Wellbeing) Aug 2020 - Oct 2021• Leading a fast paced, geographically distributed Big Data/Cloud platform engineering team on Optum Life Science's most critical data protection and cloud workbench offering that delivers to production in record time and is lauded across the organization for elite engineering capabilities.• Joined the team to make things work ('salvage expert' mode continues) and then took over the entire engineering leadership of the team and delivered several key production offerings in record time against impossible deadlines - Workbench Version 1, BYOE Version 1, etc. Further, created the roadmap for team's success.• Highly engaged and hard working team since engineers get to work on challenging problems for the most part which keeps learning and curiosity high. Consistently delivers above 100% on 95%+ sprints. • We don't work by 'big company' rules. This is a fast, innovative and aggressive team and deploys code to production every other day and maintains it's technical superiority which is reflected by how much we are consulted across the organization.• Architecture of Workbench - Datalake architecture, Infrastructure deployment architecture, unique Cohort Builder architecture(notebook deployment based), subset pipelines, innovative ingestion of non splittable format files, design for workbench consumption via notebooks, design of serverless(Athena via Sagemaker Studio notebooks) and cluster based compute for workbench consumption(EMR based via EMR notebooks).• End to end design, development and deployment of long running on-prem linux processes - multithreaded data uploads, multithreaded validation layer, etc.• End to end design and development of most data pipelines (Spark, Scala on EMRs)• Data onboarding from multiple teams, consultation to other cloud teams on best practices• Creating platform for $200m - $350m business enablement and migration• Senior Software Engineering Manager is my official designation
-
Senior Software Engineer (Big Data Platform Team)Optum (Formerly Alere Wellbeing) Mar 2018 - Aug 2020• Got Stock/RSUs for being recognized among company's top performers• Architect of data pipelines, data lake for the Optum De-identified Data Workspace team• Lead a small, fast paced engineering team on ODDW and prod delivery• Converted big and complex Oracle based data pipelines (Optum Analytics Data Warehouse) completely to big data (on prem HortonWorks cluster) in record time of 8-9 months. As part of this wrote several ETLs and optimized many more. This was a 'SWAT' team effort after the earlier teams had failed to do this and had lost 2.5 years in trying to do so! • Oracle to Spark Sql/Scala converter based on ANTLR 4 - A standalone tool used in multiple projects (OADW/CDR FE) to accelerate the conversion of Oracle pipelines to big data. It consumes oracle queries and gives out scaffolded scala files containing the transformed spark sql as well as other components required by our spark based data loader.• End to end handling of third part data called II (Impact Intelligence) to automatically trigger spark based ingestion jobs whenever a relevant json is dropped on the driver VM (edge node) - linux notify triggering python code to kick off spark jobs.• Design and development of 'merge' functionality that merges daily ecdr data with cdr. This was a very quick rewrite of an existing spark application enabling the business to meet it's deadlines - this was again a 'SWAT' team effort after earlier team had failed to deliver after wasting many months!• Design and development of CDR FE (Common Data Repository Front End) as a Spark Application and guidance to offshore team on how to develop it using Platform tools• Design and development of Scala based tools to generate Oracle schema and json based configuration for ETLs for another offshore team (OPADM)• Work on hybrid cloud (AWS and Horton Works) based orchestration for OADW - terraform based code triggering off Lambda to create EMR and run OADW on AWS and then some steps on HortonWorks post that via SQS.
-
Senior Software Engineer (Big Data Platform Team)Advisory Board Nov 2016 - Mar 2018Washington, District Of Columbia, UsJoined Advisory Board, Austin (Texas) office as SSE.My work involved enabling the Big Data Platform and the creation of Data Lake and frameworks/engines enabling it. Part of a small but very effective and fast paced team.Main Contributions - • Creation of Spark based model level de-duplication module.• Various security related development like Hive LDAP, Hive SSL, stunnel over hive, etc• Automated creation and deployment of Hive based datalake schema.• Development and extension of 'Adhoc' ETL framework - any input(to datalake), any output(from output) manifest based framework. This enables many ETL pipelines working off of the datalake.• Development of PgExport - ETL'ing datalake data to Postgres.• Interfacing with other teams based out of DC and Austin to enable them on datalake, guide them on how to use our 'Adhoc' ETL framework, etc.• Successfully showcasing alternate route to populate DHA (Data Hub Analytics - our main product) from datalake, rather than using product data from Sql Server.• Development of SqlCli - Umbrella project for all Postgres/SqlServer related operations from Big Data Platform like creation of dbs, permissions on schema, etc.• Creation of LookupUtils - another Spark based module that fetches third party data and persists in AWS S3 and stands up Hive tables on the same.• Creation and maintenance of Node scripts that fetch spec data for models and transforms them to Json and persist in repo.• Converted full Big Data stack from Spark 1.6 to Spark 2.0.1• Part of engineering team interview panel. -
Lead Data ScientistAbco India Private Ltd Jul 2015 - Oct 2016Official designation being 'Lead Engineer' at R&D, I performed the following duties - • Loaned to company's new effort for building a Java based data acquisition platform, called Xede (for about 5 months). Only one of two engineers to work on this project from India. Xede was a high performing engine for ingesting data files and processing all the way to Postgres. The main thing, however, in building this platform was speed of work. We worked as a very high speed team since the pressure to build soon this was very high, otherwise the company might have lost a huge portion of business very soon. A team had earlier failed to build this wasting about 1.5 years in the process!• Worked on Student Graduation Prediction and used GBM (Gradient Boosted Machines) to achieve good accuracy. Major portion was however, data and feature analysis. Also made a Person Matching Algorithm to find student household records using student data and US credit card data.• Lead the Xede team in India.• Lead the R&D team in India.• Continued to work on CVA and produced dashboards that showed hospitals how much could they have saved on Blood transfusions.• Part of India's interview panel for IITs (specifically IIT-Roorkee and IIT-Kharagpur) and other institutions.• Mentored interns and new engineering recruits.• Responsible for ABCO India's Advanced Data Science training from INSOFE, Hyderabad.• Only engineer/individual from India to attend the firm wide 'Elevate' at Washington, D.C.
-
Senior Software EngineerAbco India Private Ltd Aug 2013 - Jun 2015Joined R&D department of the The Advisory Board - The R&D was a very focussed, small and quick moving team and had many short term to long term projects in it's pipeline. Work areas included new technology PoCs, Product Development, statistical analysis, machine learning and prediction problems, text matching problems, person matching algorithms, etc.Contributions - • Development of Radiology, Labs and Blood codes Normalization text matching engines in just two weeks for the initial version. The matching exercises involved matching of CDM descriptions to CPT descriptions enabling cohort level analysis of item usage.• Development of CVA (Care Variation Analyzer), which used the above mentioned text matching engines and combined patient and other data to apply groupings on various patient populations and analyzing hospital item usage. It should be noted that apart from Premier Research's manual data curation, this kind of automated ability to do item analysis was not present anywhere in the Health care industry. CVA was a .Net MVC based, Angular single page app.• Lead hackathon (CrimsonU or CrimsonUniversity) to create the first cluster of Advisory Board and ported a few CCC's data loaders to Hive/Hadoop, etc.• Part of hackathon (ABC Annual Hackathon) to create end to end reporting framework using R, RHive, Sqoop, Mutt, Hadoop, Sweave, Tex, etc.• Part of IIT-Roorkee and IIT-Kharagpur interview panel and interview panel of ABCO India, in general.• Secured position in top five for Annual Innovaction challenge.• Mentored interns and new engineering recruits.• Held basic statistics, ML and Hadoop training for engineering teams.
-
Software EngineerAbco India Private Ltd Mar 2012 - Jul 2013Software development, feature development and performance enhancement related activities. Worked on .Net based products of The Advisory Board like Crimson Continuum of Care. Also developed PerformanceLabs - a wrapper over MiniProfiler that can be used in test suites with reporting capabilities all embedded in a dll.Contributions - - Development of PerformanceLabs framework- Query optimization for various modules, including Guidelines module (brought down from 1 week and TempDb crashes to half hour)- Development of Service Lines feature and it's associated data loaders- Critical bug fixes
-
Software EngineerIndus Valley Partners Oct 2010 - Mar 2012New York, Ny, UsCreated a product from scratch - Cosmos PriceMaster. PriceMaster is a price reconciliation framework for Hedge Funds and is also able to connect to other IVP products like SecMaster (or Security Master) to fetch securities data.Contributions working in a team of four - • Developed almost all of PriceMaster front end.• Developed the Scalar and Vector Pricing Models library.• Developed ASP.NET backend for major components like Feeds Setup, Data Source Setup, Attributes and Attribute Management, Rule based screen, among others.• PolarisIVP's most popular software offering, Polaris is basically an SSRS based collection of dashboards for showing Portfolio information to Hedge Funds.Contributions working in a team of three - Cross browser compatibility for Polaris custom dashboards, development of SSIS packages for data acquisition, Continuous Integration setup for Polaris using MS Build and CC.Net -
Associate Software EngineerIndus Valley Partners Sep 2009 - Oct 2010New York, Ny, Us• Security Master Cosmos Security Master is a flexible and efficient tool for managing the securities of a Hedge Fund.Contributions - Drag and drop functionalities and backend for Attribute Management and related modulesNormal bug fixes and feature development
Kumar Vaibhav Skills
Kumar Vaibhav Education Details
-
Amity UniversityComputer Science -
St. Mary'S Academy, Meerut Cantt.Mathematics And Computer Science
Frequently Asked Questions about Kumar Vaibhav
What company does Kumar Vaibhav work for?
Kumar Vaibhav works for R1 Rcm
What is Kumar Vaibhav's role at the current company?
Kumar Vaibhav's current role is Engineering Leader @ R1 RCM.
What is Kumar Vaibhav's email address?
Kumar Vaibhav's email address is ku****@****ing.com
What schools did Kumar Vaibhav attend?
Kumar Vaibhav attended Amity University, St. Mary's Academy, Meerut Cantt..
What are some of Kumar Vaibhav's interests?
Kumar Vaibhav has interest in Blogging, Slashdot, Travelling, Civil Rights And Social Action, Horse Riding, Education, Environment, Reading, Com, Science And Technology.
What skills is Kumar Vaibhav known for?
Kumar Vaibhav has skills like Asp.net Mvc, .net, Hadoop, Javascript, Microsoft Sql Server, Core Java, Artificial Intelligence, Natural Language Processing, Hive, Apache Pig, Machine Learning, Data Science.
Free Chrome Extension
Find emails, phones & company data instantly
Aero Online
Your AI prospecting assistant
Select data to include:
0 records × $0.02 per record
Download 750 million emails and 100 million phone numbers
Access emails and phone numbers of over 750 million business users. Instantly download verified profiles using 20+ filters, including location, job title, company, function, and industry.
Start your free trial