Ashwin Mathur work email
- Valid
Ashwin Mathur personal email
- Valid
NLP Engineer working on the development of applications using Large Language Models (LLMs).Some interesting projects I have worked on:𝐄𝐟𝐟𝐞𝐜𝐭 𝐨𝐟 𝐅𝐞𝐰-𝐒𝐡𝐨𝐭 𝐏𝐫𝐨𝐦𝐩𝐭𝐢𝐧𝐠 𝐚𝐧𝐝 𝐄𝐯𝐚𝐥𝐮𝐚𝐭𝐢𝐨𝐧 𝐨𝐟 𝐋𝐋𝐌𝐬 𝐮𝐬𝐢𝐧𝐠 𝐭𝐡𝐞 𝐄𝐥𝐞𝐮𝐭𝐡𝐞𝐫𝐀𝐈 𝐋𝐋𝐌 𝐇𝐚𝐫𝐧𝐞𝐬𝐬:• Evaluated the generative performance of three language models - OPT, GPTNeo, and Dolly - across benchmark datasets (AI2’s Reasoning Challenge, Adversarial Natural Language Inference, and Winograd Schema Challenge) using various prompt settings: Zero-Shot, One-Shot, Three-Shot, and Five-Shot prompts.• Observed that model performance on all the benchmarks linearly scales with an increase in model size and there is a significant increase in performance as the number of few-shot (in-context) examples increases in the prompt.𝐄𝐟𝐟𝐞𝐜𝐭 𝐨𝐟 𝐎𝐩𝐭𝐢𝐦𝐢𝐳𝐞𝐫 𝐒𝐞𝐥𝐞𝐜𝐭𝐢𝐨𝐧 𝐚𝐧𝐝 𝐇𝐲𝐩𝐞𝐫𝐩𝐚𝐫𝐚𝐦𝐞𝐭𝐞𝐫 𝐓𝐮𝐧𝐢𝐧𝐠 𝐨𝐧 𝐓𝐫𝐚𝐢𝐧𝐢𝐧𝐠 𝐄𝐟𝐟𝐢𝐜𝐢𝐞𝐧𝐜𝐲 𝐚𝐧𝐝 𝐋𝐋𝐌 𝐏𝐞𝐫𝐟𝐨𝐫𝐦𝐚𝐧𝐜𝐞:• Investigated the impact of optimizer selection and associated hyperparameters on model performance during training across diverse tasks.• Evaluated the performance of five different optimizers (AdamW, RMSProp, NAG, SGD with Momentum, and SGD) on various natural language processing tasks such as Sentiment Analysis, Question Answering, and Text Summarization. Analyzed the convergence of the best-performing models on each dataset.• Fine-tuned DistilBERT, BERT, and FinBERT models for Sentiment Analysis on the StockTwits dataset, while DistilBERT, BERT, RoBERTa were fine-tuned for Question Answering on the CoQA dataset. For Text Summarization, BART, DistillBART, and T5 models were fine-tuned on the BillSum dataset.• Empirical observations highlighted that more general optimizers like RMSProp and AdamW consistently performed as well as, if not better than, specialized optimizers like SGD, Nesterov, or Momentum, given appropriately selected hyperparameters.Active contributor to multiple popular open-source projects.𝐆𝐢𝐭𝐇𝐮𝐛: https://github.com/awinml𝐓𝐞𝐜𝐡𝐧𝐢𝐜𝐚𝐥 𝐬𝐤𝐢𝐥𝐥𝐬:• Languages: Python• Machine Learning: Numpy, Pandas, Scikit-learn, Keras, PyTorch• NLP: Hugging Face Transformers, HaystackOpen to interesting conversations and collaboration, happy to connect.𝐄𝐦𝐚𝐢𝐥: ashwinxmathur@gmail.com
Open Source
-
Open Source DeveloperOpen Source Aug 2023 - PresentWork done for the Haystack LLM framework:• Added the support for Llama.cpp to run LLMs.• Added the support for HuggingFace TEI embedding models.• Added the support for INSTRUCTOR embedding models.• Created an integration for Pinecone Vector Database.• Added the support for Diveristy Ranker for reranking tasks.• Added support for evaluation metrics like F1 Score, Exact Match and Semantic Answer Similarity.Github: https://github.com/deepset-ai/haystack/pulls?q=is%3Apr+author%3Aawinml+is%3Aclosed+sort%3Aupdated-deschttps://github.com/deepset-ai/haystack-core-integrations/pulls?q=is%3Apr+author%3Aawinml+is%3Aclosed+sort%3Aupdated-desc
-
Ai Research Intern (Llms)Alphasense Feb 2023 - Jul 2023• Built a Open-Ended Question-Answering Pipeline on Earnings Call Transcripts using Generative Large Language Models (LLMs). The pipeline reduced the time and effort required for investors and analysts to extract actionable information from earnings call transcripts, enabling them to make better investment decisions.• The pipeline used Retrieval Augmented Generation (RAG) to incorporate new information, allowing for easy attribution and minimal hallucination. The pipeline consists of an Embedding Model, Context Retriever, Prompt Generator and a Generative LLM.• Extracted text snippets from each section of the earnings call, chunked them dynamically based on similarity, and created embeddings. The embeddings were stored in a Pinecone vector database.• Experimented with and selected the best SOTA embedding models (SBERT, MPNET, SGPT and INSTRUCTOR) and context retrieval strategies for the pipeline, resulting in significant improvements in accuracy and performance. A combination of dense embedding retrieval and hybrid retrieval strategies gave the best results. The INSTRUCTOR model generated embeddings which resulted in the best context retrieval. • Created different prompt templates for entity extraction and question-answering, leveraging weak supervision techniques to dynamically generate few-shot examples. Prompt templates were based on ideas from prompt templates used in LLM frameworks like Langchain, Llama Index, OpenPrompt, and Promptify.• Carried out extensive prompt tuning by iteratively refining prompt formatting, instructions and incorporating few-shot examples.• Experimented with the SOTA instruction-tuned generative LLMs for generating answers: Llama-2, Vicuna, Alpaca, Dolly, FLAN-T5 and GPT-3. The Llama-2 and GPT-3 LLMs generated the most accurate and concise answers.• Evaluated the generated answers on Coverage, Redundancy, and Hallucination, quantitatively comparing text generation performance while ensuring accuracy. -
Open Source ContributorHugging Face Sep 2022 - Jun 2023• Developed API to enable Sequence Classification support for BioGPT.• Added functionality to pre-train Transformer-based Vision models (ViT) without using the Trainer class.• Created a section in the documentation showcasing the integration of HuggingFace Evaluate with the scikit-learn framework.View all contributions here: https://bit.ly/githftrf -
Open Source ContributorScikit-Learn Nov 2022 - May 2023• Created API to expose Out of Bag scores for Gradient Boosting based estimators.• Implemented support for sparse matrices as input to Silhouette Score.• Added multi-class support to Average Precision Score by computing it in a One-Vs-Rest manner for each class.• Improved the description of Logistic Regression in the user guide to highlight that it is a regression model implemented as a classifier and a special case of GLM.• Introduced a new section in the user guide explaining Bernoulli and Categorical Distributions in the context of Generalized Linear Models (GLM).View all contributions here: https://bit.ly/gitsklearn
Ashwin Mathur Education Details
-
9.8 Cgpa (Ranked 1ˢᵗ In Batch) -
9.64 Cgpa -
10 Cgpa
Frequently Asked Questions about Ashwin Mathur
What company does Ashwin Mathur work for?
Ashwin Mathur works for Open Source
What is Ashwin Mathur's role at the current company?
Ashwin Mathur's current role is NLP Developer | Open-Source Developer.
What is Ashwin Mathur's email address?
Ashwin Mathur's email address is as****@****ail.com
What schools did Ashwin Mathur attend?
Ashwin Mathur attended Fergusson College, Fergusson College, Delhi Public School, Pune.
Not the Ashwin Mathur you were looking for?
-
2fourseasons.com, hotmail.com
-
-
Ashwin Mathur
New Delhi -
Free Chrome Extension
Find emails, phones & company data instantly
Aero Online
Your AI prospecting assistant
Select data to include:
0 records × $0.02 per record
Download 750 million emails and 100 million phone numbers
Access emails and phone numbers of over 750 million business users. Instantly download verified profiles using 20+ filters, including location, job title, company, function, and industry.
Start your free trial