Ashwin Mathur

Ashwin Mathur Email and Phone Number

NLP Developer | Open-Source Developer @
Ashwin Mathur's Location
Pune, Maharashtra, India, India
Ashwin Mathur's Contact Details

Ashwin Mathur work email

Ashwin Mathur personal email

About Ashwin Mathur

NLP Engineer working on the development of applications using Large Language Models (LLMs).Some interesting projects I have worked on:𝐄𝐟𝐟𝐞𝐜𝐭 𝐨𝐟 𝐅𝐞𝐰-𝐒𝐡𝐨𝐭 𝐏𝐫𝐨𝐦𝐩𝐭𝐢𝐧𝐠 𝐚𝐧𝐝 𝐄𝐯𝐚𝐥𝐮𝐚𝐭𝐢𝐨𝐧 𝐨𝐟 𝐋𝐋𝐌𝐬 𝐮𝐬𝐢𝐧𝐠 𝐭𝐡𝐞 𝐄𝐥𝐞𝐮𝐭𝐡𝐞𝐫𝐀𝐈 𝐋𝐋𝐌 𝐇𝐚𝐫𝐧𝐞𝐬𝐬:• Evaluated the generative performance of three language models - OPT, GPTNeo, and Dolly - across benchmark datasets (AI2’s Reasoning Challenge, Adversarial Natural Language Inference, and Winograd Schema Challenge) using various prompt settings: Zero-Shot, One-Shot, Three-Shot, and Five-Shot prompts.• Observed that model performance on all the benchmarks linearly scales with an increase in model size and there is a significant increase in performance as the number of few-shot (in-context) examples increases in the prompt.𝐄𝐟𝐟𝐞𝐜𝐭 𝐨𝐟 𝐎𝐩𝐭𝐢𝐦𝐢𝐳𝐞𝐫 𝐒𝐞𝐥𝐞𝐜𝐭𝐢𝐨𝐧 𝐚𝐧𝐝 𝐇𝐲𝐩𝐞𝐫𝐩𝐚𝐫𝐚𝐦𝐞𝐭𝐞𝐫 𝐓𝐮𝐧𝐢𝐧𝐠 𝐨𝐧 𝐓𝐫𝐚𝐢𝐧𝐢𝐧𝐠 𝐄𝐟𝐟𝐢𝐜𝐢𝐞𝐧𝐜𝐲 𝐚𝐧𝐝 𝐋𝐋𝐌 𝐏𝐞𝐫𝐟𝐨𝐫𝐦𝐚𝐧𝐜𝐞:• Investigated the impact of optimizer selection and associated hyperparameters on model performance during training across diverse tasks.• Evaluated the performance of five different optimizers (AdamW, RMSProp, NAG, SGD with Momentum, and SGD) on various natural language processing tasks such as Sentiment Analysis, Question Answering, and Text Summarization. Analyzed the convergence of the best-performing models on each dataset.• Fine-tuned DistilBERT, BERT, and FinBERT models for Sentiment Analysis on the StockTwits dataset, while DistilBERT, BERT, RoBERTa were fine-tuned for Question Answering on the CoQA dataset. For Text Summarization, BART, DistillBART, and T5 models were fine-tuned on the BillSum dataset.• Empirical observations highlighted that more general optimizers like RMSProp and AdamW consistently performed as well as, if not better than, specialized optimizers like SGD, Nesterov, or Momentum, given appropriately selected hyperparameters.Active contributor to multiple popular open-source projects.𝐆𝐢𝐭𝐇𝐮𝐛: https://github.com/awinml𝐓𝐞𝐜𝐡𝐧𝐢𝐜𝐚𝐥 𝐬𝐤𝐢𝐥𝐥𝐬:• Languages: Python• Machine Learning: Numpy, Pandas, Scikit-learn, Keras, PyTorch• NLP: Hugging Face Transformers, HaystackOpen to interesting conversations and collaboration, happy to connect.𝐄𝐦𝐚𝐢𝐥: ashwinxmathur@gmail.com

Ashwin Mathur's Current Company Details
Open Source

Open Source

NLP Developer | Open-Source Developer
Ashwin Mathur Work Experience Details
  • Open Source
    Open Source Developer
    Open Source Aug 2023 - Present
    Work done for the Haystack LLM framework:• Added the support for Llama.cpp to run LLMs.• Added the support for HuggingFace TEI embedding models.• Added the support for INSTRUCTOR embedding models.• Created an integration for Pinecone Vector Database.• Added the support for Diveristy Ranker for reranking tasks.• Added support for evaluation metrics like F1 Score, Exact Match and Semantic Answer Similarity.Github: https://github.com/deepset-ai/haystack/pulls?q=is%3Apr+author%3Aawinml+is%3Aclosed+sort%3Aupdated-deschttps://github.com/deepset-ai/haystack-core-integrations/pulls?q=is%3Apr+author%3Aawinml+is%3Aclosed+sort%3Aupdated-desc
  • Alphasense
    Ai Research Intern (Llms)
    Alphasense Feb 2023 - Jul 2023
    • Built a Open-Ended Question-Answering Pipeline on Earnings Call Transcripts using Generative Large Language Models (LLMs). The pipeline reduced the time and effort required for investors and analysts to extract actionable information from earnings call transcripts, enabling them to make better investment decisions.• The pipeline used Retrieval Augmented Generation (RAG) to incorporate new information, allowing for easy attribution and minimal hallucination. The pipeline consists of an Embedding Model, Context Retriever, Prompt Generator and a Generative LLM.• Extracted text snippets from each section of the earnings call, chunked them dynamically based on similarity, and created embeddings. The embeddings were stored in a Pinecone vector database.• Experimented with and selected the best SOTA embedding models (SBERT, MPNET, SGPT and INSTRUCTOR) and context retrieval strategies for the pipeline, resulting in significant improvements in accuracy and performance. A combination of dense embedding retrieval and hybrid retrieval strategies gave the best results. The INSTRUCTOR model generated embeddings which resulted in the best context retrieval. • Created different prompt templates for entity extraction and question-answering, leveraging weak supervision techniques to dynamically generate few-shot examples. Prompt templates were based on ideas from prompt templates used in LLM frameworks like Langchain, Llama Index, OpenPrompt, and Promptify.• Carried out extensive prompt tuning by iteratively refining prompt formatting, instructions and incorporating few-shot examples.• Experimented with the SOTA instruction-tuned generative LLMs for generating answers: Llama-2, Vicuna, Alpaca, Dolly, FLAN-T5 and GPT-3. The Llama-2 and GPT-3 LLMs generated the most accurate and concise answers.• Evaluated the generated answers on Coverage, Redundancy, and Hallucination, quantitatively comparing text generation performance while ensuring accuracy.
  • Hugging Face
    Open Source Contributor
    Hugging Face Sep 2022 - Jun 2023
    • Developed API to enable Sequence Classification support for BioGPT.• Added functionality to pre-train Transformer-based Vision models (ViT) without using the Trainer class.• Created a section in the documentation showcasing the integration of HuggingFace Evaluate with the scikit-learn framework.View all contributions here: https://bit.ly/githftrf
  • Scikit-Learn
    Open Source Contributor
    Scikit-Learn Nov 2022 - May 2023
    • Created API to expose Out of Bag scores for Gradient Boosting based estimators.• Implemented support for sparse matrices as input to Silhouette Score.• Added multi-class support to Average Precision Score by computing it in a One-Vs-Rest manner for each class.• Improved the description of Logistic Regression in the user guide to highlight that it is a regression model implemented as a classifier and a special case of GLM.• Introduced a new section in the user guide explaining Bernoulli and Categorical Distributions in the context of Generalized Linear Models (GLM).View all contributions here: https://bit.ly/gitsklearn

Ashwin Mathur Education Details

Frequently Asked Questions about Ashwin Mathur

What company does Ashwin Mathur work for?

Ashwin Mathur works for Open Source

What is Ashwin Mathur's role at the current company?

Ashwin Mathur's current role is NLP Developer | Open-Source Developer.

What is Ashwin Mathur's email address?

Ashwin Mathur's email address is as****@****ail.com

What schools did Ashwin Mathur attend?

Ashwin Mathur attended Fergusson College, Fergusson College, Delhi Public School, Pune.

Not the Ashwin Mathur you were looking for?

Free Chrome Extension

Find emails, phones & company data instantly

Find verified emails from LinkedIn profiles
Get direct phone numbers & mobile contacts
Access company data & employee information
Works directly on LinkedIn - no copy/paste needed
Get Chrome Extension - Free

Aero Online

Your AI prospecting assistant

Download 750 million emails and 100 million phone numbers

Access emails and phone numbers of over 750 million business users. Instantly download verified profiles using 20+ filters, including location, job title, company, function, and industry.