Vinay K

Vinay K Email and Phone Number

Palantir Reporting Specialist @ Equifax
United States
Vinay K's Location
United States, United States
About Vinay K

Over 10 years of experience in data engineering, Quality Assurance, and Product Ownership, specializing in real-time data processing with Apache Kafka, Spark Streaming, and AWS services (Kinesis, Glue, Redshift, S3). Expertise in building scalable data pipelines for batch and streaming architectures.* Proficient in developing automation scripts using Python and Boto3 for AWS resource management and model monitoring with integrated CloudWatch alarms.* Automated infrastructure provisioning with CloudFormation and AWS SDK, ensuring compliant, scalable deployment environments for ML models.* Developed and deployed machine learning models using Azure ML, integrating with Azure Data Factory for data ingestion and transformation pipelines. Applied Azure Machine Learning studio for end-to-end ML Ops workflows.* Skilled in AWS services (EC2, S3, EBS, ELB, RDS, SNS, SQS, VPC, CloudFormation, Redshift) and GCP technologies like Big Query, Cloud SQL, Data Studio, and Pub/Sub for data processing and analysis.* Extensive experience with Azure services (Data Factory, Databricks, Data Lake, Cosmos DB) for both batch and streaming pipelines.* Expertise in Snowflake and Redshift for high-performance data warehousing and analytics.* Managed AWS EC2 and SAS Data Maker, automating ETL pipelines for real-time transformations. Hands-on with AWS EMR for distributed data processing.* Developed CI/CD pipelines using Jenkins, GitLab, and Azure DevOps to streamline deployments and monitor system performance using Datadog.* Implemented CI/CD pipelines using Azure DevOps, Jenkins, GitLab, and GitHub, streamlining deployment processes across Azure and AWS environments. Leveraged Terraform and CloudFormation for Infrastructure as Code, ensuring consistency and compliance in cloud deployments. Utilized Docker and Kubernetes for containerization and orchestration, enhancing scalability for distributed machine learning models.* Automated monitoring using Azure Monitor, AWS CloudWatch, and Datadog, ensuring proactive system management. Applied Boto3 and AWS SDK to script automated AWS resource management, while managing access control through IAM and RBAC in both Azure and AWS.* Experience with Informatica IDQ, DVO, and Ataccama for ensuring data quality across ETL workflows.* Implemented IAM, KMS, and VPC security measures for compliance with GDPR and PCI-DSS.* Extensive experience in data profiling, data observability, and data analysis, with SQL, PL/SQL across various databases (MySQL, MS-SQL, Oracle, DB2, Hadoop

Vinay K's Current Company Details
Equifax

Equifax

View
Palantir Reporting Specialist
United States
Website:
equifax.com
Employees:
17762
Vinay K Work Experience Details
  • Equifax
    Palantir Reporting Specialist
    Equifax
    United States
  • Equifax
    Senior Data Engineer
    Equifax Feb 2023 - Present
    Georgia, United States
    Developed scalable data pipelines using Python, Apache Airflow, Hadoop, and AWS Glue, improving data processing efficiency.* Designed MLOps platforms using AWS SageMaker, Lambda& Glue, automating ML model deployment and management with Docker & Kubernetes.* Automated CI/CD workflows with AWS CodePipeline, CodeCommit, &GitHub, reducing deployment times by 30%.* Administered Snowflake for scalable analytics and optimized data workflows using AWS Glue, Matillion, and SQL for data transformations.* Developed and deployed machine learning models using AzureML, integrating with Azure Data Factory for data ingestion and transformation pipelines. Applied Azure Machine Learning Studio for end-to-end MLOps workflows.* Implemented data quality checks using Ataccama, ensuring integrity across Snowflake environments, and optimized streaming pipelines with Kafka and Spark Streaming.* Managed AWS EMR clusters for real-time data analytics, utilizing Snowflake and Spark.* Automated monitoring using Azure Monitor, AWS CloudWatch, and Datadog, ensuring proactive system management. Applied Boto3 & AWS SDK to script automated AWS resource management, while managing access control through IAM and RBAC in both AzureML and AWS.* Developed RESTful APIs for real-time data analytics on Azure, integrating with Snowflake & optimizing security with IAM, RBAC, and VPC configurations.* Led QA processes in Agile environments, conducting batch testing for ETL workflows.* Implemented CI/CD pipelines with Jenkins, Ansible, and Bamboo, automating deployments and improving system reliability with ELK Stack and Datadog APM.* Enhanced system performance using AWS Glue, Kinesis, Lambda, and Kafka, optimizing data pipelines for improved speed and reduced latency.* Managed data security and compliance(e.g., GDPR, HIPAA) with S3, IAM roles, and VPC configurations.* Developed AI-powered models using AWS Bedrock and built real-time event-driven architectures using Kafka, AWS Lambda, and Kinesis.
  • Ge Power
    Gcp Data Engineer
    Ge Power Nov 2020 - Dec 2022
    Colorado, United States
    Developed and managed big data & analytics using GCP (BigQuery, Dataflow, Dataproc) for data transformation and aggregation. Designed ETL pipelines in Cloud Composer and Matillion, integrating data from Cloud SQL and MySQL, and leveraging MS-SQL, Oracle, Tomcat, and DB2 for optimal database performance.* Led the design and implementation of MLOps platforms using AWS SageMaker, Lambda, and Glue, streamlining machine learning model deployment and management.* Automated ML deployment pipelines utilizing AWS CodePipeline, CodeCommit, and GitHub, ensuring continuous integration and delivery (CI/CD) for data scientists.* Implemented robust automation controls and enforced compliance through infrastructure provisioning with AWS CloudFormation and IAM, enhancing security and governance.* Monitored model performance using AWS CloudWatch and automated resource management with Boto3 and AWS SDK.* Integrated AWS Step Functions to orchestrate complex workflows for machine learning and GenAI models.* Managed containerization and deployment of ML models using Docker and Kubernetes, improving scalability and efficiency across distributed systems.* Collaborated closely with data scientists and stakeholders to deliver an effective ML model deployment pipeline.* Automated ML deployment pipelines utilizing AWS CodePipeline, CodeCommit, and GitHub, ensuring continuous integration and delivery (CI/CD) for data scientists.* Implemented robust automation controls and enforced compliance through infrastructure provisioning with AWS CloudFormation and IAM, enhancing security and governance.* Monitored model performance using AWS CloudWatch and automated resource management with Boto3 and AWS SDK.* Implemented Cloud Platforms & Infrastructure like AWS (EC2, S3, EBS, ELB, RDS, SNS, SQS, VPC, Redshift, CloudFormation, CloudWatch, ELK Stack, VAST, NewT, CodePipeline, Lambda)
  • Dte Energy
    Azure Data Engineer
    Dte Energy Aug 2018 - Oct 2020
    Michigan, United States
    Developed scalable AzureML & Data Engineering data pipelines with Azure Data Factory, Databricks, and optimized Azure Data Lake for efficient high-volume data ingestion and storage. Leveraged Spark, Hive, and MapReduce and Golang for complex transformations and batch processing, enhancing analytics efficiency.* Built Distributed systems & Streaming with Apache Kafka, Flink, and Golang to process integrated NoSQL databases (MongoDB, Cassandra) for real-time, low-latency data solutions. Orchestrated data synchronization using Sqoop, Flume, Blob, and Cosmos DB for scalability and availability.* Implemented CI/CD pipelines using Azure DevOps, Jenkins, GitLab, and GitHub, streamlining deployment processes across Azure and AWS environments. Leveraged Terraform and CloudFormation for Infrastructure as Code (IaC), ensuring consistency and compliance in cloud deployments. Utilized Docker and Kubernetes for containerization and orchestration, enhancing scalability for distributed machine learning models.* Designed responsive Azure & Data Engineering dashboards using React for real-time reporting. Automated data cleaning with Python scripts and developed dynamic analytics dashboards in Power BI to enhance customer insights and product performance.* Implemented CI/CD pipelines with Azure DevOps and Git, facilitating efficient deployments. Configured Matillion for automated ETL workflows, integrating AWS Redshift and Snowflake for optimized data processing.* Led Scrum and Kanban teams to drive projects, using Project Management tools managed Jira for tracking, and ensured timely project delivery. Automated network configuration management to reduce outages and improve infrastructure resilience.* Developed Security & Cloud Functions like Azure Function Apps and WebApps for real-time data processing. Ensured data security and compliance by configuring SSH, TLS and Azure Role-Based Access Control for sensitive information protection.
  • Caterpillar Inc.
    Data Analyst
    Caterpillar Inc. Apr 2016 - Jul 2018
    North Carolina, United States
    * Involved in Data mapping specifications to create and execute detailed system test plans.* Analyzed business requirements, system requirements, data mapping requirement specifications, and responsible for documenting functional requirements and supplementary requirements in Quality Center.* Tested Complex ETL Mappings and Sessions based on business user requirements and business rules to load data from source flat files and RDBMS tables to target tables.* Responsible for different Data mapping activities from Source systems to Teradata.* Responsible for analyzing various data sources such as flat files, ASCII Data, EBCDIC Data, and Relational Data (Oracle, DB2 UDB, MS SQL Server) from various heterogeneous data sources.* Delivered files in various file formatting systems (ex. Excel file, Tab-delimited text, Coma separated text, Pipe delimited text, etc.)* Involved in testing the XML files and checking whether data is parsed and loaded to staging tables.* Executed the SAS jobs in batch mode through UNIX shell scripts.* Involved in code changes for SAS programs and UNIX shell scripts.* Created remote SAS sessions to run the jobs in parallel mode to reduce the extraction time as the datasets were generated simultaneously* Reviewed and modified SAS Programs, to create customized ad-hoc reports, and processed data for publishing business reports.* Tested several stored procedures and wrote complex SQL syntax using the case, having, connect by, etc.* Involved in Teradata SQL Development, Unit Testing, and Performance Tuning to resolve testing issues based on defect reports.* Tested the ETL process both before data validation and after the data validation process. Tested the messages published by the ETL tool and data loaded into various databases.* Experience in creating UNIX scripts for file transfer and file manipulation.
  • Byte Alpha Solutions
    Data Analyst
    Byte Alpha Solutions Jun 2014 - Aug 2015
    Hyderabad, Telangana, India
    * Analyzed Data Analysis & Machine Learning complex data sets using SQL, R, and Python for business decision-making. Applied machine learning techniques for predictive modeling and customer segmentation. Utilized Pandas for data wrangling and large dataset analysis.* Data Integration & Transformation: Integrated and transformed data with Informatica 6.1, ensuring high-quality handling. Cleaned and pre-processed data using Data Flux, Excel, and Jupyter Notebook to achieve accurate analysis-ready data.* Worked on Big Data & Storage with big data technologies such as Hadoop and Spark for large-scale data analytics. Managed and optimized data storage in Hadoop, Oracle 9i, and Teradata for efficient processing and access.* Performed statistical analysis and predictive modeling with SAS and SPSS to forecast trends and support strategic planning.* Developed Data Visualization interactive dashboards and reports in Tableau and Power BI to present data insights to stakeholders.* Technical Expertise in complex PL/SQL queries to handle data from various sources, improving data loading and transformation. Ensured data quality and accuracy using Quality Center 7.2 and TOAD.* Collaboration & Documentation with cross-functional teams to understand data needs, translating them into actionable insights. Managed tasks with Jira and created comprehensive technical documentation for data processes.Environment: Databricks, Informatica, 6.1, SQL, Excel, R, Python, Jupyter notebook, Jira, SAS, SPSS, Tableau, Power BI, Hadoop, Data Cleaning, Data Flux, Oracle 9i, Quality Center 7.2, TOAD, Statistical Analysis, PL/SQL, Flat Files, Teradata.

Vinay K Education Details

Frequently Asked Questions about Vinay K

What company does Vinay K work for?

Vinay K works for Equifax

What is Vinay K's role at the current company?

Vinay K's current role is Palantir Reporting Specialist.

What schools did Vinay K attend?

Vinay K attended Osmania University, Hyderabad.

Who are Vinay K's colleagues?

Vinay K's colleagues are Al Foreman, John Paul Alfaro, Tracy Mcarthur, Trevor Bing, Spencer Turnipseed, Fabiun Graham, Kevin Merrick.

Not the Vinay K you were looking for?

Free Chrome Extension

Find emails, phones & company data instantly

Find verified emails from LinkedIn profiles
Get direct phone numbers & mobile contacts
Access company data & employee information
Works directly on LinkedIn - no copy/paste needed
Get Chrome Extension - Free

Aero Online

Your AI prospecting assistant

Download 750 million emails and 100 million phone numbers

Access emails and phone numbers of over 750 million business users. Instantly download verified profiles using 20+ filters, including location, job title, company, function, and industry.