Vishnu Ks

Vishnu Ks Email and Phone Number

Senior Data Engineer @ Centene Corporation
United States
Vishnu Ks's Location
United States, United States
About Vishnu Ks

With over 9 years of hands-on experience in Data Engineering, I specialize in designing and building data pipelines, optimizing data architectures, and delivering end-to-end data solutions. My expertise spans across Big Data Technologies (Apache Spark, Hadoop), Cloud Platforms (AWS, Redshift, S3, Glue), ETL development, and Data Warehousing (Snowflake, BigQuery), enabling organizations to harness the power of data for actionable insights and decision-making.๐—–๐—ผ๐—ฟ๐—ฒ ๐—ฆ๐—ธ๐—ถ๐—น๐—น๐˜€ & ๐—ง๐—ฒ๐—ฐ๐—ต๐—ป๐—ผ๐—น๐—ผ๐—ด๐—ถ๐—ฒ๐˜€:โ€ข ๐—•๐—ถ๐—ด ๐——๐—ฎ๐˜๐—ฎ & ๐——๐—ถ๐˜€๐˜๐—ฟ๐—ถ๐—ฏ๐˜‚๐˜๐—ฒ๐—ฑ ๐—–๐—ผ๐—บ๐—ฝ๐˜‚๐˜๐—ถ๐—ป๐—ด: Proficient in Apache Spark, Hadoop, and Kafka for scalable data processing, analytics, and real-time data streaming.โ€ข ๐—–๐—น๐—ผ๐˜‚๐—ฑ ๐—ฃ๐—น๐—ฎ๐˜๐—ณ๐—ผ๐—ฟ๐—บ๐˜€ & ๐—œ๐—ป๐—ณ๐—ฟ๐—ฎ๐˜€๐˜๐—ฟ๐˜‚๐—ฐ๐˜๐˜‚๐—ฟ๐—ฒ: Extensive experience working with AWS services including Redshift, S3, Glue, Lambda, and EC2 for cloud-based data management and processing.โ€ข ๐—˜๐—ง๐—Ÿ & ๐——๐—ฎ๐˜๐—ฎ ๐—ฃ๐—ถ๐—ฝ๐—ฒ๐—น๐—ถ๐—ป๐—ฒ ๐——๐—ฒ๐˜ƒ๐—ฒ๐—น๐—ผ๐—ฝ๐—บ๐—ฒ๐—ป๐˜: Expertise in building and optimizing ETL workflows, integrating data from various sources, and ensuring data quality across systems.โ€ข ๐——๐—ฎ๐˜๐—ฎ ๐—ช๐—ฎ๐—ฟ๐—ฒ๐—ต๐—ผ๐˜‚๐˜€๐—ถ๐—ป๐—ด & ๐—”๐—ป๐—ฎ๐—น๐˜†๐˜๐—ถ๐—ฐ๐˜€: In-depth knowledge of Snowflake, BigQuery, and other data warehouse solutions for designing efficient, high-performance data models.โ€ข ๐—ฃ๐—ฟ๐—ผ๐—ด๐—ฟ๐—ฎ๐—บ๐—บ๐—ถ๐—ป๐—ด & ๐—ฆ๐—ฐ๐—ฟ๐—ถ๐—ฝ๐˜๐—ถ๐—ป๐—ด: Strong command of Python, SQL, and Java to develop data applications, automate processes, and build reusable data frameworks.โ€ข ๐——๐—ฎ๐˜๐—ฎ ๐— ๐—ผ๐—ฑ๐—ฒ๐—น๐—ถ๐—ป๐—ด & ๐—”๐—ฟ๐—ฐ๐—ต๐—ถ๐˜๐—ฒ๐—ฐ๐˜๐˜‚๐—ฟ๐—ฒ: Designing scalable and efficient data models and architectures to support business intelligence (BI), analytics, and machine learning (ML) applications.I have successfully contributed to numerous projects across industries, optimizing data processing workflows, implementing cloud-based solutions, and providing valuable insights through big data and advanced analytics. I thrive in collaborative environments and am driven by my passion for solving complex data challenges.Currently, I am seeking Senior Data Engineer, Big Data Developer, or AWS Engineer roles where I can leverage my extensive skills to help organizations build innovative, data-driven solutions.

Vishnu Ks's Current Company Details
Centene Corporation

Centene Corporation

View
Senior Data Engineer
United States
Website:
centene.com
Employees:
33823
Vishnu Ks Work Experience Details
  • Centene Corporation
    Senior Data Engineer
    Centene Corporation
    United States
  • Directv
    Senior Data Engineer
    Directv Jul 2023 - Present
    El Segundo, Ca, Us
    โ€ข Implemented SQL scripts to generate UAT metrics reports for the Teradata-to-Snowflake migration, leveraging Python to automate email alerts, ensuring the timely delivery of actionable insights.โ€ข Facilitated the transfer of structured data from Snowflake to AWS S3 through automated pipelines, ensuring high data quality and compatibility for machine learning model consumption.โ€ข Utilized shell scripting to automate S3 file manipulations, optimizing data structuring and organization for efficient processing in downstream workflows.โ€ข Converted existing Spark and Scala-based architecture code to a SQL-based, metadata-driven framework, streamlining the redesign and improving maintainability.โ€ข Utilized multiprocessing techniques to optimize data processing workflows, significantly reducing execution time for large-scale data tasks by parallelizing computationally intensive operations and enhancing overall system performance.โ€ข Leveraged AWS Lambda for event-driven architecture, automating data pipeline tasks, including data validation, transformation, and triggering downstream processes.โ€ข Developed a robust data quality framework supporting various metric functions and validations, implemented using stored procedures and designed for seamless integration across the data ecosystem.โ€ข Implemented comprehensive data quality checks across multiple use cases and data sources to ensure data accuracy, consistency, and reliability for critical business operations.โ€ข Utilized AWS CloudWatch for comprehensive monitoring and logging of data pipeline performance, enabling proactive issue resolution and system optimization.โ€ข Implemented AWS Lookout for anomaly detection in ad impressions, identifying unusual patterns, and enhancing data quality to improve ad performance and reporting accuracy.โ€ข Designed Airflow DAGs to orchestrate workflows, ensuring efficient scheduling, monitoring, and error handling for seamless data pipeline execution.
  • Centene Corporation
    Senior Data Engineer
    Centene Corporation Apr 2021 - Jun 2023
    Saint Louis, Mo, Us
    โ€ข Involved in writing Spark applications using Python to perform various data cleansing, validation, transformation, and summarization activities according to the requirement.โ€ข Developed multiple POCs using PySpark and deployed them on the Yarn cluster, compared the performance of Spark, with Hive and SQL/Teradata, and developed code in reading multiple data formats on HDFS using PySpark.โ€ข Loaded the data into Spark data frames and performed in-memory data computation to generate the output as per the requirements.โ€ข Worked on AWS Cloud to convert all on-premise, existing processes, and databases to AWS Cloud.โ€ข Designed and developed ETL processes in AWS Glue to migrate campaign data from external sources like S3, ORC/Parquet/Text Files into AWS Redshift.โ€ข Used AWS Redshift, S3, Redshift Spectrum, and Athena services to query large amounts of data stored on S3 to create a virtual data lake without having to go through the ETL process.โ€ข Developed PySpark job to load the CSV files into the S3 buckets and created AWS S3 buckets, performed folder management in each bucket, and managed logs, and objects within each bucket.โ€ข Involved in file movements between HDFS and AWS S3 and extensively worked with S3 bucket in AWS.โ€ข Worked on importing metadata into Hive using Python and migrated existing tables and the data pipeline from legacy to AWS cloud (S3) environment and wrote lambda functions to run the data pipeline in the cloud.โ€ข Extensively worked with partitions, dynamic partitioning, and bucketing tables in Hive, designed both managed and external tables, and worked on the optimization of Hive queries.โ€ข Created generic Databricks notebooks for data cleansing.โ€ข Integrated Snowflake with other data analytics and BI tools such as PowerBI for effective data visualization and reporting.โ€ข Devised complex data architectures in Snowflake, including schema design, and table structures.โ€ข Maintained Snowflake databases, tables, and schemas for optimized performance.
  • Questar Assessment Inc.
    Data Engineer
    Questar Assessment Inc. Aug 2019 - Mar 2021
    Apple Valley, Minnesota, Us
    โ€ข Worked with business users, business analysts, program managers, project managers, system analysts, and quality assurance analysts to review business requirements.โ€ข Created technical specification documents based on business requirements and collaborated with business analysts and other developers to discuss design and architecture. โ€ข Designed, developed, and created ETL (Extract, Transform, and Load) packages using Python, and SQL Server Integration Services (SSIS) to load data into a data warehouse (Microsoft SQL Server), from Excel workbooks and flat files into a database. โ€ข Implemented an application for cleansing and processing terabytes of data using Python and Spark.โ€ข Imported data from AWS S3 into spark RDD and performed transformations and actions on RDD.โ€ข Worked on AWS Cloud to convert all on-premise, existing processes, and databases to AWS Cloud.โ€ข Developed packages using Python, Shell scripting, and XML to automate some of the menial tasks. โ€ข Used Python to write data into JSON files for testing student item-level information. โ€ข Created scripts for data modeling.โ€ข Used Amazon CloudWatch to monitor and track resources on AWS.โ€ข Experience in deploying, configuring, and managing Hadoop clusters using AWS EMR, including customizing instance types, security settings, and software packages.โ€ข Proficient in writing AWS Kinesis applications using SQL-like queries to analyze streaming data in real-time and generate insights.โ€ข Demonstrated expertise in working with Apache Flume to ingest and transfer large volumes of data from various sources to Hadoop-based systems.โ€ข Proven ability to design, configure, and optimize Flume agents for efficient and reliable data ingestion.โ€ข Read the data from various Kafka topics and processed the data using Databricks.โ€ข Familiar with using Airflow's web UI and CLI to monitor pipeline executions, track data lineage, and manage DAGs.โ€ข Optimized queries by adding necessary non-clustered indexes and covering indexes.
  • Groupon
    Data Engineer
    Groupon Apr 2017 - Jul 2019
    Chicago, Il, Us
    โ€ข Responsible for developing a highly scalable and flexible authority engine for all of Groupon's customer data. โ€ข Worked on resetting customer attributes that provide insight about the customer, purchase frequency, marketing channel, and Groupon deal categorization. โ€ข Advocated different sources of data using SQL, Hive, and Scala.โ€ข Fetched gender, age, and other purchase histories from 3rd party data agencies and tried to integrate that data into the existing data store.โ€ข Used Kafka HDFS Connector to export data from Kafka Topic to HDFS files in a variety of formats and integrated with Apache Hive to make data immediately available for SQL querying.โ€ข Normalized the data according to the business needs by modifying the data types, performing data cleansing, and various transformations using Spark and Scala.โ€ขImplemented dynamic partitioning in BigQuery tables and used appropriate file format, and compression techniques to improve the performance of PySpark jobs in GCP Dataproc.โ€ข Built a system for analyzing the column names from all tables and identifying personal information columns of data across on-premise databases to GCP.โ€ข Processed and loaded bound and unbound data from Google Pub/Sub topic to BigQuery using Google Cloud Dataflow with Python.โ€ข Worked with Informatica Cloud for data integration projects, including cloud-to-cloud and cloud-to-on-premise integration.โ€ข Developed and maintained automated build, test, and deployment pipelines using industry-standard CI/CD tools such as Jenkins, Travis CI, or CircleCI.โ€ข Knowledge of Oozie workflow configuration files and parameters.โ€ข Designed and developed job flow using Apache Airflow in GCP.
  • Indium Software
    Etl Developer
    Indium Software Jun 2015 - Mar 2017
    Cupertino, California, Us
    โ€ข Collected business requirements from users and translated them into technical specifications and design documents for development.โ€ข Proficient in designing and implementing ETL workflows using industry-leading tools such as Informatica PowerCenter, Microsoft SSIS, and Talend.โ€ข Hands-on experience in managing metadata, creating and executing test cases, and implementing data lineage and governance using Talend Metadata Manager.โ€ข Designed and developed complex ETL mappings using Informatica PowerCenter to extract, transform and load data from various sources into target systems.โ€ข Responsible for creating ETL packages using SQL server integration services (SSIS).โ€ข Developed logging for ETL load at the package level and task level to log the number of records processed by each package and each task in a package using SSIS.โ€ข Responsible for deploying, scheduling jobs, alerting, and maintaining SSIS packages.โ€ข Designed SSIS packages using lookup transformations, merge joins fuzzy lookups, and derived columns with multiple Dataflow tasks.โ€ข Created complex stored procedures, triggers, functions, indexes, tables, views, and other SQL joins for applications following SQL coding standards.โ€ข Supported and maintained the existing SSRS reports and was responsible for source code fixes.โ€ข Created the reports and data extracts from the SQL server using Excel pivot functionality as per user requirements.โ€ข Worked on query optimization techniques for complex SQL hints statements. โ€ข Used Performance Monitor and Explain Plan for performance tuning to enhance the performance of databases.โ€ข Proficient in writing unit testing code using Unit Test/PyTest and integrating the test code with the build process worked on Python Testing frameworks.

Vishnu Ks Education Details

  • University At Buffalo
    University At Buffalo
    Data Science
  • Dhirubhai Ambani Institute Of Information And Communication Technology
    Dhirubhai Ambani Institute Of Information And Communication Technology
    Information And Communication Technology

Frequently Asked Questions about Vishnu Ks

What company does Vishnu Ks work for?

Vishnu Ks works for Centene Corporation

What is Vishnu Ks's role at the current company?

Vishnu Ks's current role is Senior Data Engineer.

What schools did Vishnu Ks attend?

Vishnu Ks attended University At Buffalo, Dhirubhai Ambani Institute Of Information And Communication Technology.

Who are Vishnu Ks's colleagues?

Vishnu Ks's colleagues are Destiny Hill, Bhavik Patel, Sarah Gunn, Georgia Counsel, Charisse Green, Kenny R., Kia Biller.

Free Chrome Extension

Find emails, phones & company data instantly

Find verified emails from LinkedIn profiles
Get direct phone numbers & mobile contacts
Access company data & employee information
Works directly on LinkedIn - no copy/paste needed
Get Chrome Extension - Free

Aero Online

Your AI prospecting assistant

Download 750 million emails and 100 million phone numbers

Access emails and phone numbers of over 750 million business users. Instantly download verified profiles using 20+ filters, including location, job title, company, function, and industry.