Ram M Email and Phone Number
As a Senior Data Engineer with 10 years of experience, I bring extensive knowledge in Big Data technologies, cloud computing, and data engineering. I specialize in Hadoop and its ecosystem, including MapReduce, Hive, Spark, Kafka, and HBase, and have strong skills in Python and Scala for data processing. I’m proficient in managing and optimizing ETL pipelines, leveraging AWS cloud services like EC2, S3, Redshift, Lambda, and Glue for efficient, scalable data solutions.I have hands-on experience with data ingestion tools like Sqoop, Flume, and Kafka for real-time and batch data processing. I’ve also developed custom UDFs for Pig and Hive, enabling seamless integration of business logic into Big Data workflows. Additionally, I am skilled in building machine learning models using Python, Scikit-learn, and PySpark, with experience in Azure Data Lake Analytics, Azure Data Factory, and Snowflake for diverse data processing tasks.My goal is to drive data innovation and deliver actionable insights by crafting and implementing scalable data architectures. Let’s connect if you’re looking for someone with a strong foundation in data engineering, cloud technologies, and Big Data ecosystems.
Charles Schwab
View- Website:
- schwab.com
- Employees:
- 20505
-
Senior Data EngineerCharles Schwab Jan 2023 - PresentWestlake, Texas• Architected end-to-end data pipelines for seamless migration of data from PostgreSQL to AWS S3 and Redshift, leveraging Talend to improve data processing efficiency by 30%.• Implemented serverless architecture using AWS Lambda and Kinesis, significantly reducing infrastructure costs and enhancing system scalability.• Developed real-time data processing applications using Apache Spark and PySpark, reducing data latency and enabling quicker, data-driven decision-making.• Optimized… Show more • Architected end-to-end data pipelines for seamless migration of data from PostgreSQL to AWS S3 and Redshift, leveraging Talend to improve data processing efficiency by 30%.• Implemented serverless architecture using AWS Lambda and Kinesis, significantly reducing infrastructure costs and enhancing system scalability.• Developed real-time data processing applications using Apache Spark and PySpark, reducing data latency and enabling quicker, data-driven decision-making.• Optimized ETL processes in collaboration with cross-functional teams, achieving a 40% reduction in data integration time.• Enhanced data security and governance by designing and managing AWS S3 buckets and implementing strict access control policies.• Collaborated with data science teams to integrate machine learning models into data pipelines, using Python and custom scripts for predictive analytics. Show less -
Senior Data EngineerTarget Feb 2021 - Dec 2022Minneapolis, Minnesota, United States• Led the migration of data pipelines to the Azure cloud platform, utilizing Azure Data Factory, Azure Data Lake, and Databricks, resulting in a 50% reduction in data processing time.• Developed and optimized Spark SQL scripts for large-scale data processing, enhancing query performance by 20% and accelerating data analysis.• Automated job scheduling and batch processing using Oozie on Azure HDInsight, reducing operational overhead and ensuring timely data processing.• Integrated… Show more • Led the migration of data pipelines to the Azure cloud platform, utilizing Azure Data Factory, Azure Data Lake, and Databricks, resulting in a 50% reduction in data processing time.• Developed and optimized Spark SQL scripts for large-scale data processing, enhancing query performance by 20% and accelerating data analysis.• Automated job scheduling and batch processing using Oozie on Azure HDInsight, reducing operational overhead and ensuring timely data processing.• Integrated multiple data sources into a centralized data lake using Hadoop and Hive, improving data accessibility for analytics teams.• Implemented Azure Key Vault and secure authentication protocols to protect sensitive data and ensure compliance with industry standards.• Provided technical leadership and mentorship to junior engineers, fostering a culture of continuous learning and improvement. Show less -
Data EngineerCentene Corporation Apr 2019 - Jan 2021St Louis, Missouri, United States• Developed scalable data ingestion pipelines using Talend, Apache Spark, and Hive, reducing data latency and improving processing capabilities.• Built machine learning models using PySpark and MLlib, providing predictive insights that supported critical business decisions.• Managed AWS S3 data lakes, ensuring reliable and secure storage of large datasets with best practices for data management.• Conducted data transformations using Scala and Spark DataFrames, streamlining data… Show more • Developed scalable data ingestion pipelines using Talend, Apache Spark, and Hive, reducing data latency and improving processing capabilities.• Built machine learning models using PySpark and MLlib, providing predictive insights that supported critical business decisions.• Managed AWS S3 data lakes, ensuring reliable and secure storage of large datasets with best practices for data management.• Conducted data transformations using Scala and Spark DataFrames, streamlining data aggregation and reporting processes.• Developed custom UDFs in Hive to handle complex data transformations, enabling the organization to meet unique business requirements.• Performed performance tuning and troubleshooting of data pipelines, ensuring smooth and efficient data processing systems. Show less -
Hadoop DeveloperAt&T Jun 2017 - Mar 2019Dallas, Texas, United States• Implemented and managed Hadoop clusters, including the setup, configuration, and optimization of HDFS, MapReduce, and other Hadoop components, ensuring high availability and performance.• Developed data pipelines using Pig, Hive, and Sqoop for large-scale ETL tasks, enabling efficient data processing and integration with downstream systems.• Configured and managed Hadoop security protocols using Kerberos and Sentry, ensuring secure access to data and compliance with AT&T’s security… Show more • Implemented and managed Hadoop clusters, including the setup, configuration, and optimization of HDFS, MapReduce, and other Hadoop components, ensuring high availability and performance.• Developed data pipelines using Pig, Hive, and Sqoop for large-scale ETL tasks, enabling efficient data processing and integration with downstream systems.• Configured and managed Hadoop security protocols using Kerberos and Sentry, ensuring secure access to data and compliance with AT&T’s security policies.• Led the migration of Hadoop jobs to a Cloudera-based environment, improving job execution times by 30% and enhancing system stability.• Conducted performance tuning and troubleshooting for Hadoop jobs, reducing errors and improving system efficiency.• Developed MapReduce jobs and Pig Latin scripts to transform and aggregate large datasets, supporting data analysis and reporting. Show less -
Technological AnalystGgk Tech Jan 2014 - Apr 2017India• Worked on Informatica Power enter Designer - Source analyser, Warehouse Designer, Mapping Designer and Transformation developer.• Used various Informatica transformations to recreate data in the data warehouse• Responsible for resolving emergency production issue for the module during the Post Implementation phase. • Had the responsibility for creating the design and implementation documents, effort estimation, planning for coding & implementation, writing and performance tuning… Show more • Worked on Informatica Power enter Designer - Source analyser, Warehouse Designer, Mapping Designer and Transformation developer.• Used various Informatica transformations to recreate data in the data warehouse• Responsible for resolving emergency production issue for the module during the Post Implementation phase. • Had the responsibility for creating the design and implementation documents, effort estimation, planning for coding & implementation, writing and performance tuning the mappings to improve the performance in production environment. • Designed and developed Aggregate, Join, Lookup transformation rules (business rules) to generate consolidated (fact/summary) data identified by dimensions• Used Lookup, Sequence generator, Router and Update Strategy transformations to insert, delete, and update the records for Slowly Changing Dimension tables.• Had the responsibility to lead a 3-member team working in technologies namely Informatica, Unix and Oracle back end Epiphany frontend. Show less
Frequently Asked Questions about Ram M
What company does Ram M work for?
Ram M works for Charles Schwab
What is Ram M's role at the current company?
Ram M's current role is Senior Data Engineer | Expert in Hadoop, Spark, Python, AWS | Data Architecture & Cloud Solutions Specialist | 10 Years in Big Data Analytics & Engineering.
Who are Ram M's colleagues?
Ram M's colleagues are Harjit Singh, Ed Kraus, Bree Morris, Meghan Riley, Arly Manigat, Jamain Simeon, Stacey Worrell.
Not the Ram M you were looking for?
Free Chrome Extension
Find emails, phones & company data instantly
Aero Online
Your AI prospecting assistant
Select data to include:
0 records × $0.02 per record
Download 750 million emails and 100 million phone numbers
Access emails and phone numbers of over 750 million business users. Instantly download verified profiles using 20+ filters, including location, job title, company, function, and industry.
Start your free trial