Mukarram Ali Email and Phone Number
Mukarram Ali work email
- Valid
- Valid
Mukarram Ali personal email
• 9+ years of experience in designing and developing data driven solutions – data warehousing, Business Intelligence, analytics, data ingestion - extraction, transformation and loading of data from Transactional databases (OLTP) to Data Warehousing Systems (OLAP) • Analyzing and understanding source systems and business requirements to Design the Enterprise Data warehousing and Business Intelligence Solutions, DataMart and Operational Data Store • Experience in designing & developing applications using Big Data technologies HDFS, Map Reduce, Sqoop, Hive, PySpark & Spark SQL, Hbase, Python, Snowflake, S3 storage, Airflow.• Experience in job workflow scheduling and monitoring tools like Airflow and Autosys.• Experienced in Designing, Developing, Documenting, Testing ETL jobs and mappings in Server and Parallel jobs using Data Stage to populate tables in Data Warehouse and Data marts. • Worked in Production support team for maintaining the mappings, sessions and workflows to load the data in Data Warehouse.• Hands-on experience in using Hadoop ecosystem components like Hadoop, Hive, Pig, Sqoop, HBase, Cassandra, Spark, Spark Streaming, Spark SQL, Oozie, Zookeeper, Kafka, Flume, MapReduce framework, Yarn, Scala, and Hue. • Extensively worked on AWS services like EC2, S3, EMR, RDS, SageMaker, Athena, Glue Data Catalog, RDS(Aurora), Redshift, DynamoDB, and Elastic Cache (Memcached & Redis) & Quick Sight and other services of the AWS family. • Worked closely with the Enterprise Data Warehouse team and Business Intelligence Architecture team to understand repository objects that support the business requirement and process.• Extensive knowledge in working with Azure cloud platforms (HDInsight, Datalake, Databricks, Blob Storage, Data Factory, Synapse, SQL, SQL DB, DWH, and Data Storage Explorer). • Extensive experience in working with NoSQL databases and its integration Dynamo DB, Cosmo DB, Mongo DB, Cassandra, and HBase. • Experience in building data pipelines using Azure Data factory, Azure data bricks and loading data to Azure data Lake, Azure SQL Database, Azure SQL Data warehouse and controlling and granting database access.
Mayo Clinic
View- Website:
- mayoclinic.org
- Employees:
- 10
-
Data Engineer At Mayo ClinicMayo Clinic Jan 2020 - PresentRochester, Minnesota, Us• Architected analytical data pipeline including but not limited to stakeholders’ interviews, data profiling, and extraction process designing from diverse sources, and data load optimization strategies. • Using Kimball four step process (Business process definition, Grain declaration, Fact and Dimension identification), designed Dimensional Data Models for Loan Servicing and Loan Origination with daily transactional facts and Customer, Account, Loan status, Credit profile etc. slowly changing dimensions. • Developed ETL using Microsoft toolset (SSIS, TSQL, MS SQL Server) to implement Type 2 Change Data Capture process for various dimensions. • Worked on development of data ingestion pipelines using ETL tool, Talend & bash scripting with big data technologies including but not limited to Hive, Impala,Spark, Kafka, and Talend. • Experience in developing scalable & secure data pipelines for large datasets. • Gathered requirements for ingestion of new data sources including life cycle, data quality check, transformations, and metadata enrichment. • Supported data quality management by implementing proper data quality checks in data pipelines. • Delivered data engineer services like data exploration, ad-hoc ingestions, and subject-matter-expertise to Data scientists in using big data technologies. • Build machine learning models to showcase big data capabilities using Pyspark and MLlib. • Enhancing Data Ingestion Framework by creating more robust and secure data pipelines. • Implemented data streaming capability using Kafka and Talend for multiple data sources. • Conducted statistical analysis to validate data and interpretations using Python and R, as well as presented Research findings, status reports and assisted with collecting user feedback to improve the processes and tools.Environment: Spark, Redshift, Python, HDFS, Hive, Pig, Sqoop, Scala, Kafka, Shell scripting, Linux, Jenkins, Eclipse, Git, Oozie, Talend, Agile Methodology. -
Data EngineerAbbott Mar 2018 - Dec 2019Abbott Park, Illinois, Us• Implemented and Transaction Monitoring (TM) Model requirements and played key role in data source requirement analysis, ETL Datastage code development and deployment.• Broad understanding of healthcare data like claims clinical data quality metrics and health outcomes..• Install and configure Apache Airflow for S3 bucket and Snowflake data warehouse and created dags to run the Airflow.• Automated resulting scripts and workflow using Apache Airflow and shell scripting to ensure daily execution in production.• Expertise in Creating, Debugging, Scheduling and Monitoring jobs using Airflow.• Presented efficient TM and CARS model enhancement strategies in terms of risk score assignment on various financial activity and profile triggers-based risk factors by applying preferred feature picking methods using Entropy, Mutual Information Gain and Decision tree to streamline High Risk Customer alert processing and SAR/CTR filing.• Played key role in design and implementation of Predictive Analytics based enrichments on CARS, TM model and in process used Bayesian Networks algorithm, coordinated with multi facet business domains and stake holders to gain knowledge regarding classification of Independent and Dependent risk factors in perspective of High Risk customer alert stacking for investigators and Customer Due Diligence(CDD) process.• Implemented TM and CARS model outlier identification algorithms using PySpark involving feature (Risk Factors) engineering, StringIndexer, Vecotrs/Vector Assembler, Linear Regression, Evaluation (RMSE, Feature Correlation check) to detect members’ Unusual Behavior which in effect tune CDD and SAM process through feedback.Environment:- ETL, Tableau, AWS EC2, AWS Lambda, AWS Glue, NoSQL , MongoDB , Python, Django, QT, C++, XML, CSS, JSON, HTML, DHTML, JavaScript and JQuery. -
Data EngineerNationwide Jan 2016 - Feb 2018Us• Implemented CARS (Customer Anti-Money Laundering Risk Scoring) and Transaction Monitoring (TM) Model requirements and played key role in data source requirement analysis, ETL Datastage code development and deployment. • Developed MapReduce programs to parse and filter the raw data store the refined data in partitioned tables in the Greenplum.• Experience on Migrating SQL database to Azure data Lake, Azure data lake Analytics, Azure SQL Database, Data Bricks and Azure SQL Data warehouse and controlling and granting database access and Migrating on premise databases to Azure Data lake store using Azure Data factory.• Experience in Developing Spark applications using Spark - SQL in Databricks for data extraction, transformation, and aggregation from multiple file formats for analyzing & transforming the data to uncover insights into the customer usage patterns.• Good understanding of Spark Architecture including Spark Core, Spark SQL, Data Frames, Spark Streaming, Driver Node, Worker Node, Stages, Executors and Tasks.• Good understanding of Big Data Hadoop and Yarn architecture along with various Hadoop Demons such as Job Tracker, Task Tracker, Name Node, Data Node.• Worked on data cleaning and reshaping, generated segmented subsets using Numpy and Pandas in Python• Developing architecture to move the project from Abinitio to pyspark and scala spark.• Implemented enterprise grade platform (Mark logic) for ETL from mainframe to NoSQL (cassandra).• Building distributed data scalable using Hadoop.• Using Sqoop to load data from HDFS, Hive, MySQL and many other sources on daily bases.• Use Python, Scala programming on a daily basis to perform transformations for applying business logic.Environment: Hadoop, Spark, Hive, Hbase, Abinitio, Scala, Python, ETL, NoSQL (Cassandra), Azure Databricks, HDFS, MapReduce, Azure Data Lake Analytics, Spark SQL, T-SQL, U-SQL, Azure SQL, Sqoop, Apache Airflow. -
Jr. Big Data DeveloperHudda Infotech Apr 2014 - Sep 2015Atlanta, Ga, Us -
Application DeveloperDhruvsoft Services Private Limited Aug 2012 - Mar 2014Hyderabad, Telangana, In
Mukarram Ali Education Details
-
Jawaharlal Nehru Technological University
Frequently Asked Questions about Mukarram Ali
What company does Mukarram Ali work for?
Mukarram Ali works for Mayo Clinic
What is Mukarram Ali's role at the current company?
Mukarram Ali's current role is Actively Seeking For New Opportunities | Data Engineer | Big Data | SQL | AWS | Spark | Hadoop | Azure | PySpark | Kafka | HDFS | Scala | ETL.
What is Mukarram Ali's email address?
Mukarram Ali's email address is al****@****ayo.edu
What schools did Mukarram Ali attend?
Mukarram Ali attended Jawaharlal Nehru Technological University.
Who are Mukarram Ali's colleagues?
Mukarram Ali's colleagues are Carrie Lahner, Kayla Marchese, Pharmd, Bcidp, Michelle Freeman, Michael Abeln, Pmp, Jerry Wille, Charles Hall, Dianne Roberts.
Free Chrome Extension
Find emails, phones & company data instantly
Aero Online
Your AI prospecting assistant
Select data to include:
0 records × $0.02 per record
Download 750 million emails and 100 million phone numbers
Access emails and phone numbers of over 750 million business users. Instantly download verified profiles using 20+ filters, including location, job title, company, function, and industry.
Start your free trial