I'm Manoj, a seasoned professional with over four years of honed expertise in data warehousing engineering, complemented by a rich six-year history as a data engineer. Currently excelling in my role as an Azure Data Engineer. I've earned a reputation for seamlessly designing, implementing, and optimizing data warehouse solutions that empower organizations to translate data into actionable insights.In my current role, I am responsible for designing, developing, and optimizing data pipelines, storage strategies, and compliance frameworks, using Azure Data Factory, Data Lake, Databricks, and Synapse Analytics. I also specialize in implementing Snowflake for advanced data warehousing and analytics, leveraging its features such as dynamic scaling, data sharing, and zero-copy cloning. I have successfully integrated Snowflake with Spark Streaming applications and ETL pipelines, ensuring efficient and scalable data processing. Additionally, I have implemented CI/CD pipelines, real-time analytics, and dynamic resource allocation, contributing to agile and robust data solutions.Expertise:Azure Services: Azure Data Factory, Azure Databricks, Azure Blob Storage, Azure Synapse Analytics, Azure Security Center, Azure Active Directory, Azure Purview, CI/CD, PowerBI, Tableau, Azure SQL Database, Azure Event Hubs, Azure Functions, Logic Apps, Azure Stream Analytics, Monitor, Cosmos DB, Notification Hubs, Queue Storage, Azure Virtual Machines, Azure DevOps, Azure HDInsightHadoop: Hadoop, YARN (Yet Another Resource Negotiator), Hive, Sqoop, MapReduce, Spark, Spark SQL, Kafka, Oozie, Linux.Snowflake: Snowflake, Snowflake-specific optimizations, Snowflake data tablesSQL: T-SQL (Transact-SQL), SQL Server 2008, MySQL, Oracle, Enterprise DBLet's ConnectPhone- (203)433-2934Email: manojkumarr0716@gmail.com I'm truly passionate about harnessing the power of Big Data, Azure services, and Snowflake to architect transformative data solutions. With a focus on innovation and scalability, I'm dedicated to optimizing data processes and unlocking valuable insights for organizations.
-
Media Data EngineerEverest Jan 2023 - PresentUnited StatesExperienced Data Engineer with a strong background in designing and implementing end-to-end data solutions on the Azure cloud platform. Skilled in data ingestion, transformation, and loading (ETL) from diverse sources into Snowflake data warehouse, I specialize in orchestrating robust data processing workflows using Azure Databricks and Apache Spark. With expertise across a wide range of Azure services, including Azure Data Factory, Azure Event Hubs, and Azure Functions, I have consistently delivered scalable and reliable data solutions to meet business needs. Additionally, I have a proven track record in implementing security and authentication measures within Azure environments, leveraging Azure Active Directory (Azure AD) to ensure data protection. Collaborative and results-driven, I excel in communicating with cross-functional teams to align data requirements and deliver impactful insights. -
Media Data EngineerHyundai Nov 2020 - Dec 2022California, United StatesIn my extensive career as a data professional, I've spearheaded transformative initiatives, demonstrating proficiency in designing, implementing, and optimizing data workflows across various cloud platforms. With a focus on real-time analytics and efficient data processing, I've built Spark Streaming applications that provide actionable insights from streaming data. Leveraging Spark scripts in Scala and Spark SQL, I've ensured the effectiveness of data processing tasks and utilized Spark Streaming to divide streaming data into batches for further processing by the Spark engine.A key aspect of my contributions involves the implementation of Continuous Integration/Continuous Deployment (CI/CD) pipelines for seamless project building and deployment in the Hadoop environment. The integration of Spark with Python (PySpark) and Spark SQL has enabled faster testing and processing of data, demonstrating a commitment to efficiency in the development lifecycle.My expertise extends to various Azure services, including Azure Data Factory, Synapse Analytics, SQL Azure, Data Factory, Azure Analysis Services, Application Insights, Azure Monitoring, Key Vault, and Azure Data Lake. I've successfully created batch and streaming pipelines for ETL processes, and enhanced ETL processes by integrating Snowflake, ensuring the efficient loading of cleansed data into Snowflake data tables for improved storage efficiency.The incorporation of Snowflake into the existing data ecosystem has elevated data warehousing capabilities, enabling efficient analytics with advanced features. I've applied Directed Acyclic Graph (DAG) representation within Apache Flink for structured and efficient data management, seamlessly integrating with AWS DynamoDB and other cloud services. Data ingestion into Azure Data Lake Storage (ADLS) Gen2 and loading it into Delta tables after cleansing showcases my commitment to data integrity and seamless integration within the Azure environment.
-
Data EngineerDentaquest Jul 2019 - Oct 2020Fountain Valley, California, United StatesWorked on the development of data ingestion pipelines using ETL tool, Informatica, and bash scripting with big data technologies including but not limited to Hadoop, Hive, Spark, and Kafka.• Experience in developing, scalable and secure data pipelines for large datasets.• Gathered requirements for ingestion of new data sources including life cycle, data quality check, transformations, and metadata enrichment.• Developed data pipeline using Flume, Sqoop, Pig, and Java Map Reduce to ingest customer behavioral data into HDFS for analysis. • Supported data quality management by implementing proper data quality checks in data pipelines.• Enhancing Data Ingestion Framework by creating more robust and secure data pipelines.• Implemented data streaming capability using Kafka and Informatica for multiple data sources.• Involved in Sqoop implementation which helps in loading data from various RDBMS sources to Hadoop systems and vice versa.• Experience with Neo4j Enterprise edition, incorporating advanced features for scalability, security, and management.• Able to maximize the capabilities of the Enterprise edition in enterprise-level projects.• Implemented query caching mechanisms to store and reuse frequently executed queries, reducing the need for redundant processing and improving overall application speed.• Established proficiency in enhancing data processing and analytics performance by effectively utilizing Snowflake's seamless scalability, ensuring the efficient scaling of computing resources as needed.• Worked with multiple storage formats (Avro, Parquet) and databases (Hive) Azure SQL.• Optimizing query performance in Hive using bucketing and partitioning techniques.• Demonstrated proficiency in leveraging Snowflake's multi-cloud capabilities to strategically deploy instances across diverse cloud providers and regions, effectively enhancing redundancy and ensuring optimal data availability. -
Hadoop DeveloperLytx, Inc. May 2018 - Jun 2019California, United States• I Installed, configured, and maintained Apache Hadoop clusters for application development.• Managed Hadoop tools such as HDFS, Hive, Zookeeper, and Sqoop.• Engineered and optimized complex ETL workflows, resulting in a 25% reduction in data processing times.• Integrated Spark Streaming with Kafka for real-time data processing, enabling instant insights and analytics for critical business applications.• Implemented partitioning, dynamic partitions, and buckets in Hive for efficient data storage.• Installed and configured Sqoop to facilitate data import/export between relational databases and Hive.• Administered large Hadoop environments, overseeing cluster setup, and providing support.• Conducted performance tuning and monitoring in enterprise environments to optimize cluster performance.• Monitored and analyzed MapReduce job executions at the task level, optimizing Hadoop cluster components for enhanced performance.• Utilized Python and SAS for extracting, transforming, and loading source data from transaction systems.• Generated reports, insights, and key conclusions from processed data.• Designed and developed data mapping procedures for ETL processes, integrating data using Python programming.• Worked on the extraction, analysis, and loading of data within Hadoop and RDBMS environments.• Created and managed Hive tables for efficient data storage.• Collaborated with cross-disciplinary teams, fostering effective communication between data engineering and server administration teams for seamless project execution.• Participated in data loading processes and developed custom Hive UDFs.• Collaborated with Linux server admin teams to administer server hardware and operating systems.• Configured Spark Streaming to receive real-time data from Kafka and store the stream data in HDFS. -
Data Warehouse DeveloperFidelity Investments Jun 2016 - Apr 2018New York, United States• Created, manipulated, and supported SQL Server databases.• Contributed to data modeling and physical and logical design of the database.• Facilitated integration of the front end with the SQL Server backend.• Developed Stored Procedures, Triggers, Indexes, User-defined Functions, and Constraints to achieve specific results.• Implemented data import and export between servers using tools like Data Transformation Services (DTS).• Wrote T-SQL statements for data retrieval and engaged in performance tuning of T-SQL queries.• Facilitated seamless data transfer between various sources such as MS Excel, MS Access, and Flat Files to SQL Server, utilizing SSIS/DTS for improved data integration.• Supported the team in resolving SQL Reporting Services and T-SQL-related issues.• Proficient in creating various types of reports including Crosstab, Conditional, Drill-down, Top N, Summary, Form, OLAP, and Sub reports.• Designed and developed ETL (Extract, Transform, Load) processes using Informatica PowerCenter to efficiently move and transform data from source systems to the data warehouse.• Developed and tested Windows command files and SQL Server queries for Production database monitoring in 24/7 support.• Developed, monitored, and deployed SSIS packages.• Introduced semantic data modeling techniques to enhance the overall understanding and interpretation of database structures, fostering improved communication between stakeholders and ensuring alignment with business requirements.• Contributed to advanced data modeling techniques, refining both the physical and logical design of databases for improved scalability and performance.• Generated multiple Enterprise reports (SSRS/Crystal/Impromptu) from SQL Server Database (OLTP) and SQL Server Analysis Services Database (OLAP).• Applied custom scripting techniques for data retrieval, offering efficient alternatives within T-SQL statements and ensuring optimized query performance. -
Data Warehouse DeveloperUbs Feb 2014 - May 2016Weehawken, New Jersey, United States• Developed intricate stored procedures, efficient triggers, and necessary functions.• Created indexes and indexed views to enhance database performance.• Monitored SQL Server performance and conducted tuning activities.• Designed ETL data flows using SSIS for data extraction from SQL Server.• Created mappings and workflows for Data Migration and Transformation from Access/Excel Sheets using SQL Server SSIS.• Worked on Dimensional Data Modeling for Data Mart design.• Identified Facts and Dimensions and developed fact tables, and dimension tables, using Slowly Changing Dimensions (SCD).• Implemented Error and Event Handling strategies in SSIS.• Utilized precedent constraints, Break Points, Check Points, and Logging.• Built Cubes and Dimensions with different Architectures and Data Sources for Business Intelligence.• Wrote MDX Scripting for Cube development.• Developed SSAS Cubes, including Aggregation, KPIs, Measures, and Partitioning Cube.• Worked on Data Mining Models and deployed and processed SSAS objects.• Developed ad hoc reports with complex formulas.• Executed database queries for Business Intelligence purposes.• Developed Parameterized, Chart, Graph, Linked, Dashboard, Scorecards, and Drill-down reports using SSRS.• Created cascading reports for enhanced data analysis.• Utilized SQL Server for database development and optimization.• Applied SSIS for ETL processes and SSAS for Cubes development.• Developed various types of reports using SSRS.
Frequently Asked Questions about Manoj Kumar
What company does Manoj Kumar work for?
Manoj Kumar works for Everest
What is Manoj Kumar's role at the current company?
Manoj Kumar's current role is Media Data Engineer.
Not the Manoj Kumar you were looking for?
-
Manoj Kumar
Senior Python Data Engineer At Change Healthcare | Python | Javascript | Angularjs | Html | Mysql | Aws | Css | Json | Apache Tomcat | Jquery |Github |Jira |Actively Looking For New Opportunities On C2C/C2HWest Hartford, Ct -
Manoj Kumar
Herndon, Va -
Manoj Kumar
Irvine, Ca2bandai.com, yahoo.com -
Manoj Kumar
Enterprise Data Solution Architect | Strategy & Architecture | Management Consulting | Technical Program & Product Management | Hands-On Data Engineer | Delivery Management | Data Management | Data GovernanceFolsom, Ca -
Free Chrome Extension
Find emails, phones & company data instantly
Aero Online
Your AI prospecting assistant
Select data to include:
0 records × $0.02 per record
Download 750 million emails and 100 million phone numbers
Access emails and phone numbers of over 750 million business users. Instantly download verified profiles using 20+ filters, including location, job title, company, function, and industry.
Start your free trial