Jon Powell

Jon Powell Email and Phone Number

Principal Data Engineer at CVS Health @ CVS Health
Jon Powell's Location
San Diego, California, United States, United States
Jon Powell's Contact Details

Jon Powell personal email

n/a
About Jon Powell

Your new position must pay about $250K/yr or $135/hr w2.Trusted w/critical clients & large initiatives--a unique combination of technical acumen, leadership & customer-facing skills.Technical: Big Data/Data Science/DevOps Principal Engineer competent w/the end-to-end architecture of data systems (pipelines, warehouses, analytics, visualization, ML), architect-level software engineering, and complex problem-solving/algorithmic skills. Competent in performing complete product life cycle: pre-sales, conception, requirements gathering/feedback, architecture, design, implementation, testing, validation, monitoring & alerting, and documentation. Leadership: commands technical and personal respect among team members by providing clear vision, guidance, and problem-solving using competence, diligence, clear communication, lucid documentation, and servant leadership.Soft Skills: superb ability to learn/research, mitigate conflict, document, speak clearly, and listen to stakeholder & audience needs.Key competencies:Data System Architecture: system development on major clouds, especially GCP, of message-based, event-driven, responsive architectures, both batch & streaming pipelines and stores for analytics, ML, visualizations, data sharing, or API’s for consumption. Also familiar with DevOps and ML Ops best practices such as monitoring, alerting, and data quality concerns.Data Pipelines: Cloud Native: GCP (PubSub, Cloud Functions, GKE/K8s, DataFlow, BigQuery(ML), Cloud Storage, StackDriver, Monitoring…), AWS (S3, Redshift, Glue, IAM, QuickSight), streaming or batch modes (Lambda, Kappa Architectures), Data ETL/ELT, validation/reconciliation / EDA, Visualization (Looker, Google Data Studio, AWS Quick Sight & Python)Data Science: growing experience with VertexAI/AutoAI/BigQuery ML, Jupyter Notebooks, Feature Development, EDA, Missing Value Imputation/Encoding, Pandas, ScikitLearn, Numpy, MatplotLib/SeabornData Warehouse: BigQuery, minimal Redshift, Snowflake/Snowpipe (knowledge), and Data Vault 2.0 methodology, HDFS.API Development: familiar with latest (and legacy) techniques (REST), formats (JSON) and protocols (HTTPS, gRPC), etc.Data Tools: Data Catalog, dbt, Great Expectations, DLP, Analytics Hub, BigQuery Omni… Database/Storage: BigQuery (+Cloud Storage, Firebase, Spanner…), BigQuery Omni, S3, Cassandra, MongoDb, Oracle, MySQL, SqlServer, Postgres, HSQLDBLanguages: Python, Go (familiarity), Scala, Java, C/C++, UX* shells, awk, SQLProgramming: Functional, MicroServices, Design Patterns, framework development, multi-threading, real-time, UX, web

Jon Powell's Current Company Details
CVS Health

Cvs Health

View
Principal Data Engineer at CVS Health
Jon Powell Work Experience Details
  • Cvs Health
    Principal Data Engineer
    Cvs Health May 2024 - Present
    Woonsocket, Ri, Us
    Anything data related -- TBD.
  • Egen
    Principal Data Engineer
    Egen Sep 2021 - May 2024
    Naperville, Il, Us
    Senior staff member reporting to CTO for GCP Architecture, Data & Software Engineering. Responsibilities: core technical resource responsible for shaping sales value propositions, project scoping, technical discovery, and technical Data Architecture. Introduced solution accelerators to reduce engagement times and improve technical precision.Examples:* Large healthcare client (225K employees): multiple projects, including GCP BigLake Lakehouse. Generated test data using ChatGPT LLM. Restored customer, Google confidence. Supply Chain Analytics using SQL & Looker viz w/visibility to CEO's effort to save $500M.* Large insurance client (~$10M deal) - led rapid requirements gathering, analysis, seminal architecture. Project continues today with ongoing, profitable work.*Massive, famous government entity: determined requirements, designed large-scale (~300M events/day w/1M events/min) webhooks GCP implementation. Delivered on schedule, with DevOps, monitoring/visualizations, test strategies to ensure consistency, reduce project risk. Used ChatGPT/Generative AI.*Large government contractor: led climate change ecosystem POC, including overall vision & roadmap for curating & analyzing large datasets, a sharing permission model, AI/ML enablement, and visualizations.*State Government Entity #1: MDM Identity project -- developed an innovative, efficient, flexible solution with auditable identity clusters for massive datasets.Large Hardware Retailer: Developed GCP system, invented DAG orchestration engine to coordinate REST microservice execution for minimal latency by depending only on declarative YAML-like data dependencies.*State Government Entity #2: quickly wowed stakeholders with data pipeline and Looker viz work. Project detractors became SpringML evangelists after seeing consistent, powerful visualizations using their data. Developed viz standards that eliminated combinatorial explosion of charts, improved look & feel, and increased data exploration capabilities.
  • Supportlogic
    Senior Data Architect
    Supportlogic Aug 2020 - Sep 2021
    San Jose, California, Us
    Hands-on Cloud Data Architect specializing in GCP, Data Pipelines and API's using Python.SupportLogic is a solid, early-stage startup that utilizes Machine Learning to revolutionize client ticket handling and service.My current role involves scaling the data architecture 10-100x, introducing REST API's to improve architectural separation of concerns, and reworking the current data warehouse for realtime analytics to be used to serve the UI and REST clients.In my previous role at the Burwood Group, we listened carefully to clients' needs, taking them through the entire development lifecycle, striving to exceed their expectations. I specialized in "Data Transformations" for external & internal customers that also involved meeting w/clients.- Educated, developed & presented marketing/sales slide decks for Business Development & Sales to sell Cloud and Data Transformation Engagements.- Complete Cloud System design from customer assessment, requirements gathering/ordering, component architecture/design/selection, implementation, monitoring/alerting, testing/validation & documentation & training handoff. Notable project: Twitter Decahose (~400M records/day, 0.9Tb data).- Data warehouse setup (analysis, import & maintenance) and fusion of data silos for multiple clients using GCP offerings for both realtime & batch processes using both ETL/ELT paradigms.- Business Analytics KPI & visualization development. Clients love these!- Machine Learning POC's for customers to show value.
  • Burwood Group
    Senior Data Engineer
    Burwood Group Jun 2019 - Jul 2020
    Oak Brook, Illinois, Us
    Company SME on Data Pipelines, Data Warehouses, Data Software Engineering, Data Science, and Data Visualization for GCP and AWS. Coming up to speed on Azure.Help drive, advise & implement data pipelines, data warehouses, and visualizations using the optimum native Google Cloud, AWS & Azure Components for many clients in higher education, medical, advertising, IoT, and retail domains. Includes Data Warehouse lift & shift, Information Architecture & DWH Data Fusion, batch/real-time ETL/ELT pipelines, and Business & Predictive Analytics (AI/ML) using visualizations. Twitter DecaHose, GSuite integration automated ELT architectures. A few of the designs required significant engineering around costs, latencies, data volumes, paging issues, and disparate ingestion APIs and are generally data configuration-driven and completely automated using NoOps services.Notable client responsibilities:Twitter Decahose: primary Data architect on daily high-volume/velocity (385M records and 0.9Tb per day) project. Balancing costs, volume and latency were a few of the challenges.Data Analytics using Cloud Functions, BigQuery and professional visualizations using Data Studio. Had additional IoT project as well as producing valuable marketing info.
  • Rackspace
    Data Engineer On Rpc Data Science Team At Rackspace
    Rackspace Mar 2018 - Feb 2019
    - Strategize with Data Science lead on new projects (Smart Ticket Auto-Routing, Server Failure Prediction, AutoAI tool) and how to position/sell to stakeholders.- Smart Ticket Auto-Routing project (save $ by routing to optimal ticket resolver): data gathering/cleaning/validation and feature generation to boost predictability. Data was loaded from legacy MS-SQL DB's and uploaded to BigQuery where feature development was performed. Moderate (accuracy) results were achieved and presented to stakeholders.- Server Failure Prediction (avoid support tickets by fixing servers proactively): data gathering/cleaning/validation for server failure detection. Hadoop data was uploaded to GCS and then uploaded to BigQuery, where it was processed.- AutoAI (automatically fit models based on input data): did overall System Design and implemented automatic data EDA Explorer, multi-processing model candidate runner as well as helping on the Flask UI.- Continual training/mentoring to become a Data Scientist by taking courses in/outside of work and doing Kaggle competitions & kernels.- Attended ODSC West 2018 Data Science Conference and reported back to the team.- Produced many BigQuery SQL-based queries and visualizations (Google Data Studio) that have proved valuable to the Operations Team, creating excitement all the way up to the C-level. Operations people were able to debug problems in minutes vs. multiple days.- Advocated/evangelized the importance of Information Architecture (IA) up to Senior Management, showing the importance of designing for the data to provide business visibility into operations, etc.- Kafka, Python & Data Studio POC: rapidly produced business analytics results that proved our team could read a realtime data feed and produce professional-looking visualizations.
  • Datapipe
    Data Engineer On Data Science Team
    Datapipe Aug 2017 - Mar 2018
    Jersey City, New Jersey, Us
    - Designed, implemented & deployed simple REST API service with Swagger & Python that provides SQL access for Ad Hoc analytics queries of AWS S3 JSON data using AWS Athena & Glue. Results are returned in cursor-like chunks. Deployed application as a Docker container using Helm charts to configure a Kubernetes cluster. - Designed tentative Kappa Architecture (realtime monitoring integrated with batch processing) for our infrastructure monitoring pipeline that utilized flexible alarming limits based on appropriate time periods. It would utilize Kafka, S3 and ostensibly Glue.- Worked on concepts and project synthesizing the ideas of Event-Sourcing, CQRS and an asynchronous ESB. Co-authored a paper published by O`Reilly that summarized our ideas.
  • Dexcom (Continuous Glucose Monitoring)
    Principal Software Engineer (Server Team)
    Dexcom (Continuous Glucose Monitoring) May 2016 - Jun 2017
    - One of two architects that designed & built a Data Pipeline that transforms JSON medical raw sensor/diagnostic (IoT) data into Google BigQuery relational tables for use by Data Scientists, Business Analysts and others. Our GCP technology stack consisted of Google PubSub, DataFlow, BigQuery, Cloud Storage and StackDriver (all PaaS solutions). Also considered were Bigtable, Cloud SQL & Spanner for persistence and Spark for data processing. Data pipeline features included duplicate & lost-record detection, data quality/reconciliation and monitoring systems to make it HIPAA compliant.- Became local expert on Looker visualization tool and produced many visualizations for monitoring, data reconciliation & analysis. Developed extensive backing BigQuery views (materialized & non) to support these visualizations, keeping efficiency, responsiveness & data freshness in mind.- Primary interface to Google Cloud team, gathering evidence & communicating issues related to their components by clearly documenting bugs we discovered and pushing to resolution.- Continued to develop data science/analysis skills by learning/writing Python, using iPython Notebook (Jupyter), applying descriptive statistics and wading thru numerous swamps of incorrect, duplicated, or just faulty data.- Attended Google Next 2017 and took Introductory 1 Day Machine Learning Practicum with Tensor Flow, and discussed our GCP issues with Google engineers. Attended Cassandra Summit 2016 and earned Developer Certificate. Attended Spark Summit 2016, expanding knowledge of Spark’s strengths, weaknesses & application areas.- One of two local experts on IaaS solutions such as Apache Spark, Cassandra, Kafka, K-Streams. Developed training material on these topics plus Scala, and presented most of them to a wide group. These presentations were very well received.
  • Mir3 (Incident Alerting)
    Software Architect (Core R&D Team)
    Mir3 (Incident Alerting) Oct 2014 - May 2016
    - Built an HTTP REST Query MicroService that accepts queries on the front end and efficiently submits them to Spark, returning the results as JSON. Identified MicroServices technology, designed the REST API, discovered easy way to submit Spark query jobs and built efficient queries (3 successive joins) using Spark Cassandra Connector on an (unfortunately relational) input schema. As a bonus, the service also reported health and status to accommodate production readiness.- Performed Apache Spark configuration/setup and recommendations to NetOps teams, ETL/sample data file ingestion and integration of Spark-Cassandra database connector. Configured Hive2 Thrift Server and Spark tables (RDD’s) for high throughput (efficient data pipeline) and JDBC access via visualization tools.- Lead team that analyzed 20 UI vendors/solutions for suitability as our AdHoc reporting & dashboard visualization solution, scoped staged tasks for evaluating vendors and hosted implementation phone/webex sessions to prove vendor’s solutions meet Mir3’s needs.- Team lead on new ground-breaking Event Sentry (Weather) project involving NoSQL DB, multi-threading, CAP Protocol and Geo location. Scoped stories for team, architected and documented design and implemented critical server-side (complex) algorithms.- New ScriptCache threading design for messaging core that utilizes weak references and is mostly lock-free, which should result in significantly improved throughput in the core event server.
  • Corelogic
    Software Architect (Data / Analytics Group)
    Corelogic Apr 2011 - Sep 2014
    Irvine, Ca, Us
    - Architected/Designed/Implemented multiple projects simultaneously on my own or in small teams, including analytics models (Rental Amount, Disposition, LendOptix, FraudmarkA/B, WatchSuite), model-backing file generation & aggregation system, Logging Service & associated database schemas. One of “goto guys” for algorithms, threading tuning, & hard problems.- Scala evangelist and one of local language “experts.” Led a language “run-off” complete with benchmarking that was used to select Scala as our language of choice. Then, two of us learned, developed and propagated the language throughout the organization.- Proposed Modeling Framework “Meta-Architecture”: investigated, documented and implemented new streamlined approach to model development using Scala. Projects that formerly had hundreds of source files and took 3-6 months to develop are now done with roughly 10 source files in 2-4 weeks! Some of these time/file reductions were due to the establishing of enabling conventions and parallel analytics development. These improvements accelerated CoreLogic development cycles, beating schedules & reducing time to market.- Greatly streamlined model/file validation process from multiple weeks down to days using innovative search/compute/join and automation/match techniques on big datasets, thereby saving time, conserving budgets & eliminating numerous headaches.- Significantly improved overall analytics model throughput by introducing reliable multi-threading. Designed & coded multi-threaded, Web-enabled test harness to validate.- One of organization-wide leaders in documentation via MediaWiki. Served on numerous committees (Collaboration, Architecture, Documentation Standards).- WatchSuite product (team lead): understood & modeled current “patched together” analytics suite for mortgage fraud detection. Proposed top level architecture & schema; interacted with multiple divisions & data sources. Project now generates over $1M annual revenue.
  • Qualcomm Qes Division
    Sr. Software Architect (Consultant)
    Qualcomm Qes Division Apr 2010 - Apr 2011
    - Data retention service: large set of stored procedures that implemented periodic database cleanup.- Trip Management Project: converted project to standardized POM’s / DM Server deployment and eased XSD parsing by introducing XPath solution. Took end-to-end responsibility for Trip Import feature (architecture/design/impl), including vendor interaction, fixed/CSV/XML adapter design & WS implementation.- Project lead on GIS accounting project. Converted unreliable, partially manual to fully automated, database-centric solution that is 100% automated and delivers easy-to-read Excel spreadsheet to business owners monthly.- Became a local expert on Spring DM Server 2.0 (OSGI-compatible) for media delivery mobile product.- Extensive Maven build / deploy responsibility for dev/integration/staging environments including Hudson, ended up suggesting standardized way to setup POM’s for maximum flexibility / maintainability.
  • Id Analytics (Fraud Detection)
    Senior Software Engineer (Contract)
    Id Analytics (Fraud Detection) Jan 2007 - Apr 2007
    - Designed, implemented & tested (JUnit) Java code for an extremely complex XML-based workflow engine that implements the company’s new Health Care Claim Fraud Detection solution. This project is the company’s largest and most complex to date. Code is very OO, configurable, parallelizable and distributed, utilizing J2EE components such as JMS. Working environment was very fast paced, so productivity and accuracy were very important.- Helped articulate design in specification document and found some critical bugs that improved overall quality.
  • Cymer
    Contract Java Software Engineer
    Cymer 2004 - 2005

Jon Powell Skills

Agile Methodologies Architecture Embedded Systems C++ Cloud Computing C Python Sql Java Databases Representational State Transfer Big Data Mysql Software Development Data Analysis Data Visualization Google Cloud Aws Aws S3 Gcp Bigquery Gcp Data Studio Scala Data Cleaning Pandas Scikitlearn Numpy Multi Threading Git Machine Learning Microsoft Office

Jon Powell Education Details

  • California State University, Chico
    California State University, Chico
    Comp Sci
  • Sonoma State University
    Sonoma State University
    Minor In Physics
  • University Of California, Davis
    University Of California, Davis
    Engineering

Frequently Asked Questions about Jon Powell

What company does Jon Powell work for?

Jon Powell works for Cvs Health

What is Jon Powell's role at the current company?

Jon Powell's current role is Principal Data Engineer at CVS Health.

What is Jon Powell's email address?

Jon Powell's email address is jp****@****ood.com

What schools did Jon Powell attend?

Jon Powell attended California State University, Chico, Sonoma State University, University Of California, Davis.

What skills is Jon Powell known for?

Jon Powell has skills like Agile Methodologies, Architecture, Embedded Systems, C++, Cloud Computing, C, Python, Sql, Java, Databases, Representational State Transfer, Big Data.

Free Chrome Extension

Find emails, phones & company data instantly

Find verified emails from LinkedIn profiles
Get direct phone numbers & mobile contacts
Access company data & employee information
Works directly on LinkedIn - no copy/paste needed
Get Chrome Extension - Free

Aero Online

Your AI prospecting assistant

Download 750 million emails and 100 million phone numbers

Access emails and phone numbers of over 750 million business users. Instantly download verified profiles using 20+ filters, including location, job title, company, function, and industry.