David T. Han

David T. Han Email and Phone Number

Software Engineer @ Zipline AI
New York, NY, US
David T. Han's Location
New York, New York, United States, United States
David T. Han's Contact Details

David T. Han personal email

n/a

David T. Han phone numbers

About David T. Han

Experienced as a data engineer, data analyst, full stack and infrastructure engineer.

David T. Han's Current Company Details
Zipline AI

Zipline Ai

View
Software Engineer
New York, NY, US
Website:
zipline.ai
Employees:
7
David T. Han Work Experience Details
  • Zipline Ai
    Software Engineer
    Zipline Ai
    New York, Ny, Us
  • Stripe
    Software Engineer
    Stripe Aug 2021 - Present
    South San Francisco, California, Us
    - Machine learning infrastructure engineer on Stripe's feature engineering platform Shepherd. https://stripe.com/blog/shepherd-how-stripe-adapted-chronon-to-scale-ml-feature-developmentSpark + Flink + Airflow + Scala + Java + Hive + IcebergBuilt out integration connecting where ML models were deployed to and how they would fetch and process streaming/batch features from Shepherd.Led migration of Stripe's early merchant fraud detection offline ML model from legacy feature engineering platform to Shepherd. Implemented the Shepherd based features and conducted extensive offline evaluation to ensure new features matched old ones. Backtested features to confirm score distributions and recall of new features were inline with the pre-existing features. Led technical implementation of first asynchronous Shepherd based ML model for merchant fraud. Built out online flow consisting of an event consumer subscribed to various Kafka topics that fetched online features from feature store and conditionally trigger model scoring downstream. Worked with team of ML engineers to assemble an automated backtesting and training data pipeline using offline computed point-in-time training data involving Airflow + Flyte + Iceberg. - Led technical design and development of first cut of Stripe's new core product infrastructure managing multi-entity accounts/businesses. See https://docs.stripe.com/get-started/account/orgs.Responsible for ideation + development of the read optimized tree graph storage strategy to represent a Stripe enterprise business with multiple accounts and/or entities. Built read + write APIs and workflows including various locking schemes to support concurrent updates to the tree while maintaining correctness. Java + RPC + Protobuf + Bazel + MySQL + Mongo- Developed real time stream processing jobs using Flink + Scala + Bazel + Kafka to monitor the merchant experience of Stripe's customer base across the world.
  • Career Break
    Personal Goal Pursuit
    Career Break Jul 2021 - Aug 2021
    Pursuing personal projects.
  • Agilemd
    Analytics And Data Engineer
    Agilemd Mar 2021 - Jul 2021
    San Francisco, California, Us
    Led and designed architecture plans to upgrade existing data pipeline infrastructure away from MongoDB including:- pros and cons proposal of replacing MongoDB with AWS Redshift or Snowflake- proof of concept detailing use of Airbyte to load from multiple sources of AgileMD data to AWS Redshift and DBT to own transformations within Redshift.- cost benefit analysis between building and hosting open source data frameworks listed above versus buying a managed ETL/ELT service
  • Amazon
    Software Engineer
    Amazon Jul 2019 - Mar 2021
    Seattle, Wa, Us
    Building asynchronous micro-services with Java using AWS CloudFormation, Lambda, DynamoDB, SQS, and EventBridge to solve problems in cross border movement for Amazon Logistics.Architected and developed end to end solution to migrate production data in AWS DynamoDB to internal Amazon data lake for further processing by dependent business intelligence and data engineer teams using AWS CloudwatchEvents, S3, and SNS.Led performance readiness assessment for team’s services (> 10) in preparation for increased traffic due to 2020 Q4 holiday season. Analyzed previous traffic patterns and initiated load testing of individual services to understand if services’ limits would handle the forecasted traffic. Horizontally or vertically scaled team’s services on a case by case basis (larger AWS EC2 instances, more EC2 instances, configuring AWS DynamoDB to autoscaling, etc).Completed integration of Amazon Devices Logistics onto team’s platform and served as lead and primary point of contact for new technical issues and features.
  • Petal
    Software Engineer Ii
    Petal Jun 2019 - Jul 2019
    New York, Ny, Us
  • Petal
    Software Engineer
    Petal Jun 2018 - Jun 2019
    New York, Ny, Us
    Transforming traditional credit underwriting to better serve credit invisibles.- Led development in internal web application to expose sensitive customer information (PII) to stakeholders using jquery, Python + Flask, Postgres + Redshift, and secret management with Hashicorp Vault. Deployed application using Docker + Hashicorp Nomad, and AWS EC2. Web application was tightly integrated with the day to day operations for risk, finance, and customer experience teams.- Architected solution to automate and centralize all data to AWS Redshift to build business dashboards showcasing insights from application/platform data combined with third party reporting. Used change data capture in AWS Database Migration Service to stream data from RDS Postgres to AWS S3, and batch processed to AWS Redshift. Managed infrastructure using Terraform.- Coordinated data engineering efforts with stakeholders such as risk, finance, and customer support analysts to model data in Redshift and better increase operational efficiency. Created dashboards for tech and non-tech teams to monitor business and engineering metrics with SQL and business intelligence tool (Periscope).- Drove initiatives to increase observability in production backend systems. Created monitoring and alerting systems around backend Flask app through Prometheus + Grafana, PagerDuty, and Slack in order to provide coverage around the health status and related APIs of the registration and dashboard website. - Re-architected backend mono-repository testing by segregating each component of repo to have its own seeded Postgres Docker database in order to parallelize unit testing. Decreased average testing time from 30 to 10 minutes.- Developed ETL pipelines in Python to batch process files in variety of formats stored in AWS S3 and ingested to AWS Redshift. Used open source Concourse CI and later Jenkins to schedule cron jobs. Reduced latency of ETL jobs across company by three times through multiprocessing via Python.
  • Capital One
    Software Engineering Intern
    Capital One May 2017 - Aug 2017
    Mclean, Va, Us
    - Migrated existing audit flow containing logs for customers of the commercial bank website Intellix from IBM’s DB2 mainframe to Postgres database using AWS CloudFormation.- Interacted with database administrators to implement new security groups, various login privileges, unique indexes, and schema for Postgres database using pgAdmin4.- Developed REST API for Postgres database using Java Spring, Spring Boot, and Jersey to allow faster access of customer logs in order to better assist customer support with troubleshooting and developers with software bugs for Capital One’s commercial bank website, Intellix. Implemented caching feature and configured time-to-live to reduce latency for database calls. Developed feature that exported and encrypted large collections of data to text file, and uploaded to Amazon S3 bucket.
  • Sql Sentry
    Ios Engineer Intern
    Sql Sentry May 2016 - Aug 2016
    Charlotte, North Carolina, Us
    - Developed UI that displayed all user-flagged data from a SQLite backend by creating a table view with collapsible sections and dynamic changing of table cell heights. - Handled user interaction for de-flagging of data and resulting removal of table cells.
  • Duke University
    Electrical Engineering Intern
    Duke University Jun 2015 - May 2016
    Durham, North Carolina, Us
    Research under the Nanomaterials and Thin Films Lab at Duke University. Programming an algorithm in Mathematica that obtains structural parameters of carbon nanotube using various statistical distributions.
  • Elon University
    Research Assistant
    Elon University Sep 2015 - Dec 2015
    Elon, Nc, Us
    Research under Dr. Scott Spurlock. Applying features extraction and dimensional reduction (machine learning) on video frames to analyze human movement using Matlab.
  • Elon University
    Research Assistant
    Elon University Sep 2013 - May 2015
    Elon, Nc, Us
    Conducted research under Dr. Benjamin Evans. Investigated a model to explore the feasibility of a variation on traditional field geometries for magnetic nanoparticle hyperthermia cancer treatment. Used video microscopy to characterize magnetic forces on a novel magnetic microsphere.https://cismm.web.unc.edu/wp-content/uploads/sites/9983/2016/02/Evans_Han_High-Permeability_Elsevier.pdf

David T. Han Education Details

  • Columbia University
    Columbia University
    Computer Engineering
  • Elon University
    Elon University
    Engineering Physics

Frequently Asked Questions about David T. Han

What company does David T. Han work for?

David T. Han works for Zipline Ai

What is David T. Han's role at the current company?

David T. Han's current role is Software Engineer.

What is David T. Han's email address?

David T. Han's email address is dh****@****lon.edu

What is David T. Han's direct phone number?

David T. Han's direct phone number is (336)-584*****

What schools did David T. Han attend?

David T. Han attended Columbia University, Elon University.

Free Chrome Extension

Find emails, phones & company data instantly

Find verified emails from LinkedIn profiles
Get direct phone numbers & mobile contacts
Access company data & employee information
Works directly on LinkedIn - no copy/paste needed
Get Chrome Extension - Free

Aero Online

Your AI prospecting assistant

Download 750 million emails and 100 million phone numbers

Access emails and phone numbers of over 750 million business users. Instantly download verified profiles using 20+ filters, including location, job title, company, function, and industry.