Wei Chu Email and Phone Number
Wei Chu work email
- Valid
Wei Chu personal email
Olewave.com provides extensive customizable multimodal datasets in various languages and topics, amounting to millions of hours. Our pricing is 1/10th of traditional vendors, with 5x data efficacy. We also offer a robust and customizable data collection and cleaning pipeline than can run on your site or cloud. Our data service and pipeline enable you to train top-notch GPT-4o-style multimodal or Generative AI models in-house. Specializing in data since 2015, we ensure quality and reliability.See the introduction of Olewave's large-scale conversational speech dataset with speaker labels and high quality transcriptions. The data samples are also presented:https://www.linkedin.com/pulse/olewave-large-scaled-convesational-speech-dataset-olewave-927ic/Links to Olewave's Youtube channel -- mostly paper reading on Speech and NLPhttps://www.youtube.com/channel/UCm99ZwZ1bODHskkHwMOue0wI have a mirror Youtube channel for my Chinese audience (most of them cannot access Youtube):https://space.bilibili.com/12477508The organizer of Speech and NLP Meetup Grouphttps://www.meetup.com/speech-and-language-technology-meetup-group* Any institute who is interested in sponsoring/hosting an offline meetup with Speech and NLP people in Bay Area, feel free to contact me.
Olewave
View-
Co-Founder And CeoOlewave Aug 2023 - Present** Customized Large-Scale Datasets ** Olewave delivers customized, labeled, and validated large-scale real-world NLP/CV/speech/multimodal datasets of various scenarios such as dictation and conversation in multi accents/dialects/languages, and of diverse topics such as education, finance, legal, entertainment, healthcare, retail, and customer service. Our datasets include: • topic-specific text datasets for training your own LLM model • visual/video/image datasets with tags/prompts for training your own CV model • speech/audio datasets of different languages and dialects for training your own ASR/Whisper/SeamlessM4T/TTS model. • and multimodal datasets.We constantly collect timely data from languages including Latin America Spanish, Arabic, Southeast Asian, Chinese, Japanese, Korean.1/3-1/10 in data pricing vs. traditional data vendors with the same or better data efficacy.** Proprietary/Private Data Cleaning and Labeling **Olewave offers bespoke solutions for proprietary data labeling, normalization, and transformation.Tired of inaccurate transcriptions and frustrating APIs? We offer a superior solution with:• AI-powered Accuracy: Transcribe any audio, regardless of language, dialect, accent, or topic, with exceptional accuracy. We surpass the competition in understanding even the most challenging recordings.• Detailed Insights: Gain valuable insights with word/character-level confidence scores, precise timestamps, and advanced speech analytics.• Privacy Guaranteed: Keep your data secure. Integrate our powerful data labeling tool directly into your platform, eliminating risks associated with external APIs.• Competitive Pricing: Enjoy high-quality service at accessible prices, outperforming both tech giants and human-intensive transcription solutions.Ready to experience the difference? Don't settle for mediocrity. Contact info@olewave.com and give us a try! -
Staff Research ScientistPing An Oct 2019 - Aug 2023Shenzhen, Guangdong, CnTo PhD students who are interested in doing research on but not limited to Seq2Seq-based large vocabulary continuous speech recognition (RNN-T) and spoken language understanding (End-to-End-based Conversational AI) with me, feel free to send your CV/resume to wei.w.chu@gmail.com.I have supervised 10+ interns, most of them ended their internship with publications, launched project, and systematic ways of doing academic research. -
Staff Research ScientistLiulishuo Jan 2018 - Oct 2019Shanghai, Shanghai, Cn1) Full-stack development and improvement of acoustic, pronunciation, language modeling, and decoder optimization for streaming non-native large vocabulary continuous speech recognitiona. Acoustic Modeling (Research & Development): i. Multi-Task Learning on a limited amount of training data collected from each task. ii. Using Adversarial training to train for dictation task using read speech iii. Research on speaker/accent-embeddings and neural network topologiesb. Pronunciation Modeling/Dictionary (Research & Development): i. Automatically detect mispronunciation then generate a pronunciation dictionary for English learners.c. Language Modeling and Decoding (Development) i. Push the accuracy (10-15%) and latency (0.3-0.4s) of a streaming decoder to their limit. ii. Adding confidence score and domain-special re-scoring for PARTIAL hypothesis.2) End-2-End Speech Recognition: a. Compared the wordpiece vs character on hybrid CTC-Attention based network. Paper published.3) Video Synthesis and Smoothing a. Fast Audio/Text-driven Talking Head Synthesis on Limited Amount of Training b. Automatically smooth head movement when stitching video clipstag: kaldi, openfst, language model, acoustic model, pronunciation, deep learning, machine learning, speech synthesis, voice conversion, hotword detection, spoken keyword spotting, keyword recognition, DSP, speech activity detection, audio classification, automatic closed captioning, music composition, speech recognition, natural language processing, spoken dialogue system, computer vision, computer graphics, lip sync, talking head animation, noise robust speech recognition, deep neural network, recurrent neural network, convolutional neural network, AI, LSTM, End-to-End speech recognition, DNN, RNN, CNN, CTC, CUDA, GPU, cluster, hadoop, mapreduce, big data.. -
Research ScientistSnapchat, Inc. Feb 2016 - Dec 2017Santa Monica, California, UsAs the first Speech Research Scientist and Engineer at Snap Research, promptly prototype, build and customize speech recognition services, keyword spotting in audio services, and audio/speech processing libraries for product teams.1. Robust Closed Captioning for Audio Content-- Worked with News/Ads/User Content teams to customize the above speech recognizer to generate closed captions for sound tracks with music and noise.2. Spoken Keywords Spotting-- Co-developed and deeply customized a BLSTM-based keyword spotting engine on the cloud for curse word detection and brand name detection in noisy oral speech.3. Hotwords Detection on Mobile-- Worked with colleagues to deliver a DNN-based light-weight hotwords detector on iOS. Evaluated noise robust features. Co-invented a way of automatically generating data with transcriptions for training.3. Audio Signal Processing on iOS/Android\\-- Worked out purely DSP-based speech detection and enhancement libraries.-- Evaluated and improved audio fingerprint algorithm (Shazam-like). -- Designed and built a PSOLA-based spoken voice to singing voice conversion process (Patent filed). -- Worked with our intern on a RNN-based music composition machine that can automatically generate pleasant MIDI music with multiple instruments. Paper accepted. 4. Acoustic Event Detection-- Worked out a pipeline to extracting data for deep learning purpose. Co-tried different network structures. Customize it as a service for engineering teams.5. Deep Learning-based Voice Conversion-- Co-worked out a many-to-one BLSTM-based voice conversion algorithm. Paper submitted.6. Audio-driven Lip Sync-- Co-worked out a DNN-based audio driven lip sync algorithm which is well working and runs fast. Paper published.7. Multi-modal Emoji Prediction-- Co-worked out a deep learning-based Emoji predictor given audio and text. (Patent filed).8. Consulting for Internal Teams-- Audio Transformation (Voice Filter), Beamforming, Audio quality testing. -
Staff Software EngineerSony Computer Entertainment America Mar 2014 - Feb 2016San Mateo, California, Us-- Built and delivered production-level Deep Neural Network (DNN)-based models for speech recognition.-- Built a low-cost but yet high-performance CPU+GPU computing cluster for deep learning experiments.-- Designed and implemented new practical algorithms for improving the performance of speech recognition.-- Handled practical aspects of a production-level speech recognizer, which include: * Speech and text normalization and lexicon modeling for training multi-language recognizers, * Improving a recognizer's robustness to speaker and environmental variations by performing adaptation, * Customization and optimization for speech recognition on different tasks and platforms, * Towards reliablely rejecting background noises and unrelated speech input to a recognizer, * Designing workflow and preparing data for validation and QA purpose. -
Application EngineerIntel Corporation Apr 2013 - Mar 2014Santa Clara, California, UsLearning the skills of handling issues from vendors, customers, and internal develop and debug teams. Designed and built a test system for internal teams, customers and vendors. Learned how a corporation operates. -
Sr. Speech ScientistLumenvox Dec 2012 - Apr 2013San Diego, California, UsUnderstood the key aspects in a successful Interactive Voice Response (IVR) system. Also learned how a middle-sized company operates. -
Speech ScientistVoci Technologies Jan 2012 - Nov 2012Earned 500k funding for the company. Also learned how a startup operates.
-
Research InternMicrosoft Jun 2011 - Sep 2011Redmond, Washington, UsAudio-visual-depth automatic speech recognition on Kinect. -
Research AssistantUcla 2007 - 2011Los Angeles, Ca, UsWorked out very accurate noise robust pitch estimation and tracking algorithm.Pioneer study in bird song classification, recognition, and detection. -
Summer InternDisney Research 2010 - 2010Los Angeles, Ca, UsMicrphone array-based kid speech recognition.. -
Summer InternRosetta Stone Jun 2009 - Aug 2009Arlington, Virginia, UsStatistically-based Letter-To-Sound conversion. -
Summer InternMitsubishi Electric Research Labs Jun 2008 - Sep 2008Cambridge, Ma, UsDiscriminative training for speech recognition. -
Visiting StudentMicrosoft Apr 2007 - Aug 2007Redmond, Washington, UsAcoustic event detection. -
Research InternIntel Jul 2006 - Oct 2006Santa Clara, California, UsSpeaker segmentation and clustering.
Wei Chu Skills
Wei Chu Education Details
-
UclaElectrical Engineering -
Tsinghua UniversityElectrical Engineering -
North China University Of TechnologyElectronic Engineering -
UclaSignal And Systems
Frequently Asked Questions about Wei Chu
What company does Wei Chu work for?
Wei Chu works for Olewave
What is Wei Chu's role at the current company?
Wei Chu's current role is GPT-4o/multimodal dataset collection pipeline & service by Olewave.com; large-scale, from different languages and topics, 1/10 in data pricing with 5x data efficacy vs. other vendors. 20 yoe in speech/voice/audio R&D..
What is Wei Chu's email address?
Wei Chu's email address is we****@****cla.edu
What schools did Wei Chu attend?
Wei Chu attended Ucla, Tsinghua University, North China University Of Technology, Ucla.
What are some of Wei Chu's interests?
Wei Chu has interest in Cooking, Skiing, Rock Climbing, Biking, Running.
What skills is Wei Chu known for?
Wei Chu has skills like Machine Learning, Signal Processing, Pattern Recognition, Speech Recognition, Speech Processing, Image Processing, Python, Statistical Modeling, Algorithms, Perl, Artificial Intelligence.
Free Chrome Extension
Find emails, phones & company data instantly
Aero Online
Your AI prospecting assistant
Select data to include:
0 records × $0.02 per record
Download 750 million emails and 100 million phone numbers
Access emails and phone numbers of over 750 million business users. Instantly download verified profiles using 20+ filters, including location, job title, company, function, and industry.
Start your free trial