Experienced in building data lakes and data warehouses, always ensuring data integrity, availability, performance while enabling advanced analytics and reporting. I have a strong grasp of API integrations to seamlessly connect data sources and systems.Skilled in designing and implementing ETL pipelines, optimizing data architecture, and leveraging cloud technologies to unlock business decisions insights.Additionally, I have a solid understanding of infrastructure-as-code principles to automate the provisioning and management of cloud resources, ensuring reliable and reproducible environments.Technologies:- Python (Pandas, PySpark, Numpy)- SQL (PostgreSQL, MySQL, AuroraDB, Firebird, BigQuery, Redshift)- NoSQL (DynamoDB, DocumentDB)- Clouds (AWS, GCP)- S3, Redshift, Glue, CloudWatch, EventBridge, RDS, DMS, Kinesis, Athena, EC2- Orchestration (Glue Workflow, Airflow)- Data visualization (Power BI, Datastudio, Qlik, Tableau)- CI/CD, TerraformEnthusiastic about turning raw data into meaningful insights that drive business success, I continuously explore new technologies and concepts. Collaboration plays a crucial role in my work as I stay in constant contact with colleagues to understand their needs and contribute to the company's development, always mindful of cost-effectiveness.GitHub: https://github.com/gs-costa
Critical Techworks
View- Website:
- criticaltechworks.com
- Employees:
- 2778
-
Data EngineerCritical TechworksLisbon, Portugal -
Data EngineerKanastra Aug 2023 - PresentUberlândia, Minas Gerais, Brazil- Implemented row-level quality tests, which provided more granularity to pinpoint and remediate data quality issues. Using the library Great Expectations (GX) and Spark, there's been about a 5% increase on average in task duration, which it's great compared to the benefits.- Reduced 70% costs with MERGE operations in BigQuery native tables, migrating BigQuery native tables to external tables hosted in GCS.- Refactored an API service in FastAPI (Python) to improve readability and maintainability. The service processes document uploads and extracts data using AI, leveraging object-oriented programming techniques to enhance code structure and efficiency.- Developed an API in FastAPI to provide clients with direct access to data stored in the data warehouse. Empowered clients with autonomy and ensured high availability for seamless data retrieval.Technologies:- Python (Pyspark, Pandas, FastAPI)- API (Rest, Soap)- SQL, NoSQL- Airflow- GCP Cloud Storage, BigQuery, Dataproc, Kubernetes, Artifacts, Cloud Composer, Secret Manager- Mac OS- GitHub- CI/CD- Docker -
Data EngineerLinker Apr 2022 - Aug 2023São Paulo, BrasilAchievements:- Reduced in 80% the dashboards failures in production by fixing ETL pipeline jobs;- Increased in 2x the efficiency of data delivery to partners;- Reduced costs in 70% by parallelizing data pipelines.Context:At Linker, a bank exclusively focused on serving companies, my role involves building consistent ETL pipelines that source data from APIs and both SQL and NoSQL databases. I work with batch processing and near real-time updates to ensure the data is up-to-date and accurate. I also collaborate directly with business teams, helping them extract meaningful data and creating automation solutions using refined data to build business rules. Technologies:- Python (PySpark, Pandas, Numpy)- API (Get, Post)- SQL, NoSQL- PostgreSQL, MySQL, AuroraDB- DynamoDB, DocumentDB- AWS S3, Redshift, Glue, DMS, Kinesis, RDS, Lambda, Athena- Linux- CI/CD -
Data EngineerBowe | B2B Growth Feb 2022 - Apr 2022Uberlândia, Minas Gerais, BrasilAchievements:- Reduced 80% of manually extraction of data from marketing tools (social medias, CRM, mail automations, leads journeys);- Reducted execution time in 85% of ETL Python code extraction from API;Context:At Bowe, a marketing and sales consulting company, my role involved engaging with clients to gain insights into their data sources and identify any requirements for customized dashboards. To ensure seamless data integration, consolidated customer data from diverse sources into a centralized data lake and warehouse, employing ETL pipelines. Additionally, I collaborated closely with data analysts, lending my expertise to assist in the development of interactive and visually appealing dashboards, empowering our clients to closely monitor and evaluate the performance of their marketing campaigns.Technologies:- Python (Pandas, Numpy, Requests)- SQL- AWS Lambda, Redshift, S3, Athena, EventBridge, CloudWatch- Power BI, DataStudio, Qlik- AppScript- API -
Senior Business AnalystBowe | B2B Growth Aug 2021 - Feb 2022Achievements:- Increased 30% (7 to 5 hours) in agility of dashboard implementation by standardizing data.- Increased team revenue by 20% in creation of a product for marketing and sales analysis.Context:During that time, main focus was on maintaining client dashboards and creating a product that prioritized delivering analytics to track marketing campaigns that generated sales.Technologies:- Python- Google Functions, BigQuery- AWS Lambda, S3, Redshift, Athena, EventBridge, CloudWatch- Power BI, DataStudio, QlikView- API -
Business AnalystBowe Martech Mar 2021 - Aug 2021Uberlândia, Minas Gerais, BrasilAchievements:- Reducted storage costs in 40% and increased agility in 80% the execution of functions through the migration of the Google Cloud Platform (GCP) cloud server to Amazon Web Services (AWS);- Reduced costs in 20% by consuming API with Python, instead of payed extractors.Context:During that period, the stability of customer dashboards with daily updates had been achieved. Consequently, the focus changed to pipeline efficiency and implementing a data lake solution.Technologies:- Python- Google Functions, BigQuery- AWS Lambda, S3, Redshift, Athena, EventBridge, CloudWatch- Power BI, DataStudio, QlikView -
Junior Business AnalystBowe Aug 2020 - Mar 2021Uberlândia, Minas Gerais, BrasilAchievements:- Increased 25% of revenue with BI works by assuming new clients;- Increased 60% number of active dashboards by building new dashes and keeping up to date twice a day;- Reduced 80% of manually extraction of data from marketing tools (social medias, CRM, mail automations, leads journeys);Context:As the first member of Bowe's Business Intelligence (BI) team had the responsibility of creating intuitive dashboards for our clients and automate the update process.Technologies:- Python- Google Functions- BigQuery- Power BI, DataStudio- Stitch, Pluga
Gustavo Costa Education Details
-
Bowe AcademyMarketing -
Aeronautical Engineering -
Mechanical And Aerotechnical Engineering -
Automação E Programação Em Labview
Frequently Asked Questions about Gustavo Costa
What company does Gustavo Costa work for?
Gustavo Costa works for Critical Techworks
What is Gustavo Costa's role at the current company?
Gustavo Costa's current role is Data Engineer.
What schools did Gustavo Costa attend?
Gustavo Costa attended Fiap, Bowe Academy, Ufu - Universidade Federal De Uberlândia, Ecole Nationale Supérieure De Mécanique Et D'aérotechnique, Universidade Federal De Uberlândia.
Who are Gustavo Costa's colleagues?
Gustavo Costa's colleagues are Fábio Oliveira, Miguel Pedrosa, Lara Faustino, Luis Ferreira, Daniel Rodrigues, Joana Fernandes, Bernardo Câmara.
Not the Gustavo Costa you were looking for?
-
Gustavo Costa
São Paulo, Sp -
Gustavo Costa
Tech Lead | Product Owner | Data Engineering Specialist At GloboRio De Janeiro, Brazil -
Gustavo Costa
São Paulo, Sp -
-
Gustavo Costa
Rio De Janeiro, Rj
Free Chrome Extension
Find emails, phones & company data instantly
Aero Online
Your AI prospecting assistant
Select data to include:
0 records × $0.02 per record
Download 750 million emails and 100 million phone numbers
Access emails and phone numbers of over 750 million business users. Instantly download verified profiles using 20+ filters, including location, job title, company, function, and industry.
Start your free trial