Основна информация
Реф. номер
Последен ден за кандидастване
Основно местоположение
Държава
Тип обява
Work Style
Описание и изисквания
Our team is expanding due to significant growth in new projects and data volume. We are seeking motivated and skilled Data Engineers to join our dynamic data warehousing team. This is not a backfill; you will be joining a growing function responsible for building and scaling the data infrastructure that powers our internal operations.
You will play a crucial role in developing and maintaining robust ETL pipelines using a modern tech stack centered on Google Cloud Platform (GCP) and Python. While our ideal candidate has GCP experience, we are committed to investing in the right talent and are open to cross-training individuals with strong foundational database and Python skills from other cloud (AWS/Azure) or on-premise environments.
This is a hands-on development role where you will take ownership of your work from design through to a post-deployment warranty period, ensuring the quality and reliability of our data flows.
Key Responsibilities:
Design, develop, and maintain scalable and efficient ETL pipelines to extract data from various sources, transform it, and load it into Google BigQuery.
Write high-quality, maintainable Python code for data processing and transformation, leveraging core libraries such as Pandas and NumPy.
Author and troubleshoot complex SQL queries to support data extraction, transformation, and analysis within our data warehouse.
Collaborate with team members and stakeholders in an Agile/Scrum environment to deliver features in sprints.
Provide post-deployment support for your developed solutions for a 3-month "warranty period," which includes monitoring jobs, identifying issues, and deploying fixes.
Engage with teams in different time zones (primarily Central America) for requirements clarification and issue resolution, requiring some flexibility in work hours.
Contribute to the overall architecture and design of our data solutions, ensuring they are robust and performant.
Допълнително описание на позицията
Qualifications & Skills:
Must-Have:
Experience: 3-5 years of professional experience in a data warehousing, ETL development, or data engineering role.
Database Expertise: Expert-level proficiency in SQL. You must be able to write, optimize, and troubleshoot complex queries, including multi-level joins, window functions, and aggregations.
Python Proficiency: Strong, hands-on experience using Python for ETL and data manipulation. Demonstrable expertise with data-centric libraries, specifically Pandas and NumPy, is required.
ETL Fundamentals: A deep understanding of ETL principles, data modeling, and data warehouse architecture.
Preferred & Advantageous:
Cloud Experience: Hands-on experience with Google Cloud Platform (GCP), particularly BigQuery.
Alternative Cloud Skills: Experience with other cloud platforms like AWS (Redshift) or Azure (Synapse Analytics) is a significant plus.
BI Tool Familiarity: Experience with any BI visualization tool (e.g., Tableau, Looker, Sisense, Power BI, Cognos) is considered a valuable secondary skill.
Agile Methodology: Prior experience working in an Agile/Scrum development team.
Work Environment & Expectations:
Work Hours: Standard working hours are 11:00 AM to 8:00 PM IST. Flexibility to occasionally extend hours for meetings with international teams is expected.
Role Type: This is a full-time, non-billable position as part of an internal shared-cost team.
Ownership: A strong sense of ownership is critical, especially concerning the 3-month post-deployment support period.
EEO Statement