Job Description
What’s the opportunity?
We are searching for an accountable, multitalented Data Engineer to facilitate the operations of
our data solutions. This Data Engineer position will be responsible for sustaining and improving
our Business Intelligence platform, performance, scalability and processes. During various
aspects of this process, you should collaborate with coworkers to ensure that your approach
meets the needs of each request.
What will I be doing?
Sustaining and improving our BI PlatformBoost Amazon Redshift performanceDesign and build the Medallion Architecture in order to optimize data pipelines for different tiers (Bronze, Silver and Gold) to support efficient data processing and analysisImplement and maintain data pipelines, considering internal and external schemas, and ensuring data quality and consistencyDetecting and correcting errors in our BI Platform and process in order to guarantee reliabilityProactively addressing future BI Platform needsDevelop and maintain data documentation of the BI Platform Architecture, processes, and proceduresEnsure timeliness, resilience, availability and accessibility of data present in our BI PlatformEvaluate and test new tools with a focus on efficiency, cost reduction and qualityRemaining up-to-date with industry standards and technological advancements that will improve the quality of our deliverablesContribute to the definition of standards and best practices for the development of data solutionsWork closely with the Data teams to provide better solutionsSupport the team with new tools adoptionSupport ad-hoc requests related with data extractionWork with leadership in defining the company's BI data strategy and in actions of Data GovernanceWhat skills do I need?
Bachelor's degree in Data Engineering, Big Data Analytics, Computer Engineering, or related fieldProven experience as a Data Engineer, Data Platform Engineering or similarProficiency in SQL and Python and one or more programming languages (Java, C++, etc.)Strong knowledge of Amazon Redshift optimization/tuning techniquesFamiliarity with Hadoop or suitable equivalentStrong analytical and problem-solving skillsExperience in Big Data tools and platforms, such as Hadoop, Spark, Docker, Kubernetes, EKS, Terraform and DevOps CI/CDBackground in big data solutions within cloud infrastructures such as AWS, Azure Cloud, Google CloudExperience with data warehousing tools such as RedshiftExperience with AWS (Lambda, EKS, EC2, Spark (EMR), S3, DynamoDB, Glue, Athena, AWS Cost Monitoring)Solid knowledge of process orchestration tools such as AirflowKnowledge of transformation tools DBT (Data Build Tool), Pentaho Data IntegrationAbility to work independentlyTeam playerExcellent communication (written and spoken English) and collaboration skills