Ohita ja siirry sisältöön

Yleiset tiedot

Sijainnit: Vancouver, British Columbia, Canada 
  • Sijainti: Vancouver
  • Osavaltio:
  • Maa: Canada


Roolin tunnus
206327
Työntekijän tyyppi
Regular Employee
Studio/osasto
CTO - Studio Data Insights
Joustava työjärjestely
Hybrid

Kuvaus ja vaatimukset

Olemme maailmanlaajuinen tiimi, joka koostuu sisällöntuottajista, tarinankertojista, teknologian asiantuntijoista, alkuperäisten kokemusten tekijöistä, innovaattoreista ja todella monista muista. Uskomme, että upeat pelit ja kokemukset alkavat tiimeistä, jotka ovat yhtä monimuotoisia kuin pelaajat ja yhteisöt, joita palvelemme. Electronic Artsilla rajana on vain mielikuvituksesi.

As part of the Adrenaline team, join our team and drive the development of our data products that align with our business objectives. You will collaborate with cross-functional teams, including leadership, architects, analysts, product management, project management, and data scientists, to understand and address data requirements effectively. You will report to the Director of Data Architecture. 

What you will do:

  • Collaborate with cross-functional stakeholders, including analyst, governance, project management, and data scientists, to comprehend data needs and ensure alignment with business goals.

  • Design, build, and launch efficient and reliable data pipelines to move and transform data effectively, handling both large and small amounts.

  • Possessing the capability to interpret and construct code, as well as provide insightful feedback during code reviews, focusing on elements such as scalability, performance, and adherence to best practices.

  • Ensure scalability and efficiency by designing and implementing complex features, including multi-layered data workflows.

  • Deploy data quality checks in collaboration with quality engineering and data governance teams to uphold high-quality data standards.

  • Uphold a high-quality code base by writing and reviewing performant, scalable, and well-tested code to deploy effectively using source control products such as GitLab/GitHub.

  • Define and manage SLAs for all data sets in allocated areas of ownership.

  • Utilize job scheduling tools like Airflow and GitLab Runners to streamline processes.

  • Maintain data engineering and architecture best practices and standards within the team and the wider organization, fostering a culture of quality, innovation, and experimentation.

  • Take ownership of the end-to-end data engineering component of the product.

  • Identify, design, and implement internal process improvements in collaboration with data architects, including automating manual processes and optimizing data delivery.

  • Create and maintain technical documentation.

Qualifications:

  • BS/MS in Computer Science, Engineering, or a related field with 8+ years of data engineering experience.

  • A must experience with Snowflake, Git,  DBT & Airflow

  • 7+ years of experience in ETL orchestration and workflow management tools.

  • Experience using containerization technologies such as Docker or Kubernetes.

  • Proficiency in programming languages such as Python (preferred)  & SQL

  • Strong understanding of data modeling, warehousing, and building ETL/ELT/EtLT pipelines.

  • Knowledge or Experience in Data mesh and Data Lake House Designs.

  • Excellent written and verbal communication skills, ability to simplify and synthesize complex constructs.