5+ years of work experience in relevant field (Data Engineer, DWH Engineer, Software Engineer, etc)
Experience with data-lake and data-warehousing technologies and relevant data modeling best practices (Presto, Athena, Glue, etc)
Proficiency in at least one of the main programming languages used: Python and Scala. Additional programming languages expertise is a big plus!
Experience building data pipelines/ETL in Airflow, and familiarity with software design principles.
Excellent SQL and data manipulation skills using common frameworks like Spark/PySpark, or similar.
Expertise in Apache Spark, or similar Big Data technologies, with a proven record of processing high volumes and velocity of datasets.
Experience with business requirements gathering for data sourcing.
Bonus - Kafka and other streaming technologies like Apache Flink.
Unless a specific application deadline is stated in the job posting, applications are accepted on an ongoing basis.
Please note, applicants are permitted to redact or remove information on their resume that identifies age, date of birth, or dates of attendance at or graduation from an educational institution.
We consider qualified applicants with criminal histories for employment on our team, assessing candidates in a manner consistent with the requirements of the San Francisco Fair Chance Ordinance.
Kraken is powered by people from around the world and we celebrate all Krakenites for their diverse talents, backgrounds, contributions and unique perspectives. We hire strictly based on merit, meaning we seek out the candidates with the right abilities, knowledge, and skills considered the most suitable for the job. We encourage you to apply for roles where you don't fully meet the listed requirements, especially if you're passionate or knowledgable about crypto!
Responsibilities
Building scalable and reliable data pipelines that collect, transform, load, and curate data from internal systems.
Augmenting the Data Platform with additional data pipelines sourced externally to ensure comprehensive coverage of necessary information streams for traders and institutions.
Ensuring high quality in built pipelines through rigorous auditing processes while maintaining performance standards that support near real-time accessibility, crucial for timely analytics and product development decisions.
Driving the design and deployment process to create a distributed data store serving as the central source of truth across Kraken's organization by integrating with company IT systems through built connections.
Developing customized self-service tools that empower users within Kraken to extract, analyze, and interact directly with large volumes of internal data for informed decision making without requiring direct technical support intervention.
Evaluating new technologies on a continuous basis by building prototypes aimed at improving the efficiency and capabilities of existing Data Engineering processes in line with evolving industry standards and practices.
Benefits
This is the target annual salary range for this role. This range is not inclusive of other additional compensation elements, such as our Bonus program, Equity program, Wellness allowance, and other benefits [US Only] (including medical, dental, vision and 401(k)).