Senior Data Engineer
Upload My Resume
Drop here or click to browse · PDF, DOCX, DOC, RTF, TXT
Requirements
• You are fluent in modern Python and SQL. You are used to work with modern software development tooling and best practices (Docker, CI/CD, unit testing, evals). We do not run data pipelines built on Jupyter Notebooks in production. Bonus point if you are have already worked with other programming languages (e.g. Rust, Scala, Elixir). • You know how to read and understand an EXPLAIN plan. • Proven expertise in building robust data pipelines and evolving data architectures. Previous exposure to modern data storage formats (e.g. Arrow, Iceberg), query engines (e.g. DataFusion) and database design patterns (CDC, queues, message brokers) is highly beneficial. We have a bias to keep things simple - but sometimes we’ll have to push the bar up to cope with challenging problems. • Especially at the beginning, an end-to-end data ownership mindset. It will not happen every day but - if necessary - you might be asked to set up a no-code ETL in our customer data platform, to debug third-party tools behaviour, or to support teammates with an important one-off empirical analysis. • Fluency in English and Italian. • Bonus point: at least 7-8 years of experience. You have already coached and mentored junior colleagues.
Responsibilities
• You will report directly to Dario, Chief Data Officer & Co-Founder. You will own, iterate and build all the pipelines and systems that empower the people within our team to take good, sound, data-driven decisions. We strongly believe that the foundations behind strong business intelligence is a well-crafted data platform. • This will not be a one-man standing. It will take a lot of cross-functional work with the people in our analytics, system engineering, product and business teams. • We sit on top of the most business-critical company data. Money flows, bank transactions, invoices, cash register records, payables and receivables, general ledger records. We are the accounting & finance operating system of our customers. On top, core internal data is originated across multiple systems (CRM, billing, messaging interfaces, etc). You will build systems that connect this information and store it so that it can be retrieved (i) easily (ii) quickly (iii) safely by our teammates & AI agents. • Build and iterate fast, efficient and robust data pipelines (batch, stream ETL, etc) • Manage our analytics data warehouse - together with our system engineers - to make it scalable and reliable. Queries (when well-written…) should execute fast • Build and scale data catalogs, ontologies and - more generally - a semantic layer for our data infrastructure • Ensure data quality, anonymisation of sensitive information and, in general, technical compliance to privacy laws and adherence to technical standards • Examples of projects you could work on from month one at Sibill: • Centralise our ETL tasks in a single, well-architected data platform • Migrate our current analytics data warehouse off Postgres onto an OLAP-minded database
Benefits
• RAL: range between 55-75k €, based on your experience • Meal voucher: (8€/business day) • Full remote (based in Italy). We will just ask you to come to Milan for a few weeks during the onboarding (trip is on us) • You will work within a team of smart, hard-working, passionate and nice human beings. People in Sibill tend to be very happy. Happiness is an underrated benefit ;-) • Recruiting process • a quick chat with Gaia, our HR Manager • 30 minutes “know-each-other” session with Dario (Chief Data Officer) • 45 minutes technical interview • you will meet some of the key people you will work with (engineering, data, product) • 45 minutes with Mattia (CEO) and Lorenzo (Chief Product Officer) • By applying, you consent to Sibill’s processing of your personal data for recruitment and hiring purposes, in accordance with our Privacy Policy available at: https://sibill.com/trattamento-dei-dati/