Sensorfact’s mission is to eliminate all waste in the industry. Our machine-level sensors allow our customers to track their energy consumption at 30 second granularity, and our extensive analysis suite provides them with personalised savings advice. We continually expand our sensor suite, from electricity to water, gas and steam. We have extended our utilities suite with vibration sensors, allowing us to help customers reduce maintenance downtime. And the latest value proposition reaching our customers consists of production speed sensors to optimise the use of their machinery.
As a Data Engineer, you are the glue that makes it all possible, from raw sensor measurement to actionable insight for the customer. You will drive the development of:
Real-time ingestion of incoming sensor data from various sources, persisting it in our time series database
Batch processing pipelines and analytics microservices that provide personalised advice to our customers over our GraphQL API
Advanced real-time detection of machine faults and energy waste, powered by machine learning
Tooling for state of the art machine learning operations (MLOps), serverless and event-driven architecture and cloud services in AWS.
Being part of a scale-up, you are proactive in prioritising and solving the needs of our fast growing group of customers.
The key technologies you will be working with
Our AWS stack is focused on ingesting raw sensor data into Kafka, stream processing it using Flink and exposing it through Clickhouse. Batch processing is done using Prefect and Fargate, on-demand services are deployed using Lambda. We have a powerful internal GraphQL API to expose data to end users, managed by Hasura in combination with Typescript.
How we do it
We do Scrum with 2-week sprints, sprint planning and retrospective sessions. Our stand-ups are at 9:30 and if you're not there you can chime in over Meet. We know how important it is to get in the zone and write beautiful code so we schedule most meetings in the morning and keep the afternoon quiet (we try). We work from home about 70% of the time, but we enjoy meeting each other in the office regularly.
You will be in the Data team, which along with IoT and Platform make up the technology departments. The course is determined by quarterly goals, set collaboratively with the teams themselves. We don't believe in silos, so you will work in a multidisciplinary team with colleagues from multiple departments, represented by a product manager.
Your profileMSc (or PhD) in Computer Science, Distributed Systems, A.I., or a comparable field
Medior (2+ years) data engineer fluent in Python
Experience with modern cloud and data technologies such as Spark, Kubernetes, Docker, AWS Lambda, AWS Fargate, etc.
Experience with relational and/or OLAP database systems such as Postgres, Redshift, Bigquery, or time series databases such as Clickhouse or InfluxDB
Experience with software engineering best practices (version control, testing, code quality, CICD)
Are fluent in English
One or more of the following is a plus:
Experience with batch processing pipelines e.g. with Airflow or Prefect
Experience working with streaming platforms such as Kafka, as well as stream processing frameworks such as Spark Streaming, Flink, Storm, etc.
Experience working with Typescript, for example in backend programming
You can play a key role in creating a smart and sustainable industry;
We have an open culture where you get a lot of freedom and where taking ownership is valued;
We like to have you around, but you also get the opportunity to work remote (abroad) for up to 2 months per year;
You get to choose your own laptop or opt in for a reimbursement when you bring your own;
Legendary lunches, drinks and other activities with colleagues (our office manager Elise exceeds our expectations every time);
…And we obviously offer you the usual: a contract for 32-40 hours per week, employer contribution to your pension, 27 holidays, €500 contribution to set up/improve your home office and a NS business card to get to our great office. LI - MIFTAH