Data Engineer
Knihobot & Bookbot is the largest re-commerce bookstore in the Czech Republic. Every day, we return pre-read books back into circulation, helping to protect the environment. So far, we've processed over 5 million books. Currently, we operate in four markets (DE, AT, CZ, SK), with ambitious growth plans, especially in Germany. Our success is directly tied to our expertise in data.
We have over a million books in our warehouse, and every day we receive another 20,000 while sending out about the same amount. This is why we need to leverage data fully. We aim to get books—with all the right information and prices—onto our e-shop as quickly as possible, ensuring they reach the right customers efficiently.
We work at a fast pace, with minimal bureaucracy. Sustainability and positive environmental impact are core to what we do. You'll see the results of your work daily and play a crucial role in driving the company's growth.
What to expect:
- You’ll be part of an internal data team of 5–8 people, working closely with IT, BizDev, Marketing, and Product teams. You’ll collaborate with key people like Markéta (Head of Data), Mikuláš (Product Owner, Books), and Patrik (AI Engineer).
- Delivering data products for projects across the entire Knihobot world.
- Developing in the cloud, building data warehouses and pipelines capable of handling tens of millions of records daily.
Example projects you’ll work on:
- Building a pipeline that searches the data warehouse for newly received titles and enriches their records in the production database.
- Integrating data from a third-party system used for planning shifts for hundreds of temporary workers and transforming it into a reportable format.
- Analyzing data on the most searched authors and titles across our e-shops.
Who we’re looking for:
- At least 3 years of experience in a similar role.
- Creating ETL/ELT pipelines and orchestrating them has been a regular part of your work.
- Strong analytical skills, understanding of data modeling, and database systems, and the ability to analyze data using SQL/Python.
- Experience with basic cloud infrastructure (Azure, GCP, AWS). You know what storage, virtual machines, proxies, IAM, and serverless computing services are.
- Familiarity with data development best practices: version control (GitHub), testing, automated QA.
- AWS experience is a plus.
- Experience working with big data is a plus.
- You are independent, proactive, and take ownership of your tasks, but also enjoy working in a team.
- Fluent in English for professional discussions.
Tech stack:
- Database systems: MySQL, Snowflake, BigQuery.
- Visualization tools: PowerBI, Metabase.
- Data warehousing and orchestration: AWS, Keboola, Dataform.
- Analytics: SQL, Python, Google Sheets.
What we offer:
- A role in a fast-growing, dynamic company that has been sustainably expanding for several years. Check out a post from our CEO Dominik for more details.
- One extra week of vacation after your first year, two extra weeks after two years.
- An informal work environment with a pragmatic approach.
- 20% discount on books, VIP selling tariff, and parking at our Kolbenova office.
- MultiSport card.
- Budget for personal development.
- Option to work from home two days a week.