Fornova is a fast-growing technology company specializing in advanced data collection and analytics for the travel industry. We operate unique scraping pipelines and proprietary data processing technologies to deliver high-quality, actionable insights.
TimeZone: Europe
Location: Fully Remote. We believe that location should not be a barrier to finding the best talent.
Work hours: 9:00 AM — 5:00 PM in the candidate’s local timezone, with flexible scheduling options (Sun—Thu/Mon—Fri).
Job description: We are looking for a Strong Junior Scraping Expert to join one of our technical operations teams. This role focuses on mastering Fornova’s pipelines and proprietary scraping technology to resolve production issues, implement new developments, optimize infrastructure, and improve cost-efficiency. The ideal candidate is technically proactive, self-driven, and experienced in building, scaling, and hardening web scraping solutions.
Must-have skills: — 1+ years of experience in web scraping, data collection, or similar development roles. — Strong proficiency in Python and building/maintaining scraping scripts. — Experience with headless browsers and browser automation frameworks (e.g., Playwright, Puppeteer, Selenium). — Understanding of user fingerprints, anti-bot detection mechanisms, and strategies to randomize or evade them. — Familiarity with web protocols, HTTP methods, API reverse engineering, and debugging tools (browser dev tools, Postman). — Practical experience with Linux command line and common developer tools (pip, Git/Bitbucket). — Ability to independently troubleshoot complex scraping and extraction issues. — Good communication skills in English, both written and spoken. — BA in Computer Science, Information Systems, Industrial Engineering, or related field preferred.
It would be very useful if you have experience working with the following: Scanners & Crawlers, Splash, Lua script, Fiddler, Postman, MongoDB, Cassandra, ElasticSearch, TypeScript, basic Docker knowledge, Linux, JIRA, Robot framework, Kibana, Tableau
Responsibilities: — Performing website analysis to overcome bot detection mechanisms, and implementing these solutions as a highly efficient and maintainable automated data collector in Python. — Adding enhancements to existing infrastructure. — Understanding internal data pipelines, troubleshooting, spotting and overcoming issues on the different stages of data flows. — Integrating the external partner’s solutions into our infrastructure. — Taking a logical, data-backed approach to prioritizing projects, while cooperating with colleagues from different departments at the same time. — Taking ownership of challenging tasks and enjoying being the “go-to” person. — Enhancing our knowledge base with new high-end solutions.
If you find this position interesting and suitable for you, please share with our recruiters answers for the questions below. You do not have to answer them right now, but if you do, you will definitely move faster in our recruiting process!
1. What was the most challenging website you managed to collect data from? Please elaborate on how you overcame it. 2. What was the strongest blockage/bot detection/challenge you faced during the data collection? How did you manage to overcome it? 3. What was the biggest number of records you managed to collect in 24 hours?
Feel free to share any other information that you think is relevant to your candidacy.