We are looking for a Senior Java Developer to join our growing team and contribute to building a new product focused on Container Risk Analysis.
As an Experienced Backend Engineer, you will play a crucial role in the design, development, and maintenance of our data platform. This platform supports critical data-driven decision-making throughout the organization.
You will ensure smooth data flow and optimize our services by working closely with cross-functional teams, including data scientists, analysts, and other software engineers. Requirements: * Bachelor’s degree in Computer Science, Engineering, or a related field; * A minimum of 5 years of professional experience as a Backend Engineer; * Full proficiency in Java (specifically Spring Boot) and Kafka, with a strong background in developing micro-services architecture and streaming solutions; * Proficiency with Generative AI (GenAI) and Large Language Model (LLM) tools; * Proactively utilized and consistently engaged with LLM and AI tools (e.g., Claude) throughout the development lifecycle; * Expertise in AWS services for data storage, processing, and analytics; * Demonstrated experience in designing and implementing robust ETL (Extract, Transform, Load) processes for data transformation and integration; * Good working knowledge of Python; * Experience with MySQL; * Exceptional problem-solving skills and meticulous attention to detail; * Ability to work independently and take ownership of tasks; * Good command of English, both written and verbal.
Nice to Have: * Familiarity with data processing technologies like Spark; * Experience with SingleStore.
Responsibilities: * Utilize Java, including Spring Boot, to build robust and high-performance data processing services within our data platform; * Implement real-time data streaming solutions using Kafka, ensuring timely data ingestion and availability; * Collaborate closely with cross-functional teams to comprehend data requirements, identify opportunities for data optimization, and support data-driven initiatives; * Lead the design, development, and maintenance of efficient and scalable data pipelines, facilitating data collection, processing, and transformation from diverse sources; * Leverage AWS services for data storage, processing, and analytics, adhering to security and performance best practices; * Monitor and troubleshoot service performance, proactively identifying bottlenecks and implementing optimizations; * Uphold data integrity, reliability, and availability by implementing effective ETL processes and conducting data quality checks.
What we offer: * We value your skills and ensure you’re rewarded accordingly; * Enjoy 15 days of annual leave to recharge and maintain a healthy work-life balance; * Fully remote work with a flexible setup; * Our supportive and approachable HR team is always here to help you thrive.