Vortexa is a fast-growing international technology business founded to solve the immense information gap that exists in the energy industry. By using massive amounts of new satellite data and pioneering work in artificial intelligence, Vortexa creates an unprecedented view on the global seaborne energy flows in real-time, bringing transparency and efficiency to the energy markets and society as a whole.
The Data Production Team is responsible for all of Vortexa's data, ranging from mixing raw satellite data from 600,000 vessels to generating high-value forecasts such as vessel destinations, cargo onboard, ship-to-ship transfer detection, dark vessels, congestion, and future prices.
The team has built a variety of procedural, statistical, and machine learning models that enable us to provide the most accurate and comprehensive view of energy flows. We take pride in applying cutting-edge research to real-world problems in a robust and maintainable way. The quality of our data is continuously benchmarked and assessed by experienced in-house market and data analysts to ensure the quality of our predictions.
You'll be instrumental in designing and building infrastructure and applications to propel the design, deployment, and benchmarking of existing and new pipelines and ML models. Working with software and data engineers, data scientists, and market analysts, you'll help bridge the gap between scientific experiments and commercial products by ensuring 100% uptime and bulletproof fault-tolerance of every component of the team's data pipelines.
You Are:
Experienced in building and deploying distributed scalable backend data processing pipelines that can handle terabytes of data daily using AWS, K8s, and Airflow.
Fluent in both Java and Python (with Rust being a plus).
Knowledgeable about data lake systems like Athena, and big data storage formats like Parquet, HDF5, and ORC, with a focus on data ingestion.
Driven by working in an intellectually engaging environment with top minds in the industry, where constructive and friendly challenges are encouraged.
Excited about working in a start-up environment: not afraid of challenges, eager to bring new ideas to production, and possess a positive can-do attitude.
Passionate about coaching developers, helping them improve their skills and grow their careers.
Experienced in the full software development life cycle (SDLC), including technical design, coding standards, code review, source control, build, test, deploy, and operations.
Awesome If You:
Have experience with Apache Kafka and streaming frameworks, e.g., Flink.
Are familiar with observability principles such as logging, monitoring, and tracing.
Have experience with web scraping technologies and information extraction.
Are motivated by being collaborative, working, and achieving together.
Value a flexible working policy accommodating remote and home working, with regular staff events.
Enjoy private health insurance offered via Vitality to help you look after your physical health.
Appreciate a global volunteering policy to help you ‘do good' and feel better.
#J-18808-Ljbffr