Contact

#
All open positions

Data Engineer

Dec 5, 2024

About SailPlan

SailPlan is a cutting-edge technology company that is dedicated to transforming the future of maritime transportation. Founded in 2019, the company is on a mission to make shipping more sustainable by reducing its carbon footprint and ensuring that it becomes NetZero. SailPlan offers a range of innovative solutions and services that enable its clients to optimize their operations and reduce their environmental impact.

SailPlan works with some of the most important names in the shipping industry to deliver a cleaner future for the world. SailPlan’s team comprises of experts with a diverse range of skills and experience, including naval architects, data scientists, and software engineers. The company’s collaborative and dynamic work culture fosters innovation and creativity, allowing the team to develop cutting-edge solutions that drive the industry forward. By combining state-of-the-art technology and a commitment to sustainability, SailPlan is leading the way towards a greener and more efficient maritime industry.

At SailPlan, you will be part of a fast-growing team, will wear many hats and have ownership over building key aspects of our platform. You will work within a collaborative environment to build the next generation of technology for the maritime industry.

If you think you have the right stuff, we are looking for YOU.

Location

This position may be located remotely or from our Headquarters in Miami / Fort Lauderdale, Florida, as determined on a case by case basis. Remote candidates must be US citizens located in the United States or Canada. Remote candidates are expected to travel to office periodically as necessary.

Role Description and Responsibilities

Role Overview:

We are looking for a skilled data engineer to join our growing team and play a critical role in developing the data infrastructure that powers our digital twin platform. This position will focus on leveraging the GCP stack, specifically dbt for data transformations and Looker for visualization, while building robust pipelines to handle data from real-time ingestion through BigQuery. The data engineer will collaborate closely with other engineers/analysts, and report to the data science tech lead.

Key Responsibilities:

  • Pipeline Development:
    • Build and maintain data pipelines to transform and model data using dbt, ensuring efficient and scalable integration with GCP tools such as BigQuery
    • Utilizing already existing BigQuery direct-ingestion to create efficient pipelines for data transformations 
    • Ensure pipelines are optimized for performance, cost efficiency, and scalability
  • Data Modeling:
    • Develop and maintain reusable, maintainable, and performant data models using dbt that support analytical and machine learning use cases
    • Collaborate with data analysts and cloud engineers to design and refine data models that meet visualization needs in Looker
  • Collaboration:
    • Work closely with cloud engineers to design and manage the underlying infrastructure required for the pipelines
    • Coordinate with the sensor/instrumentation engineer to ensure seamless ingestion of sensor data into GCP
    • Partner with the data science tech lead to align data engineering efforts with analytical and machine learning goals
  • Data Quality & Governance:
    • Implement monitoring and validation systems to ensure data accuracy, consistency, and completeness
    • Develop documentation for data models, pipelines, and workflows to ensure transparency and team-wide understanding
  • Visualization Enablement:
    • Enable the creation of effective dashboards and visualizations by building and managing the data foundation required for Looker and LookML

Core Requirements and Qualifications

  • Bachelor’s or Master’s degree in Computer Science, Software Engineering, or a related field
  • Strong experience with Google Cloud Platform (GCP) tools, including BigQuery, dbt, and Looker
  • Proficiency in SQL and Python for data manipulation, modeling, and pipeline development
  • Experience with ETL/ELT workflows, particularly using dbt for data transformation on time-series data
  • Hands-on experience with CI/CD practices and version control (e.g., Git)
  • Excellent problem-solving skills and a proactive, collaborative mindset

 

Preferred Qualifications

  • Familiarity with the maritime/shipping domain, including knowledge of sensor data and operational challenges
  • Experience with LookML and designing models for Looker dashboards
  • Knowledge of real-time data ingestion and streaming technologies (e.g., Pub/Sub)
  • Understanding of data security and compliance considerations, particularly for the maritime industry

Other information or Special Requirements

This is a full time position. US citizen or work permit required.