Job description
Are you smart, driven, curious, resourceful, and not afraid to fail? Then we want to meet you! Our team of bold, innovative, and creative teammates is what makes us a top startup to work for. FreightWaves delivers news and commentary as well as data and analytics which empower risk management and actionable market insights in the logistics and supply chain industry. If you are ready to join our team, it is time for YOU to apply!
FreightWaves is on the hunt for a curious, tenacious, and team-oriented Senior Data Engineer to join our fast paced engineering team. The ideal candidate is inquisitive, versatile, team oriented, thrives on change, and has a positive attitude. If you are ready to be challenged, learn new and exciting technologies, and have the unique opportunity to work with some of the most talented developers in the country, we want you to apply!
**This position is fully remote.**
**Must RESIDE in the United States and be eligible to work.**
What you will be doing:
- Implementing ingestion pipelines, using Airflow as the orchestration platform, for consuming data from a wide variety of sources (API, SFTP, Cloud Storage Bucket, etc.).
- Implementing transformation pipelines using software engineering best practices and tools (DBT)
- Working closely with Software Engineering and DevOps to maintain reproducible infrastructure and data that serves both API-only customers and in-house SaaS products
- Defining and implementing data ingestion/transformation quality control processes using established frameworks (Pytest, DBT)
- Building pipelines that use multiple technologies and cloud environments (for example, an Airflow pipeline pulling a file from an S3 bucket and loading the data into BigQuery)
- Create and ensure data automation stability with associated monitoring tools.
- Review existing and proposed infrastructure for architectural enhancements that follow both software engineering and data analytics best practices.
- Working closely with Data Science and facilitating advanced data analysis (like Machine Learning)
What you bring to the table:
- Strong working knowledge of Apache Airflow
- Experience supporting a SaaS or DaaS product, bonus points if you were creating new data products/features
- Strong in Linux environments and experience in scripting languages
- Python Expert
- Strong understanding of software best practices and associated tools.
- Experience in any major RDBMS (MySQL, Postgres, SQL Server, etc.).
- Strong SQL Skills, bonus points for having used both T-SQL and Standard SQL
- Experience with NoSQL (Elasticsearch, MongoDB, etc.)
- Multi-cloud and/or hybrid-cloud experience
- Strong interpersonal skills
- Comfortable working directly with data providers, including non-technical individuals
- Experience with the following (or transitioning from equivalent platform services):
- Cloud Storage
- Cloud Pubsub
- BigQuery
- Apache Airflow
- dbt
- DataFlow
Bonus knowledge/experience:
- Experience implementing cloud architecture changes
- Working knowledge of how to build and maintain APIs using Python/FastAPI
- Transforming similar data from disparate sources to create canonical data structures
- Surfacing data to BI platforms such as Looker Studio
- Data Migration experience, especially from one cloud platform to another
- Certification: Professional Google Cloud Certified Data Engineer
Our Benefits:
- An excellent work environment, flat hierarchies, and short decision paths
- A generous benefits package including 100% employer-paid health, dental, vision and Life insurance, STD, LTD
- Stock options
- Appealing 401k matching plan
- Career Mentorship Opportunities
- Personal Development Credit (Can be used toward Student loans or relevant PD Courses)
- Annual life achievement bonus for having a baby, buying a house, or getting married (max one per year)
- No set days off Vacation policy (our team takes time off as needed with supervisor approval)
- Up to $50 for Gym or Virtual Gym membership.
- Audible or Kindle Unlimited subscription
- Discount on Ford vehicles
UkDeIkUkPj