Here’s How You Make an Impact:
- You’re a builder. You’ll be responsible for designing, building and deploying the components of a modern data stack, including CDC ingestion (using Debezium), a centralized Hudi data lake, and a variety of batch, incremental and stream-based pipelines.
- You’ll make things better. You enjoy the challenge of helping build and manage a fault tolerant data platform that scales economically, while balancing innovation with operational stability by maintaining legacy Python ELT scripts and accelerating the transition to dbt models in Redshift.
- You’re all about collaboration and relationships. You will collaborate within a cross-functional team in planning and rolling out data infrastructure and processing pipelines that serve workloads across analytics, machine learning and GenAI services. You enjoy working with different teams across Wave and helping them to succeed by ensuring that their data, analytics, and AI insights are reliably delivered.
- You’re self-motivated and can work autonomously. We count on you to thrive in ambiguous conditions by independently identifying opportunities to optimize pipelines and improve data workflows under tight deadlines.
- You will resolve and mitigate incidents: You will respond to PagerDuty alerts and proactively implement monitoring solutions to minimize future incidents, ensuring high availability and reliability of data systems.
- You're a strong communicator. As a data practitioner, you’ll have people coming to you for technical assistance, and your outstanding ability to listen and communicate with people will reassure them as you help answer their concern.
- You love helping customers. You will assess existing systems, optimize data accessibility, and provide innovative solutions to help internal teams surface actionable insights that enhance external customer satisfaction.
You Thrive Here By Possessing the Following:
- Data Engineering Expertise: Bring 3+ years of experience in building data pipelines and managing a secure, modern data stack. This includes CDC streaming ingestion using tools like Debezium into a Hudi data lake that supports AI/ML workloads and a curated Redshift data warehouse.
- AWS Cloud Proficiency: At least 3 years of experience working with AWS cloud infrastructure, including Kafka (MSK), Spark / AWS Glue, and infrastructure as code (IaC) using Terraform.
- Strong Coding Skills: Write and review high-quality, maintainable code that enhances the reliability and scalability of our data platform. We use Python, SQL, and dbt extensively, and you should be comfortable leveraging third-party frameworks to accelerate development.
- Data Lake Development: Prior experience building data lakes on S3 using Apache Hudi with Parquet, Avro, JSON, and CSV file formats.
- Workflow Automation: Build and manage multi-stage workflows using serverless Lambdas and AWS Step Functions to automate and orchestrate data processing pipelines.
- Data Governance Knowledge: Familiarity with data governance practices, including data quality, lineage, and privacy, as well as experience using cataloging tools to enhance discoverability and compliance.
- CI/CD Best Practices: Experience developing and deploying data pipeline solutions using CI/CD best practices to ensure reliability and scalability.
- Data Integration Tools: Working knowledge of tools such as Stitch and Segment CDP for integrating diverse data sources into a cohesive ecosystem.
- Analytical and ML Tools Expertise: Knowledge and practical experience with Athena, Redshift, or Sagemaker Feature Store to support analytical and machine learning workflows is a definite bonus!
0 applies
2 views
Other Jobs from Wave
Senior Software Engineer, Full Stack
Senior Data Analyst
Manager, Engineering
Principal Engineer
Senior Machine Learning Engineer
Similar Jobs
Senior Backend Engineer
Data Analyst 4
Data Engineer I
Data Engineer, eCS Data Engineering and Analytics
There are more than 50,000 engineering jobs:
Subscribe to membership and unlock all jobs
Engineering Jobs
60,000+ jobs from 4,500+ well-funded companies
Updated Daily
New jobs are added every day as companies post them
Refined Search
Use filters like skill, location, etc to narrow results
Become a member
🥳🥳🥳 401 happy customers and counting...
Overall, over 80% of customers chose to renew their subscriptions after the initial sign-up.
To try it out
For active job seekers
For those who are passive looking
Cancel anytime
Frequently Asked Questions
- We prioritize job seekers as our customers, unlike bigger job sites, by charging a small fee to provide them with curated access to the best companies and up-to-date jobs. This focus allows us to deliver a more personalized and effective job search experience.
- We've got about 70,000 jobs from 5,000 vetted companies. No fake or sleazy jobs here!
- We aggregate jobs from 5,000+ companies' career pages, so you can be sure that you're getting the most up-to-date and relevant jobs.
- We're the only job board *for* software engineers, *by* software engineers… in case you needed a reminder! We add thousands of new jobs daily and offer powerful search filters just for you. 🛠️
- Every single hour! We add 2,000-3,000 new jobs daily, so you'll always have fresh opportunities. 🚀
- Typically, job searches take 3-6 months. EchoJobs helps you spend more time applying and less time hunting. 🎯
- Check daily! We're always updating with new jobs. Set up job alerts for even quicker access. 📅
What Fellow Engineers Say