🚀 Join Our Data Products and Machine Learning Development Remote Startup! 🚀
Muttdata is a dynamic startup committed to crafting innovative systems using cutting-edge Big Data and Machine Learning technologies.
We’re looking for a DataOps Engineer to help take our expertise —and our internal optimization platform— to the next level. If you consider yourself a data nerd like us, we’d love to connect!
As a key contributor, you'll play a pivotal role in designing, deploying, and maintaining ML models that power our product for marketing campaign budget optimization and allocation. You’ll work closely with the product team and business stakeholders, bringing robust systems into production, introducing fresh technical insights, and ensuring our models deliver real business impact.
If you thrive on building scalable ML pipelines, solving complex challenges, and shaping products that drive decision-making at scale, we’re eager to connect with you! 🐶🚀
🚀 What We Do
- Leveraging our expertise, we build modern Machine Learning systems for demand planning and budget forecasting.
- Developing scalable data infrastructures, we enhance high-level decision-making, tailored to each client.
- Offering comprehensive Data Engineering and custom AI solutions, we optimize cloud-based systems.
- Using Generative AI, we help e-commerce platforms and retailers create higher-quality ads, faster.
- Building deep learning models, we enhance visual recognition and automation for various industries, improving product categorization, quality control, and information retrieval.
- Developing recommendation models, we personalize user experiences in e-commerce, streaming, and digital platforms, driving engagement and conversions.
🌟 Our Partnerships
- Amazon Web Services
- Astronomer
- Databricks
🌟 Our Values
- 📊 We are Data Nerds
- 🤗 We are Open Team Players
- 🚀 We Take Ownership
- 🌟 We Have a Positive Mindset
🔍 Curious about what we’re up to? Check out
our case studies and dive into our
blog post to learn more about our culture and the exciting projects we’re working on! 🚀
Responsibilities 🤓
- Orchestration & Integration: Build and manage scalable data pipelines using Apache Airflow, dbt, and Airbyte, ensuring seamless data ingestion and movement.
- Product Development: Work hand in hand with Mixilo’s product team to address real client and internal users' issues, design technical solutions, and help prioritize the roadmap.
- Infrastructure as Code (IaC): Use Terraform to provision and manage cloud resources on AWS, maintaining a secure and cost-effective infrastructure.
- Kubernetes & GitOps: Manage containerized applications and services on Kubernetes (EKS), implementing continuous delivery practices to keep Mixilo running smoothly.
- Enhance DX (Developer Experience): Abstract away complex DAG and dbt logic to reduce manual work for the team and optimize our time to market.
- Data Reliability: Implement rigorous testing frameworks—specifically leveraging dbt tests—to ensure data quality and catch errors before they impact client recommendations.
- CI/CD for Data: Maintain and improve our CI/CD pipelines to automate testing, deployment, and infrastructure changes.
Required Skills 💻
- Senior Data Engineering Foundations (Minimum 5+ Years): Strong experience in Python and a deep mastery of SQL and Postgres.
- Modern Data Stack: Hands-on experience with Airflow, dbt, and Airbyte.
- Analytical Data Systems: Experience in constructing analytical data systems over Data Lakes (e.g.,AWS S3, Athena, EMR, Glue, Iceberg/Delta, etc.).
- Cloud Mastery: Solid understanding of AWS services (S3, EC2, RDS, IAM, etc.)
- .Infrastructure & Containers: Proficiency with Terraform, Docker, and Kubernetes.
- Operations Mindset: A strong understanding of GitOps, CI/CD principles, and a passion for automation.
- Teamwork: Great capacity for collaborative, async, and written communication. You take ownership and follow through on your commitments.
Nice to have skills 😉
- Observability: Experience setting up monitoring and alerting systems to ensure the health of data pipelines and infrastructure.
- Code Hygiene: A sharp sense of code hygiene, including code review, documentation, testing, and CI/CD (Continuous Integration/Continuous Delivery).
- Stream Processing Knowledge: Experience with stream processing tools like Kafka Streams, Kinesis, or Spark.
- Python's Scientific Stack: Proficiency in Python's Scientific Stack, including numpy, pandas, jupyter, matplotlib, scikit-learn, and related tools.
- English Proficiency: Solid command of the English language for writing technical documents, such as Design Documents.
🎁 Perks
- Remote-first culture – work from anywhere! 🌍
- AWS, DBT, Google Cloud, Azure & Databricks certifications fully covered
- In-Company English Lessons.
- Birthday off + an extra vacation week (Mutt Week! 🏖️)
- Referral bonuses – help us grow the team & get rewarded!
- Maslow: Monthly credits to spend in our benefits marketplace.
- ✈️🏝️ Annual Mutters' Trip – an unforgettable getaway with the team!