We seek a skilled and versatile Senior Data Engineer with strong backend development experience and a solid foundation in DevOps practices. In this role, you will contribute to developing and maintaining our backend services, perform DevOps tasks such as CI/CD, infrastructure management, and participate in data pipeline design and implementation using Scala and Python. You should be comfortable working in a fast-paced environment and take full ownership of your tasks from development to testing and deliverings.
Key Responsibilities:
- Develop and maintain services written in Scala/Java.
- Design, build, and optimize data pipelines, integrating with various data sources and ensuring data quality.
- Manage CI/CD pipelines, infrastructure (AWS/GCP/Azure), container orchestration (Kubernetes, Docker), and system monitoring.
- Participate in code reviews, pair programming, and other agile practices to ensure high-quality code.
- Automate deployment, monitoring, and alerting tasks for system reliability.
- Collaborate with front-end engineers, data engineers, and data scientists.
- Ensure adherence to software development best practices, security guidelines, and performance standards.
Qualifications:
- Strong proficiency in Scala and the ability to design and build robust backend services.
- Experience with Python for data pipeline development, ETL processes, or data integration Solid understanding of database systems (SQL/NoSQL) and experience working with large datasets.
- Experience with K8S, writing and maintaining Helm Charts.
- Familiarity with DevOps tools and practices (CI/CD, Docker, Kubernetes, Terraform,Jenkins, etc.).
- Experience working with cloud platforms (AWS, GCP, Azure) and managing infrastructure-as-code.
- Strong problem-solving skills, with the ability to work both independently and as part of a team.
- Nice to Have:
- Experience with other programming languages like Java, JavaScript, or Node.js.
- Knowledge of big data tools (e.g., Spark, Kafka).
- Experience with Airflow, Dagster, or other data orchestration tools.
- Exposure to Machine Learning and Data Science workflows.