Job Description
DevOps Engineer
New York, New York (100% remote)
6-month Contract
$55-65/hr
We are seeking a highly skilled DevOps Engineer to join a Fortune 50 Broadcast Media & Entertainment leader based in New York, New York. As the DevOps Engineer, you will be responsible for building & maintaining scalable, reliable, and efficient data infrastructures. In this role, you will have the opportunity to work closely with data engineers, developers, and operations teams to ensure seamless CI/CD pipelines, automate data workflows, and support the implementation of robust monitoring and observability practices.
Minimum Qualifications:
- 2+ years of relevant experience in Data Engineering, Operations, or SRE roles.
- Experience with Terraform for infrastructure as code and automation.
- Proven experience building and maintaining CI/CD pipelines using GitHub Actions, Concourse, or similar tools.
- Hands-on experience with Apache Airflow for managing data workflows.
- Proficiency in Python, Java, Scala, R, or SQL for automating data processes and workflows.
- Familiarity with Real-Time and Batch Data Pipelines, especially in the context of Big Data Engineering
- Practical experience building distributed, scalable, and highly available systems using Google Cloud or AWS
- Experience with Kubernetes, Apache Beam, Apache Spark, Snowflake, and Databricks or similar tools.
- Strong understanding of SRE best practices for system observability, including tools like Grafana.
- Bachelor's degree in Computer Science, Engineering, Physics, or a related quantitative field (or equivalent industry experience).
- Excellent communication skills, with the ability to collaborate effectively across cross-functional teams.
Responsibilities:
- Implement and maintain infrastructure as code (IaC) using HashiCorp Terraform for scalable, reliable infrastructure management.
- Develop and maintain CI/CD pipelines, utilizing modern tools like GitHub Actions or Concourse to ensure seamless code deployment and integration.
- Work with Apache Airflow to design and manage graph-based data workflows, automating and optimizing data pipelines for both real-time and batch processing.
- Write clean, efficient, and reusable code in Python, Java, Scala, R, SQL, or similar languages to automate data processes, analysis, and workflows.
- Design and implement scalable, distributed, and highly available systems, with hands-on experience in Google Cloud and/or AWS platforms.
- Collaborate with teams to integrate Apache Spark, Apache Beam, Snowflake, Databricks, and other tools into the data pipeline architecture.
- Apply SRE best practices to ensure the observability and reliability of data pipelines, using monitoring tools such as Grafana.
- Develop and maintain monitoring and alerting systems for real-time data pipeline performance and operational health.
- Continuously improve the automation, scalability, and efficiency of data engineering processes.
- Collaborate with cross-functional teams to meet the operational and development needs of the business.
What’s in it for you?
- Work with a globally recognized media streaming organization at the forefront of innovation.
- Collaborate with high-level business professionals and technical teams, gaining valuable cross-functional experience.
- Opportunity to accelerate your career in a fast-paced, evolving industry.