Software Dev Engineer I – Yahoo!
Yahoo!
About Yahoo!
Yahoo is an American web portal that provides the search engine Yahoo Search and related services including My Yahoo, Yahoo Mail, Yahoo News, Yahoo Finance, Yahoo Sports, y!entertainment, yahoo!life, and its advertising platform, Yahoo Native.
Yahoo builds, improves, and maintains one of the highest scaling platforms in the world. Their amazing team of engineers work on next-generation big data platforms that transform how users connect every single day. Yahoo’s Big Data Platform drives some of the most demanding applications in the industry. The system handles billions of requests a day and runs on some of the largest Hadoop clusters ever built! 50,000 nodes strong and several multi-thousand node clusters bring scalable computing to a whole new level. They work on problems that cover a wide spectrum – from web services to operating systems and networking layers. Their biggest challenges ahead are designing efficient cloud-native big data platforms.
About the Role
You will be part of the Big Data Solutions engineering team, contributing to the design and implementation of robust Big Data solutions on AWS, GCP, and Databricks platforms. You will be involved in the design, implementation, and testing of end-to-end solutions, ensuring sound development practices are followed to achieve well-tested technical solutions, ideally in a Unix environment. You will gain experience with containerization (ideally Kubernetes) and Big Data technologies, challenging security constraints, and myriad AWS/GCP services, with Python and Java as primary development languages.
Responsibilities
A solid working understanding of cloud technologies is a must. Knowledge of developing clean APIs and integrating with existing web services is required, as well as a solid understanding of core architectural principles and a cloud-native solutions mindset. You will contribute to projects
involving distributed systems, cloud technologies, Agentic AI workflows, and ML use cases. Experience with Spark, Airflow, Vertex AI, and Sagemaker is a plus. You will be highly motivated and able to work independently or as part of a cross-functional team on complex projects.
- Big Data platform: implementing and scaling solutions using Big Data technologies like Hadoop, Spark, Airflow, ensuring high performance and scalability.
- Data Orchestration: Utilizing tools like Airflow to manage the scheduling, execution, andmonitoring of data workflows across cloud platforms such as AWS and GCP.
- Integration and Support: Collaborating with different teams to integrate data processing, MLOps and AIOps.
- Feature Development: Implementing new features to support advanced processing capabilities, including cost and performance optimization.
- Must be self-motivated, proactive, driven and able to execute tasks both independently and within a team environment.
- Ability to comprehend and communicate in a complex technical environment.
- Understand and apply common design and architectural patterns.
- Share knowledge and experience through thorough code review processes.
- Display a history of creativity, innovative thought, awareness of current technical trends and high levels of motivation.
Required skills
- 3+ years of experience in software development.
- Programming skills: Python/Java/Golang.
- Excellent knowledge of Spark, Airflow, Terraform, AWS/GCP and cloud native services like EMR, Sagemaker, Dataproc, Bigquery, VertexAI, Airflow.
- Familiarity with Kubernetes, Docker, and Containerization.
- Operating systems: Mac/Linux.
- Bachelor’s degree in Computer Science, Software Engineering or a related field or relevant experience.
To apply, please send your resume to resume-champaign-data@yahooinc.com.