Site Name: London The Stanley Building, USA - California - San Francisco, USA - Massachusetts - Cambridge, USA - Washington - Seattle-OnyxPosted Date: Dec 13 2023At GSK, we want to supercharge our data capability to better understand our patients and accelerate our ability to discover vaccines and medicines. The Onyx Research Data Platform organization represents a major investment by GSK R&D and Digital & Tech, designed to deliver a step-change in our ability to leverage data, knowledge, and prediction to find new medicines.We are a full-stack shop consisting of product and portfolio leadership, data engineering, infrastructure and DevOps, data / metadata / knowledge platforms, and AI/ML and analysis platforms, all geared toward:Building a next-generation, metadata- and automation-driven data experience for GSK's scientists, engineers, and decision-makers, increasing productivity and reducing time spent on "data mechanics"Providing best-in-class AI/ML and data analysis environments to accelerate our predictive capabilities and attract top-tier talentAggressively engineering our data at scale, as one unified asset, to unlock the value of our unique collection of data and predictions in real-timeWe are looking for a skilled Data Operations Engineer II to join our growing team. The DataOps team accelerates biomedical and scientific data product development and ensures consistent, professional-grade operations for the Data Science and Engineering organization by building templated projects (code repository plus DevOps pipelines) for various Data Science/Data Engineering architecture patterns in the challenging biomedical data space.A Data Operations Engineer II knows the metrics desired for their tools and services and iterates to deliver and improve on those metrics in an agile fashion. A Data Operations Engineer II is a technical contributor who can take a loosely defined specification for a function, pipeline, service, or other sort of component, devise a technical solution, and deliver it at a high level. They have a strong focus on operability of their tools and services, and develop, measure, and monitor key metrics for their work to seek opportunities to improve those metrics. They are aware of, and adhere to, best practice for software development in general (and their specialization in particular), including code quality, documentation, DevOps practices, and testing. They ensure robustness of our services and serve as an escalation point in the operation of existing services, pipelines, and workflows.A Data Operations Engineer II should be deeply familiar with the most common tools (languages, libraries, etc) within their specialization, and aware of the open-source communities that revolve around these tools. They should be constantly seeking feedback and guidance to further develop their technical skills and expertise and should take feedback well from all sources in the name of development.In this role you willBe a highly technical individual contributor, building modern, cloud-native systems for standardizing and templatizing data engineering:Develop and support delivery of high-performing, high-impact data ops products and services, from a loosely defined data engineering problem or requirementBuilds modular code / libraries / services / etc using tools appropriate to their area of specializationImplement data governance aligned to modern standards and technologiesProduces well-engineered software, including appropriate automated test suites and technical documentationDevelop, measure, and monitor key metrics for all tools and services and consistently seek to iterate on and improve themEnsure consistent application of platform abstractions to ensure quality and consistency with respect to logging and lineageFully versed in coding best practices and ways of working, and participates in code reviews and partner