The National Center for Biotechnology Information (NCBI) is part of the National Library of Medicine (NLM) at the National Institutes of Health (NIH). NCBI is the world's premier biomedical center hosting over six million daily users that seek research, clinical, genetic, and other information that directly impacts biomedical research and public health – at NCBI you can literally help to accelerate cures for diseases! NCBI's wide range of applications, platforms (node, python, Django, C++, you name it) and environments (big data [petabytes], machine learning, multiple clouds) serve more users than almost any other US Government Agency according to https://analytics.usa.gov/
The Sr. Data Engineer is responsible for the availability and reliability of the biological databases, as well as the development of solutions within those systems. The These can be Commercial Off the Shelf (COTS) systems or systems custom developed in house. They will primarily be working on major projects and high-profile enhancements, as well as be responsible for reviewing the deliverables of other analysts on their team.
Responsibilities
Support the development and maintenance of biological databases.
Supports technology stack supporting data management, including data warehousing, transport visualization, master data management, and other tools
Designs and provides oversight for development best practices for the data landscape at Enviva
Defines and is responsible for the execution of maintenance activities supporting their technology stack
Performs unit testing on software and/or system changes
Keeps up to date with system documentation, knowledge base articles, and potential software updates and roadmaps.
Works with a Business Analyst to translate business requirements into technical requirements
Trains business analysts and end users on proper system functionality
Keeps system documentation up to date as necessary
Participates in team agile ceremonies
Maintains, enforces, and follows ITGC and SDLC standards, as they relate to systems analysis activities
Qualifications
7+ Years of Experience
Bachelor’s Degree in Computer Science or Data Science or related required, Master’s degree preferred.
Specialized knowledge or Industry Experience
Agile Full lifecycle software development
Expert knowledge and experience in Big data and Cloud technologies – AWS, GCS or MCS, and Hadoop
Production expertise with DBMS technology, including SQLServer, Oracle, MongoDB and MySQL
Experience CI/CD building pipelines
Expertise in writing PL/SQL, T-SQL Queries, Dynamic-queries, sub-queries, Stored Procedures, Views and Cursors, CTE, temp tables, Advanced Analytical functions
Experience using Agile methodologies
Expert knowledge in tuning SQL queries, ETL process for optimal performance
Optimize pipelines, dashboards, frameworks, and systems to facilitate easier development of data artifacts.
Very strong written and oral communication skills required.
Strong computer skills, including DevOps, as well as deep knowledge for the systems they will support.
Working knowledge of solutions for various deployment models (Infrastructure as a Service (IaaS), Platform as a service (PaaS), (SaaS) and emerging cloud services to optimize the essential characteristics of cloud computing.
Ability to obtain a Public Trust security clearance.