BioFire Diagnostics, LLC. ETL Developer I in Salt Lake City, Utah
BioFire Diagnostics, LLC. is looking to add an ETL Developer I to our team!
ETL Developer will analyze, design, develop, test and maintain the data pipeline component of BioFire’s Data Warehouse. This includes refinement of stored procedures to improve daily and real-time data loads in effort to make them more efficient and scalable.
ETL Developer will work closely with BI Analysts and Engineers to build a Data Warehouse that provides quality, accurate, accessible and governed data capable to deliver business insight to various BioFire business domains through reporting, dashboards and visualizations
ETL Developer will promote Kimball methodology and utilize current technology stack to find innovative solutions to complex data problems. Ability to work as a team player alongside a talented group of individuals for the same cause is essential to success in the role.
Principal Job Duties and Responsibilities:
Perform all work in compliance with company policy and within the guidelines of BioFire’s Quality System.
Understand BioFire’s technical and scientific mission.
Understand technical and functional components of BioFire Diagnostics
Perform role in development of BI infrastructure including design, build, management, maintenance, and optimization of a data warehouse
Achieve optimal target architecture by analyzing current ETL landscape and determine if pipelines are best handled through an ETL tool, stored procedures or a combination of both
Participate in requirements, design, and analysis sessions to ensure sound team decision-making and effective ETL strategy that aligns with BioFire company objectives
Adapt ETL processes to accommodate changes in source systems and new business user requirements
Develop, test, monitor and troubleshoot ETL processes
Automate the ETL process through scheduling and exception-handling routines as well as source to target mapping development, support and maintenance
Thoroughly document ETL process and overall Data Warehouse landscape within Data Catalog
Participate in collaboration efforts with global BI development teams (e.g. bioMerieux France, Cognizant development team in India)
Collaborate with internal data analyst/science teams to provide accurate analytics to business and a data warehouse/lake that allows for flexibility and usability
Where applicable, build out new and useful technologies within our Microsoft Azure environment such as Azure Databricks, Azure Machine Learning, and big data technologies such as Hive, Spark, and Kafka
Prioritize and complete data requests in a timely manner
Complete projects on time with minimal supervision.
Continually update technical knowledge and skills.
Participate in sprint planning, standup and retrospective meetings as needed
Domestic and international travel may be required.
Recommend process improvement opportunities related to BI and the Data Warehouse
BI is an integral member of the data architecture team. ETL Developer will contribute meaningfully to CAB discussions and data governance discussions to ensure we have a Data Warehouse that provides quality, accurate, accessible, and governed data
Training and Education:
BS degree in Computer Science, IS, IT or related field required; advanced degree preferred
Professional BI certifications preferred
0-2 years of relevant experience required
Experience in BI / data integration and solution implementation projects required (BI architecture and design, solution implementation, and/or data warehouse development); MS Azure, Amazon AWS, and/or SAP Hana (SCP) preferred
Previous experience in BI / data integration and implementation projects required (BI architecture and design, solution implementation, and/or data warehouse development)
Understands ERP and CRM systems data infrastructure and how it relates to transformation requirements (data warehouse) for analysis.
Expert level in SQL (MS SQL/TSQL, MySQL, and/or Amazon Redshift SQL)
Experience with one or more ETL Tools preferred (SAP Data Services, MS Data Factory, Oracle Data Integrator, Informatica, SSIS)
Experience with stored procedures
Experience with Data Warehouse design methodologies (Kimball)
Experience in BI Software (Power BI, DOMO, Tableau, etc.) preferred
Experience in one of the following programming/analytical languages preferred: R, Python, C#, Java
Experience working with data from SAP and Salesforce a plus
Experience with Jira preferred
Proficient in MS Office Suite and flow charts software
Drive for continuous improvement
Must demonstrate ability to communicate effectively and work well with team
Must demonstrate ability to complete objectives without high levels of supervision.