Ep 32: Best Practices - Pipeline Development, Part Two
Update: 2024-03-23
Description
PHA4GE Ten Best Practices for Public Health Bioinformatics Pipelines:
https://github.com/pha4ge/public-health-pipeline-best-practices/blob/main/docs/pipeline-best-practices.md
Summary
In this episode, Kevin Libuit and Andrew Page discuss the 10 best practices for public health pipeline development. They start by emphasizing the use of common file formats and the importance of avoiding reinventing the wheel. They highlight the benefits of standard file formats and the availability of parsers for different languages. They also discuss the implementation of software testing, including the use of automated testing and the integration of testing with Docker containers. They emphasize the need for accessibility to benchmark or validation data sets and the importance of reference data requirements. They also touch on the significance of hiring bioinformaticians and the documentation practices that should be followed.
Takeaways
Use common file formats to avoid reinventing the wheel and enable compatibility with other programs.
Implement software testing, including automated testing, to ensure functionality and identify bugs.
Provide benchmark or validation data sets to allow users to compare and evaluate the performance of the pipeline.
Consider the reference data requirements and ensure accessibility to curated databases.
Hire bioinformaticians with domain expertise to navigate the complexities of pipeline development.
Follow documentation practices, including communication of authorship, pipeline maintenance statements, and community guidelines for contribution and support.
https://github.com/pha4ge/public-health-pipeline-best-practices/blob/main/docs/pipeline-best-practices.md
Summary
In this episode, Kevin Libuit and Andrew Page discuss the 10 best practices for public health pipeline development. They start by emphasizing the use of common file formats and the importance of avoiding reinventing the wheel. They highlight the benefits of standard file formats and the availability of parsers for different languages. They also discuss the implementation of software testing, including the use of automated testing and the integration of testing with Docker containers. They emphasize the need for accessibility to benchmark or validation data sets and the importance of reference data requirements. They also touch on the significance of hiring bioinformaticians and the documentation practices that should be followed.
Takeaways
Use common file formats to avoid reinventing the wheel and enable compatibility with other programs.
Implement software testing, including automated testing, to ensure functionality and identify bugs.
Provide benchmark or validation data sets to allow users to compare and evaluate the performance of the pipeline.
Consider the reference data requirements and ensure accessibility to curated databases.
Hire bioinformaticians with domain expertise to navigate the complexities of pipeline development.
Follow documentation practices, including communication of authorship, pipeline maintenance statements, and community guidelines for contribution and support.
Comments
Top Podcasts
The Best New Comedy Podcast Right Now – June 2024The Best News Podcast Right Now – June 2024The Best New Business Podcast Right Now – June 2024The Best New Sports Podcast Right Now – June 2024The Best New True Crime Podcast Right Now – June 2024The Best New Joe Rogan Experience Podcast Right Now – June 20The Best New Dan Bongino Show Podcast Right Now – June 20The Best New Mark Levin Podcast – June 2024
In Channel