
We have moved our website to embl.org/events. The content below is no longer being updated.
EMBL Courses and Conferences during the Coronavirus pandemic
With the onsite programme paused, many of our events are now being offered in virtual formats.
Registration is open as usual for many events, with back-up plans in place to move further courses and conferences online as necessary. Registration fees for any events affected by the COVID-19 disruption are fully refundable.
More information for participants of events at EMBL Heidelberg can be found here.
Course Overview
Computation is an integral part of today's research as data has grown too large or too complex to be analysed by hand. An ever-growing fraction of science is performed computationally and many wet-lab biologists spend part of their time on the computer. Many scientists struggle with this aspect of research as they have not been properly trained in the necessary set of skills. The result is that too much time is spent using inefficient tools when progress could be faster. This course provides training in several key tools, with a focus on good development practices that encourage efficient and reproducible research computing.
Audience
This is a course for researchers in the life sciences who are using computers for their analyses, even if not full time. The target student will be familiar with some command line/programmatic computer usage, will want to become more confident using these tools efficiently and reproducibly. A target student will have written a for loop in some language before, but will not know what git is (or at least not be very comfortable using git).
Modules/Resources
- Introduction to Python scripting
- Introduction to the Unix shell and usage of cluster resources
- Version control with Git and Github
- Analysis pipeline management
- Scientific Python & working with biological data
- Literate programming with Jupyter notebooks
Learning Outcomes
After the workshop, participants will be able to:
- write and organise their own scripts for data analysis
- use version control to keep track of, and revert changes to their files
- implement efficient and reproducible pipelines that combine multiple tools and scripts