-
Notifications
You must be signed in to change notification settings - Fork 1
/
manuscript.tex
139 lines (86 loc) · 16.7 KB
/
manuscript.tex
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
% Modified from the eLife template, which is (C) eLife 2017 and released under the Creative Commons CC BY 4.0 license
% This template is (C) R. Stuart Geiger 2020 and released under the Creative Commons CC BY 4.0 license
\documentclass[11pt]{elife}
\usepackage{ragged2e}
\usepackage{authblk}
\setlength\parindent{0pt}
\setlength{\parskip}{1em}
\titlespacing\section{0pt}{0pt}{0pt}
\title{LaTeX Template for BIDS Best Practices Reports}
\subtitle{A Report from the Berkeley Institute for Data Science's \textit{Best Practices in Data Science Series}}
\author[1\authfn{1}*]{R. Stuart Geiger}
\author[1,10\authfn{1}]{Dan Sholler}
\author[3\authfn{2}]{Aaron Culich}
\author[1,4\authfn{2}]{Ciera Martinez}
\author[5\authfn{2}]{Fernando Hoces de la Guardia}
\author[1,8,9\authfn{2}]{Fran\c{c}ois Lanusse}
\author[1,2\authfn{2}]{Kellie Ottoboni}
\author[1,6\authfn{2}]{Marla Stuart}
\author[1,7\authfn{2}]{Maryam Vareth}
\author[1,2\authfn{2}]{Nelle Varoquaux}
\author[1,2\authfn{2}]{Sara Stoudt}
\author[1\authfn{2}]{St\'efan van der Walt}
\affil[1]{ Berkeley Institute for Data Science, University of California, Berkeley}
\affil[2]{ Department of Statistics, University of California, Berkeley}
\affil[3]{ D-Lab, University of California, Berkeley}
\affil[4]{ Department of Molecular and Cell Biology, University of California, Berkeley}
\affil[5]{ Berkeley Initiative for Transparency in the Social Sciences, University of California, Berkeley}
\affil[6]{ School of Social Welfare, University of California, Berkeley}
\affil[7]{ Department of Radiology and Biomedical Imaging, University of California, San Francisco}
\affil[8]{ Berkeley Center for Cosmological Physics, University of California, Berkeley}
\affil[9]{ Foundations of Data Analysis Institute, University of California, Berkeley}
\affil[10]{ rOpenSci}
\contrib[*]{Corresponding author: [email protected]}
\contrib[\authfn{1}]{{These authors contributed equally to this work}}
\contrib[\authfn{2}]{{These authors contributed equally to this work, order is alphabetical}}
\contrib{\bigskip \textbf{Published}: 21 May 2020}
\contrib{ \textbf{DOI}: \href{https://doi.org/10.17605/osf.io/ctfqn}{10.17605/osf.io/ctfqn} }
\contrib{ \textbf{License}: \href{https://creativecommons.org/licenses/by/4.0/}{Creative Commons \\ Attribution (CC BY 4.0 Intl)}}
%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%
%%% ARTICLE START
%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%
\begin{document}
\maketitle
\begin{abstract}
\justify
This is the LaTeX template for the reports produced by the BIDS Best Practices in Data Science group, which has been modified from a freely-licensed template released by the editors of eLife \citep{elife_template}. You should feel free to further adapt and modify this template for your own work, as you see fit. This field is for the abstract, which can be a decent length, but not so long --- especially if you have a lot of authors. The front page also has space for a ``recommended citation'' element below, which is currently produced manually. The remainder of this article and the author list have been taken from our first publication, ``Challenges of Doing Data-Intensive Research in Teams, Labs, and Groups: Report from the BIDS Best Practices in Data Science Series.''
\end{abstract}
\small \noindent \textbf{Recommended citation:} Dan Sholler, Sara Stoudt, Chris Kennedy, Fernando Hoces de la Guardia, François Lanusse, Karthik Ram, Kellie Ottoboni, Marla Stuart, Maryam Vareth, Nelle Varoquaux, Rebecca Barter, R. Stuart Geiger, Scott Peterson, and Stéfan van der Walt. "Best Practices Template." \textit{BIDS Best Practices in Data Science Series.} Berkeley Institute for Data Science: Berkeley, California. 2019. \href{http://doi.org/10.31235/osf.io/qr8cz}{10.31235/osf.io/qr8cz}
\normalsize
\clearpage
\section{Introduction}
This paper is a summary of the first meeting of the Best Practices lunch and discussion series, sponsored by the Berkeley Institute for Data Science (BIDS), in which we bring people together from across the UC-Berkeley campus and beyond to discuss a particular challenge or issue in doing data-intensive research. The goal of the series is to informally share experiences and ideas on how to do data science well (or at least better) from many disciplines and contexts. The topic for this week was doing data-intensive research in teams, labs, and other groups. For this first meeting, we focused on just identifying and diagnosing the many different kinds of challenges. In future meetings, we will dive deeper into some of these specific issues and try to identify best practices for dealing with them.
We prepared for this series by reviewing many of the papers and series around ``best practices'' in scientific computing \citep[e.g.][]{Wilson2014,Noble2009}, ``good enough practices'' \citep{Wilson2017}, and PLOS Computational Biology’s ``ten simple rules'' series \citep[e.g.][]{Sandve2013,Goodman2014,PerezRiverol2016}. There is also an extensive literature relevant to collaboration and teamwork in data-intensive groups, including work in ethnography of scientific labs \citep[e.g.][]{traweek1992}, library and information science \citep[e.g.][]{borgman2015}, team science \citep[e.g.][]{NAP19007}, and industry guides \citep[e.g.][]{patil2011building}. We also see this series as a successor to the collection of case studies in reproducible research published by several BIDS fellows \citep{Kitzes2018}. One reason we chose to identify issues with doing data science in teams and groups is because many of us felt like we understood how to best practice data-intensive research individually, but struggled with how to do this well in teams and groups.
\section{Compute and data challenges}
\subsection{Getting on the same stack}
Some of the major challenges in doing data-intensive research in teams is around technology use, particularly in using the same tools. Today’s computational researchers have an overwhelming number of options to choose in terms of programming languages, software libraries, data formats, operating systems, compute infrastructures, version control systems, collaboration platforms, and more. One of the major challenges often faced is that members of a team often have been trained to work with different technologies, which also often come with their own ways of working on a problem. Getting everyone on the same technical stack often takes far more time than is anticipated, and new members can spend much time learning to work in a new stack.
One of the biggest divides our group had experienced was in the choice of using programming languages, as many of us were more comfortable with either R or Python. These programming languages have their own extensive software libraries, like the tidyverse~\citep{wickham2017tidyverse} vs. the numpy/pandas/Matplotlib stack~\citep{vanderwalt2011,Hunter2007}. There are also many different software environments to choose from at various layers of the stack, from development environments like Jupyter notebooks versus RStudio and RMarkdown to the many options for package and dependency management. While most of the people in the room were committed to open source languages and environments, many people are trained to use proprietary software like MATLAB or SPSS, which raises an additional challenge in teams and groups.
Another major issue is where the actual computing and data storage will take place. Members of a team often come in knowing how to run code on their own laptops, but there are many options for groups to work, including a lab’s own shared physical server, campus clusters, national grid/supercomputer infrastructures, corporate cloud services, and more.
\subsection{Workflow and pipeline management}
Getting everyone to use an interoperable software and hardware environment is as much of a social challenge as it is a technical one, and we had a great discussion about whether a group leader should (or could) require members to use the same language, environment, or infrastructure. One of the technical solutions to this issue --- working in staged data analysis pipelines --- comes with its own set of challenges. With staged pipelines, data processing and analysis tasks are separated into modular tasks that an individual can solve in their own way, then output their work to a standardized file for the next stage of the pipeline to take as input.
The ideal end goal is often imagined to be a fully-automated (or ‘one click’) data processing and analysis pipeline, but this is difficult to achieve and maintain in practice. Several people in our group said they personally spend substantial amounts of time setting up these pipelines and making sure that each person’s piece works with everyone else’s. Even with groups that had formalized detailed data management plans, a common theme was that someone had to constantly make sure that team members were actually following these standards so that the pipeline keep running.
\subsection{External handoffs to and from the team}
Many of the research projects we discussed involved not only handoffs between members of the team, but also handoffs between the team and external groups. The “raw” data a team begins with is often the final output of another research team, government agency, or company. In these cases, our group discussed issues that ranged from technical to social, from data formats that are technically difficult to integrate at scale (like Excel spreadsheets) to not having adequate documentation to be able to interpret what the data actually means. Similarly, teams often must deliver data to external partners, who may have very different needs, expectations, and standards than the team has for itself. Finally, some teams have sensitive data privacy issues and requirements, which makes collaboration even more difficult. How can these external relationships be managed in mutually beneficial ways?
\section{Team management challenges}
Beyond technical challenges, a number of management issues face research groups aspiring to implement best practices for data-intensive research. Our discussion highlighted the difficulties of composing a well-balanced team, of dealing with fluid membership, and of fostering generative coordination and communication among group members.
\subsection{Composing a well-balanced team}
Data-intensive research groups require a team with varied expertise. A consequence of varied expertise is varied capabilities and end goals, so project leads must devote attention to managing team composition. Whereas one or two members might be capable of carrying out tasks across the various stages of research, others might specialize in a particular area. How then can research groups ensure that no one member of the team departing would collapse the project and that the team holds the necessary expertise to accomplish the shared research goal? Furthermore, some members may participate simply to acquire skills, while others seek to establish or build an academic track record. How might groups achieve alignment between personal and team goals?
\subsection{Dealing with voluntary and fluid membership}
A practical management problem also relates to the quasi-voluntary and fluid nature of research groups. Research groups largely rely extensively on students and postdocs, with an expectation that they join the team temporarily to gain new skills and experience, then leave. Many members also only work part-time for a research group, with other professional obligations (like classes) that make real-time collaboration difficult. In the long-term, turnover becomes a problem when processes, practices, and tacit institutional knowledge are difficult to standardize or document. What strategies might project leads employ to alleviate the difficulties associated with voluntary, fluid membership? What kinds of collaboration platforms, documentation practices, data management strategies, and workflows best support groups with regular turnover?
\subsection{Fostering open and inclusive coordination and communication}
The issues of team composition and voluntary or fluid membership raise a third challenge: fostering open and inclusive communication among group members. Previous research and guidelines for managing teams \citep{Edmondson1999,GooglereWork2017} emphasize the vital role of psychological safety in ensuring that team members share knowledge and collaborate effectively. Adequate psychological safety ensures that team members are comfortable speaking up about their ideas and welcoming of others’ feedback. Yet fostering psychological safety is a difficult task when research groups comprise members with various levels of expertise, career experience, and, increasingly, communities of practice (as in the case of data scientists working with domain experts). How can projects establish avenues for open communication between diverse members?
\subsection{Not abandoning best practices when deadlines loom}
One of the major issues that resonated across our group was the tendency for a team to stop following various best practices when deadlines rapidly approach. In the rush to do everything that is needed to get a publication submitted, it is easy to accrue what software engineers call “technical debt.” For example, substantial “collaboration debt” or “reproducibility debt” can be foisted on a team when a member works outside of the established workflow to produce a figure or fails to document their changes to analysis code. These stressful moments can also be difficult for the team’s psychological safety, particularly if there is an expectation to work late hours to make the deadline.
\section{Concluding thoughts and plans}
\subsection{Are there universal best practices for all cases and contexts?}
At the conclusion of our first meeting, we evaluated topics for future discussions, thinking about identifying potential solutions to the challenges faced by data-intensive research groups. In doing so, we were quickly confronted with the diversity of technologies, research agendas, disciplinary norms, team compositions, and governance structures, and other factors that characterize scientific research groups. Are solutions that work for large teams appropriate for smaller teams? Do cross-institutional or inter-disciplinary teams face different problems than those working in the same institution or discipline? Are solutions that work in astronomy or physics appropriate for ecology or social sciences? Dealing with such diversity and contextuality, then, might require adjusting our line of inquiry to the following question: At what level should we attempt to generalize best practices?
\subsection{Our future plans}
The differences within and between research groups are meaningful and deserve adequate attention, but commonalities do exist. This semester, our group will aggregate and develop input from a diverse community of practitioners to construct sets of thoughtful, grounded recommendations. For example, we will aim to provide recommendations on issues such as how to build and maintain pipelines and workflows, as well as strategies for achieving diversity and inclusion in teams. In our next post, we will offer some insights on how to manage the common problem of perpetual turnover in team membership. On all topics, we welcome feedback and recommendations.
\subsection{Combating impostor syndrome}
Finally, many people who attended told us afterwards how positive and valuable it was to share these kinds of issues and experiences, particularly for combating the “impostor syndrome” that many of us often feel. In scientific research, we typically only present the final end-product of research. Even sharing one’s final code and data in perfectly reproducible pipelines can smooth over and obscure all the messy, complex, and challenging work that inevitably takes place in any research process \citep{neff2017critique}. And terms like ``data science'' and ``machine learning'' are often presented to the public as kinds of magic rather than as kinds of methods, which can set dangerous expectations \citep{elish2018}.
In our group, people appreciated hearing others talk openly about the difficulties and challenges that come with doing data-intensive research and how they tried to deal with them. The format of sharing challenges followed by strategies for dealing with those challenges may be a meta-level best practice for this work, versus the more standard approach of listing more abstract rules and principles. Through these kinds of conversations, we hope to continue to shed light on the doing of data science in ways that will be constructive and generative across the many fields, areas, and contexts in which we work.
\vspace{-10px}
\section*{Acknowledgments}
This work was supported by the Gordon \& Betty Moore Foundation (GBMF3834) and Alfred P. Sloan Foundation (2013-10-27), as part of the Moore-Sloan Data Science Environments.
\bibliography{biblio}
\end{document}