Management Science, Accepted October 2025
Author(s): Xinlan Emily Hu, Mark E. Whiting, Linnea Gandhi, Duncan J. Watts, and Abdullah Almaatouq
This repository contains the data and code required to reproduce the results in The Task Space: An Integrative Framework for Team Research. This paper presents a novel multidimensional representation of tasks, in which the task performed by a group can be described along 24 quantifiable dimensions that are grounded in prior theory. It also presents a case study in which the 24-dimensional Task Space is used to select a set of 20 diverse tasks, and subsequently to predict heterogeneity in group advantage (the ability of groups to outperform an equivalent individual) across the 20 tasks.
The data and code in this paper are divided into two distinct components:
(1) The Task Map: A database of 102 tasks sourced from the interdisciplinary literature on group collaboration, annotated by human raters (sourced from Amazon Mechanical Turk) along the 24 dimensions. Here, the relevant empirical data is from the annotation process, and the analysis code transforms the raw ratings into final task dimensions.
(2) Group Advantage Integrative Experiment: A large-scale experiment in which 1,200+ workers on Amazon Mechanical Turk are recruited to complete tasks in real time. We then use the 24 task dimensions to predict the extent to which groups outperform equivalent individuals (group advantage). Here, the relevant empirical data comes from this experiment, and the analysis explores the question, when (on which tasks) do groups outperform individuals?
For more information about this work, please visit our companion website, https://taskmap.seas.upenn.edu/
This folder contains all raw input data associated with the paper (from both the Task Map annotation and the Group Advantage Integrative Experiment).
A data dictionary (data_dictionary.md) details all variables from the datasets in this folder.
Data from Task Map annotation are at the top level of the folder; data from the Group Advantage experiment, which took place in three waves, are available in subfolders Wave 1 data/, Wave 2 data/, and Wave 3 data/, respectively.
-
Information about curating the panel (e.g., rater training and pre-testing) can be found under
task space resources/rating pipelines/, with the rater panel (with real MTurk ID's hashed) and their pre-test scores included inrating pipelines/survey_workflows/our_panel.csv. -
Information about the participants in the Group Advantage study (with real MTurk ID's hashed) can be found in
data/players/.
This folder contains all analysis scripts required to reproduce key figures, tables, and findings.
-
Analysis for the Task Map can be found under
analysis_task_space/. -
Analysis for the Group Advantage study can be found under
analysis_group_advantage/. -
A master script,
run_master.py, consolidates all key analysis in a single file and logs outputs inoutputs/logs/.
This folder contains all outputs that are generated by the code in analysis/. Figures, tables, and results are reproduced here.
-
processed_data/are data files generated from the raw inputs. This includes the Task Map (ouptuts/processed_data/task_map.csv), which is the data object associated with this paper's primary contribution. -
figures/are images and visuals generated by the code. -
cached_pkls/are.pklfiles that cache objects that may take slightly longer to reproduce, such as models. -
logs/are printed outputs from running the master script.
This folder contains the primary assets for using and reproducing the crowd-annotated Task Space. It contains two subfolders: writeups/ (in which each task included in the repository is a separate Markdown file) and rating pipelines/ (which documents the process for training crowd annotators).
The top level of this folder contains cleaned up .csv files of the the 24 task dimensions, 102 tasks, and the questions used for annotation.
This folder contains the assets and code required for running the large-scale group experiments on Empirica. Its contents are a copy of an external repository that manages its version control:
- Link to External Repository: https://github.com/Watts-Lab/multi-task-empirica
- Link to the
customTasks/Folder (Implements Our 20 Tasks): https://github.com/Watts-Lab/multi-task-empirica/tree/master/multi-task-app/customTasks/
Data provenance: This paper relies on data, and all data necessary to reproduce the results of the paper are included. Below is a detailed description of how the data was obtained, allowing future researchers to reproduce the dataset.
All data originates from the authors' own data collection and is provided freely in the data/ directory for other researchers' usage in academic applications.
Data collection took place online, via Amazon Mechanical Turk. Task Space ratings were collected between July 2022 and March 2023. The Group Advantage experimental data was colleced between April 2023 and April 2024.
For detailed descriptions of the data collection process, please refer to our manuscript.
-
players/— Contains information about participants in the Group Advantage experiment, including individual-level data (individuals.csv) and player data separated by experimental wave and epoch (e.g., players_wave1_epoch1.csv). -
Wave 1 data/,Wave 2 data/, andWave 3 data/— Contain the primary experimental data for each of the three data collection waves of the Group Advantage experiment. Each wave includes one or more epochs (e.g., Epoch 1, Epoch 2) that share a common structure of CSV and JSON tables (e.g., games, players, rounds, stages, treatments, etc.). The contents and variables of these files are described in detail indata_dictionary.md. -
raw_map.csv— Contains the raw crowd annotation data used to characterize tasks along the 24 task-dimension ratings for the construction of the Task Space. -
hand_labeled_task_mcgrath_sectors.csv— Contains the hand-labeled mappings of tasks into canonical task categories following McGrath’s task typology. Used for visualization/comparison purposes.
The steps to process the Task Space rating data are detailed in analysis/analysis_task_space/generate_task_map_from_raw.R. Following these steps produces the main Task Map data object.
The steps to process the Group Advantage experimental data are detailed in analysis/analysis_group_advantage/raw_data_cleaning.R. Following these steps produces the data table used for answering the primary research question for the case study: when (on which tasks) do groups outperform individuals?
Materials associated with data collection are provided in this repository and located in the following subfolders:
This folder contains the materials documenting the process of creating the Task Space. Clean versions of the 102 task descriptions, 24 dimensions, and rater questions are provided at the top level.
The rating pipelines subfolder documents the process of recruiting and training humans to annotate the Task Map (our repository of 102 tasks, annotated along 24 quantitative dimensions by raters recruited from Amazon Mechanical Turk).
The following materials may be useful to future researchers wishing to reproduce our rating pipeline:
rating pipelines/rater_training/contains the Qualtrics survey used to instruct raters on how to annotate the tasks.rating pipelines/resources_for_rating/contains data used in the task rating process, including the questions and exact descriptions provided to the raters.rating pipelines/survey_workflows/contains code associated with grading the pre-test for filterings raters, as well as the panel of rater comprising our final pool.rating pipelines/writeup to html pipeline/contains code and materials associated with translating the written task summaries (which are stored in.mdformat in thewriteups/folder) into html files, which was the format in which they were displayed to raters.
Materials required to conduct real-time group experiments on the Empirica platform are provided in multi-task-empirica/ (which is maintained in the external repository noted above). Task details, including all stimuli and participant instructions, can be found in multi-task-empirica/multi-task-app/customTasks/. Folders in that directory correspond to the code for implementing each of the 20 tasks in the experiment.
A data dictionary for all datasets can be found in data/data_dictionary.md.
Both R and Python are required to reproduce this package. The code has been reproduced with R version 4.5.1 and Python version 3.13.7.
Nonstandard package requirements are listed separately for the two components:
-
Requirements for the Task Map can be found under
analysis_task_space/requirements.txt. -
Requirements for the Group Advantage study can be found under
analysis_group_advantage/requirements.txt.
For Python, you can easily install the dependencies correspending to each analysis using the following commands (a virtual environment is ideal):
python3 -m pip install -r analysis/analysis_group_advantage/requirements.txt
python3 -m pip install -r analysis/analysis_task_space/requirements.txtFor R, you can install the commonly used package set across this repo with one command:
Rscript scripts/install_r_dependencies.RNotes:
- This installs a superset of packages referenced across R scripts and Rmd files (Shiny apps, visualization, modeling, and utilities).
- If any package fails because it’s not on CRAN in your environment, install it manually (e.g., from GitHub) or skip if not needed for your workflow.
All programs and code associated with the project can be found in the analysis directory.
Main analyses associated with each of the two components can be reproduced by running commands via the master script (analysis/run_master.py). Use the following commands from the root of the directory:
# list available steps
python analysis/run_master.py --list
# run only the task space pipeline
python analysis/run_master.py --steps task_space
# run group advantage pipeline end-to-end
python analysis/run_master.py --steps group_advantage
# run both
python analysis/run_master.py --steps task_space,group_advantage
- Figure 4 (Heterogeneity in group advantage across conditions):
raw_data_cleaning.R - Figure 5 (Bar plots of the root mean squared errors of models predicting condition-level group advantage):
viz.ipynb(underlying models generated bymodels.ipynb) - Figure 6 (Feature importance among task dimensions):
viz.ipynb - Table 1 (A summary of the 24 dimensions included in the Task Space) is included as
task space resources/24_dimensions_clean.csv.
- Appendix Figure 4 (A heatmap in which each row represents a task, and each column represents a task dimension):
viz.ipynb - Appendix Figure 5 (A two-dimensional projection of the Task Space using PCA):
clean_task_space_visuals.R - Demographic distribution of participants (Appendix Figures 7, 9-11):
demographics_viz.ipynb - Appendix Figure 46 (Empirical difference distributions in model performance for the primary ElasticNet models):
viz.ipynb - Appendix Figure 47 (Robustness of model performance to adding noise to training data columns):
viz.ipynb - Appendix Figure 48 (Results of the second robustness check):
viz.ipynb
Remaining figures in the main manuscript are illustrations and are not generated via empirical data.
Additional analysis (associated with the Supplementary Information) can be found in the relevant subfolders under the analysis subdirectory.