DataPackageR is used to reproducibly process raw data into packaged, analysis-ready data sets.

CRAN R-CMD-check Codecov test coverage Project Status: Active – The project has reached a stable, usable state and is being actively developed. DOI


You can install the latest CRAN release of DataPackageR with:


You can install the latest development version of DataPackageR from GitHub with:


What problems does DataPackageR tackle?

You have diverse raw data sets that you need to preprocess and tidy in order to:

Why package data sets?

Definition: A data package is a formal R package whose sole purpose is to contain, access, and / or document data sets.



DataPackageR provides a number of benefits when packaging your data.

Blog Post - building packages interactively.

See this rOpenSci blog post on how to build data packages interactively using DataPackageR. This uses several new interfaces: use_data_object(), use_processing_script() and use_raw_dataset() to build up a data package, rather than assuming the user has all the code and data ready to go for datapackage_skeleton().



# Let's reproducibly package up
# the cars in the mtcars dataset
# with speed > 20.
# Our dataset will be called cars_over_20.
# There are three steps:

# 1. Get the code file that turns the raw data
# into our packaged and processed analysis-ready dataset.
# This is in a file called subsetCars.Rmd located in exdata/tests of the DataPackageR package.
# For your own projects you would write your own Rmd processing file.
processing_code <- system.file(
  "extdata", "tests", "subsetCars.Rmd", package = "DataPackageR"

# 2. Create the package framework.
# We pass in the Rmd file in the `processing_code` variable and the names of the data objects it creates (called "cars_over_20")
# The new package is called "mtcars20"
  "mtcars20", force = TRUE,
  code_files = processing_code,
  r_object_names = "cars_over_20",
  path = tempdir())
#> ✔ Creating '/tmp/RtmpU0javb/mtcars20/'
#> ✔ Setting active project to '/tmp/RtmpU0javb/mtcars20'
#> ✔ Creating 'R/'
#> ✔ Writing 'DESCRIPTION'
#> Package: mtcars20
#> Title: What the Package Does (One Line, Title Case)
#> Version:
#> Authors@R (parsed):
#>     * First Last <[email protected]> [aut, cre] (YOUR-ORCID-ID)
#> Description: What the package does (one paragraph).
#> License: `use_mit_license()`, `use_gpl3_license()` or friends to pick a
#>     license
#> Encoding: UTF-8
#> Roxygen: list(markdown = TRUE)
#> RoxygenNote: 7.3.1
#> ✔ Writing 'NAMESPACE'
#> ✔ Setting active project to '<no active project>'
#> ✔ Setting active project to '/tmp/RtmpU0javb/mtcars20'
#> ✔ Added DataVersion string to 'DESCRIPTION'
#> ✔ Creating 'data-raw/'
#> ✔ Creating 'data/'
#> ✔ Creating 'inst/extdata/'
#> ✔ Copied subsetCars.Rmd into 'data-raw'
#> ✔ configured 'datapackager.yml' file

# 3. Run the preprocessing code to build the cars_over_20 data set
# and reproducibly enclose it in the mtcars20 package.
# packageName is the full path to the package source directory created at step 2.
# You'll be prompted for a text description (one line) of the changes you're making.
# These will be added to the file along with the DataVersion in the package source directory.
# If the build is run in non-interactive mode, the description will read
# "Package built in non-interactive mode". You may update it later.
package_build(packageName = file.path(tempdir(),"mtcars20"))
#> ✔ 1 data set(s) created by subsetCars.Rmd
#> • cars_over_20
#> ☘ Built all datasets!
#> Non-interactive file update.
#> * Added: cars_over_20
#> ✔ Creating 'vignettes/'
#> ✔ Creating 'inst/doc/'
#> ℹ Loading mtcars20
#> Writing 'NAMESPACE'
#> Writing 'mtcars20.Rd'
#> Writing 'cars_over_20.Rd'
#> ── R CMD build ─────────────────────────────────────────────────────────────────
#> * checking for file ‘/tmp/RtmpU0javb/mtcars20/DESCRIPTION’ ... OK
#> * preparing ‘mtcars20’:
#> * checking DESCRIPTION meta-information ... OK
#> * checking for LF line-endings in source and make files and shell scripts
#> * checking for empty or unneeded directories
#> * looking to see if a ‘data/datalist’ file should be added
#> * building ‘mtcars20_1.0.tar.gz’
#> Next Steps 
#> 1. Update your package documentation. 
#>    - Edit the documentation.R file in the package source data-raw subdirectory and update the roxygen markup. 
#>    - Rebuild the package documentation with  document() . 
#> 2. Add your package to source control. 
#>    - Call  git init .  in the package source root directory. 
#>    -  git add  the package files. 
#>    -  git commit  your new package. 
#>    - Set up a github repository for your pacakge. 
#>    - Add the github repository as a remote of your local package repository. 
#>    -  git push  your local repository to gitub.
#> [1] "/tmp/RtmpU0javb/mtcars20_1.0.tar.gz"

# Update the autogenerated roxygen documentation in data-raw/documentation.R.
# edit(file.path(tempdir(),"mtcars20","R","mtcars20.R"))

# 4. Rebuild the documentation.
#> ℹ Loading mtcars20
#> [1] TRUE

# Let's use the package we just created.
# During actual use, the temporary library does not need to be specified.
temp_lib <- file.path(tempdir(),"lib")
                 type = "source", repos = NULL, lib = temp_lib)
library(mtcars20, lib.loc = temp_lib)
data("cars_over_20") # load the data

# We have our dataset!
# Since we preprocessed it,
# it is clean and under the 5 MB limit for data in packages.
#>    speed dist
#> 44    22   66
#> 45    23   54
#> 46    24   70
#> 47    24   92
#> 48    24   93
#> 49    24  120
#> 50    25   85

?cars_over_20 # See the documentation you wrote in data-raw/documentation.R.

# We can easily check the version of the data
#> [1] '0.1.0'

# You can use an assert to check the data version in  reports and
# analyses that use the packaged data.
assert_data_version(data_package_name = "mtcars20",
                    version_string = "0.1.0",
                    acceptable = "equal")

Reading external data from within R / Rmd processing scripts.

When creating a data package, your processing scripts will need to read your raw data sets in order to process them. These data sets can be stored in inst/extdata of the data package source tree, or elsewhere outside the package source tree. In order to have portable and reproducible code, you should not use absolute paths to the raw data. Instead, DataPackageR provides several APIs to access the data package project root directory, the inst/extdata subdirectory, and the data subdirectory.

# This returns the datapackage source
# root directory.
# In an R or Rmd processing script this can be used to build a path to a directory that is exteral to the package, for
# example if we are dealing with very large data sets where data cannot be packaged.

# This returns the
# inst/extdata directory.
# Raw data sets that are included in the package should be placed there.
# They can be read from that location, which is returned by:

# This returns the path to the datapackage
# data directory. This can be used to access
# stored data objects already created and saved in `data` from
# other processing scripts.


yaml configuration guide

a more detailed technical vignette

Preprint and publication

The publication describing the package, (Finak et al., 2018), is now available at Gates Open Research .

The preprint is on bioRxiv.

Similar work

DataPackageR is for processing raw data into tidy data sets and bundling them into R packages. (Note: datapack is a different package that is used to “create, send and load data from common repositories such as DataONE into the R environment”.)

There are a number of tools out there that address similar and complementary problems:

See the publication for further discussion.

Code of conduct

Please note that this project is released with a Contributor Code of Conduct. By participating in this project you agree to abide by its terms.


  1. Gentleman, Robert, and Duncan Temple Lang. 2004. “Statistical Analyses and Reproducible Research.” Bioconductor Project Working Papers, Bioconductor project working papers,. bepress.

  2. Finak G, Mayer B, Fulp W, et al. DataPackageR: Reproducible data preprocessing, standardization and sharing using R/Bioconductor for collaborative data analysis. Gates Open Res 2018, 2:31 (DOI: 10.12688/gatesopenres.12832.1)