• Stars
    star
    145
  • Rank 248,616 (Top 5 %)
  • Language
    R
  • License
    Other
  • Created over 9 years ago
  • Updated almost 2 years ago

Reviews

There are no reviews yet. Be the first to send feedback to the community and the maintainers!

Repository Details

An R package to enable reproducible data processing, packaging and sharing.

README

DataPackageR

DataPackageR is used to reproducibly process raw data into packaged, analysis-ready data sets.

CRAN R-CMD-check Coverage status Project Status: Active – The project has reached a stable, usable state and is being actively developed. DOI

Important Note: datapack is a different package that is used to “create, send and load data from common repositories such as DataONE into the R environment.”

This package is for processing raw data into tidy data sets and bundling them into R packages.

What problems does DataPackageR tackle?

You have diverse raw data sets that you need to preprocess and tidy in order to:

  • Perform data analysis
  • Write a report
  • Publish a paper
  • Share data with colleagues and collaborators
  • Save time in the future when you return to this project but have forgotten all about what you did.

Why package data sets?

Definition: A data package is a formal R package whose sole purpose is to contain, access, and / or document data sets.

  • Reproducibility.

    As described elsewhere, packaging your data promotes reproducibility. R’s packaging infrastructure promotes unit testing, documentation, a reproducible build system, and has many other benefits. Coopting it for packaging data sets is a natural fit.

  • Collaboration.

    A data set packaged in R is easy to distribute and share amongst collaborators, and is easy to install and use. All the hard work you’ve put into documenting and standardizing the tidy data set comes right along with the data package.

  • Documentation.

    R’s package system allows us to document data objects. What’s more, the roxygen2 package makes this very easy to do with markup tags. That documentation is the equivalent of a data dictionary and can be extremely valuable when returning to a project after a period of time.

  • Convenience.

    Data pre-processing can be time consuming, depending on the data type and raw data sets may be too large to share conveniently in a packaged format. Packaging and sharing the small, tidied data saves the users computing time and time spent waiting for downloads.

Challenges.

  • Package size limits.

    R packages have a 5MB size limit, at least on CRAN. BioConductor has explicit data package types that can be larger and use git LFS for very large files.

    Sharing large volumes of raw data in an R package format is still not ideal, and there are public biological data repositories better suited for raw data: e.g., GEO, SRA, ImmPort, ImmuneSpace, FlowRepository.

    Tools like datastorr can help with this and we hope to integrate the into DataPackageR in the future.

  • Manual effort

    There is still a substantial manual effort to set up the correct directory structures for an R data package. This can dissuade many individuals, particularly new users who have never built an R package, from going this route.

  • Scale

    Setting up and building R data packages by hand is a workable solution for a small project or a small number of projects, but when dealing with many projects each involving many data sets, tools are needed to help automate the process.

DataPackageR

DataPackageR provides a number of benefits when packaging your data.

  • It aims to automate away much of the tedium of packaging data sets without getting too much in the way, and keeps your processing workflow reproducible.

  • It sets up the necessary package structure and files for a data package.

  • It allows you to keep the large, raw data and only ship the packaged tidy data, saving space and time consumers of your data set need to spend downloading and re-processing it.

  • It maintains a reproducible record (vignettes) of the data processing along with the package. Consumers of the data package can verify how the processing was done, increasing confidence in your data.

  • It automates construction of the documentation and maintains a data set version and an md5 fingerprint of each data object in the package. If the data changes and the package is rebuilt, the data version is automatically updated.

Similar work

There are a number of tools out there that address similar and complementary problems:

  • datastorr github repo

    Simple data retrieval and versioning using GitHub to store data.

    • Caches downloads and uses github releases to version data.
    • Deal consistently with translating the file stored online into a loaded data object
    • Access multiple versions of the data at once

    datastorrr could be used with DataPackageR to store / access remote raw data sets, remotely store / access tidied data that are too large to fit in the package itself.

  • fst github repo

    fst provides lightning fast serialization of data frames.

  • The modern data package pdf

    A presentation from @noamross touching on modern tools for open science and reproducibility. Discusses datastorr and fst as well as standardized metadata and documentation.

  • rrrpkg github repo

    A document from ropensci describing using an R package as a research compendium. Based on ideas originally introduced by Robert Gentleman and Duncan Temple Lang (Gentleman and Lang (2004))

  • template github repo

    An R package template for data packages.

See the publication for further discussion.

Installation

You can install the latest version of DataPackageR from github with:

library(devtools)
devtools::install_github("ropensci/DataPackageR")

Blog Post - building packages interactively.

See this rOpenSci blog post on how to build data packages interactively using DataPackageR. This uses several new interfaces: use_data_object(), use_processing_script() and use_raw_dataset() to build up a data package, rather than assuming the user has all the code and data ready to go for datapackage_skeleton().

Example (assuming all code and data are available)

library(DataPackageR)

# Let's reproducibly package up
# the cars in the mtcars dataset
# with speed > 20.
# Our dataset will be called cars_over_20.
# There are three steps:

# 1. Get the code file that turns the raw data
# into our packaged and processed analysis-ready dataset.
# This is in a file called subsetCars.Rmd located in exdata/tests of the DataPackageR package.
# For your own projects you would write your own Rmd processing file.
processing_code <- system.file(
  "extdata", "tests", "subsetCars.Rmd", package = "DataPackageR"
)

# 2. Create the package framework.
# We pass in the Rmd file in the `processing_code` variable and the names of the data objects it creates (called "cars_over_20")
# The new package is called "mtcars20"
datapackage_skeleton(
  "mtcars20", force = TRUE, 
  code_files = processing_code, 
  r_object_names = "cars_over_20", 
  path = tempdir()) 

# 3. Run the preprocessing code to build the cars_over_20 data set 
# and reproducibly enclose it in the mtcars20 package.
# packageName is the full path to the package source directory created at step 2.
# You'll be prompted for a text description (one line) of the changes you're making.
# These will be added to the NEWS.md file along with the DataVersion in the package source directory.
# If the build is run in non-interactive mode, the description will read
# "Package built in non-interactive mode". You may update it later.
dir.create(file.path(tempdir(),"lib"))
package_build(packageName = file.path(tempdir(),"mtcars20"), install = TRUE, lib = file.path(tempdir(),"lib"))
#> Warning: package 'mtcars20' is in use and will not be installed

# Update the autogenerated roxygen documentation in data-raw/documentation.R. 
# edit(file.path(tempdir(),"mtcars20","R","mtcars20.R"))

# 4. Rebuild the documentation.
document(file.path(tempdir(),"mtcars20"), install = TRUE, lib = file.path(tempdir(),"lib"))
#> Warning: package 'mtcars20' is in use and will not be installed

# Let's use the package we just created.
install.packages(file.path(tempdir(),"mtcars20_1.0.tar.gz"), type = "source", repos = NULL)
#> Warning: package 'mtcars20' is in use and will not be installed
library(mtcars20)
data("cars_over_20") # load the data
cars_over_20  # Now we can use it.
?cars_over_20 # See the documentation you wrote in data-raw/documentation.R.

# We have our dataset!
# Since we preprocessed it,
# it is clean and under the 5 MB limit for data in packages.
cars_over_20

# We can easily check the version of the data
data_version("mtcars20")

# You can use an assert to check the data version in  reports and
# analyses that use the packaged data.
assert_data_version(data_package_name = "mtcars20",
                    version_string = "0.1.0",
                    acceptable = "equal")

Reading external data from within R / Rmd processing scripts.

When creating a data package, your processing scripts will need to read your raw data sets in order to process them. These data sets can be stored in inst/extdata of the data package source tree, or elsewhere outside the package source tree. In order to have portable and reproducible code, you should not use absolute paths to the raw data. Instead, DataPackageR provides several APIs to access the data package project root directory, the inst/extdata subdirectory, and the data subdirectory.

# This returns the datapackage source 
# root directory. 
# In an R or Rmd processing script this can be used to build a path to a directory that is exteral to the package, for 
# example if we are dealing with very large data sets where data cannot be packaged.
DataPackageR::project_path()

# This returns the   
# inst/extdata directory. 
# Raw data sets that are included in the package should be placed there.
# They can be read from that location, which is returned by: 
DataPackageR::project_extdata_path()

# This returns the path to the datapackage  
# data directory. This can be used to access 
# stored data objects already created and saved in `data` from 
# other processing scripts.
DataPackageR::project_data_path()

Preprint and publication.

The publication describing the package, (Finak et al., 2018), is now available at Gates Open Research .

The preprint is on biorxiv.

Code of conduct

Please note that this project is released with a Contributor Code of Conduct. By participating in this project you agree to abide by its terms.

References

  1. Gentleman, Robert, and Duncan Temple Lang. 2004. “Statistical Analyses and Reproducible Research.” Bioconductor Project Working Papers, Bioconductor project working papers,. bepress.

  2. Finak G, Mayer B, Fulp W et al. DataPackageR: Reproducible data preprocessing, standardization and sharing using R/Bioconductor for collaborative data analysis [version 1; referees: 1 approved with reservations]. Gates Open Res 2018, 2:31 (doi: 10.12688/gatesopenres.12832.1)

ropensci_footer

More Repositories

1

drake

An R-focused pipeline toolkit for reproducibility and high-performance computing
R
1,331
star
2

skimr

A frictionless, pipeable approach to dealing with summary statistics
HTML
1,095
star
3

targets

Function-oriented Make-like declarative workflows for R
R
880
star
4

rtweet

🐦 R client for interacting with Twitter's [stream and REST] APIs
R
785
star
5

tabulizer

Bindings for Tabula PDF Table Extractor Library
R
518
star
6

pdftools

Text Extraction, Rendering and Converting of PDF Documents
C++
489
star
7

magick

Magic, madness, heaven, sin
R
440
star
8

visdat

Preliminary Exploratory Visualisation of Data
R
439
star
9

stplanr

Sustainable transport planning with R
R
414
star
10

RSelenium

An R client for Selenium Remote WebDriver
R
332
star
11

rnoaa

R interface to many NOAA data APIs
R
326
star
12

osmdata

R package for downloading OpenStreetMap data
C++
307
star
13

charlatan

Create fake data in R
R
289
star
14

software-review

rOpenSci Software Peer Review.
R
279
star
15

iheatmapr

Complex, interactive heatmaps in R
R
259
star
16

taxize

A taxonomic toolbelt for R
R
250
star
17

rrrpkg

Use of an R package to facilitate reproducible research
248
star
18

elastic

R client for the Elasticsearch HTTP API
R
244
star
19

tesseract

Bindings to Tesseract OCR engine for R
R
236
star
20

qualtRics

Download ⬇️ Qualtrics survey data directly into R!
R
214
star
21

git2r

R bindings to the libgit2 library
R
212
star
22

biomartr

Genomic Data Retrieval with R
R
207
star
23

writexl

Portable, light-weight data frame to xlsx exporter for R
C
202
star
24

rnaturalearth

An R package to hold and facilitate interaction with natural earth map data 🌍
R
191
star
25

googleLanguageR

R client for the Google Translation API, Google Cloud Natural Language API and Google Cloud Speech API
HTML
190
star
26

textreuse

Detect text reuse and document similarity
R
188
star
27

piggyback

📦 for using large(r) data files on GitHub
R
179
star
28

tokenizers

Fast, Consistent Tokenization of Natural Language Text
R
179
star
29

rentrez

talk with NCBI entrez using R
R
178
star
30

rcrossref

R client for various CrossRef APIs
R
165
star
31

osmextract

Download and import OpenStreetMap data from Geofabrik and other providers
R
163
star
32

dataspice

🌶️ Create lightweight schema.org descriptions of your datasets
R
156
star
33

tic

Tasks Integrating Continuously: CI-Agnostic Workflow Definitions
R
153
star
34

rgbif

Interface to the Global Biodiversity Information Facility API
R
151
star
35

webchem

Chemical Information from the Web
R
149
star
36

geojsonio

Convert many data formats to & from GeoJSON & TopoJSON
R
148
star
37

MODIStsp

An "R" package for automatic download and preprocessing of MODIS Land Products Time Series
R
147
star
38

tsbox

tsbox: Class-Agnostic Time Series in R
R
146
star
39

dev_guide

rOpenSci Packages: Development, Maintenance, and Peer Review
R
141
star
40

ghql

GraphQL R client
R
141
star
41

jqr

R interface to jq
R
139
star
42

osfr

R interface to the Open Science Framework (OSF)
R
139
star
43

osmplotr

Data visualisation using OpenStreetMap objects
R
130
star
44

opencv

R bindings for OpenCV
C++
130
star
45

ssh

Native SSH client in R based on libssh
C
126
star
46

tarchetypes

Archetypes for targets and pipelines
R
120
star
47

RefManageR

R package RefManageR
R
114
star
48

ezknitr

Avoid the typical working directory pain when using 'knitr'
R
112
star
49

spocc

Species occurrence data toolkit for R
R
109
star
50

hunspell

High-Performance Stemmer, Tokenizer, and Spell Checker for R
C++
106
star
51

crul

R6 based http client for R (for developers)
R
102
star
52

gistr

Interact with GitHub gists from R
R
101
star
53

spelling

Tools for Spell Checking in R
R
101
star
54

weathercan

R package for downloading weather data from Environment and Climate Change Canada
R
100
star
55

rfishbase

R interface to the fishbase.org database
R
100
star
56

UCSCXenaTools

📦 An R package for accessing genomics data from UCSC Xena platform, from cancer multi-omics to single-cell RNA-seq https://cran.r-project.org/web/packages/UCSCXenaTools/
R
99
star
57

gutenbergr

Search and download public domain texts from Project Gutenberg
R
98
star
58

bib2df

Parse a BibTeX file to a tibble
R
97
star
59

ckanr

R client for the CKAN API
R
97
star
60

git2rdata

An R package for storing and retrieving data.frames in git repositories.
R
97
star
61

rsvg

SVG renderer for R based on librsvg2
C
95
star
62

EML

Ecological Metadata Language interface for R: synthesis and integration of heterogenous data
R
94
star
63

nasapower

API Client for NASA POWER Global Meteorology, Surface Solar Energy and Climatology in R
R
93
star
64

FedData

Functions to Automate Downloading Geospatial Data Available from Several Federated Data Sources
R
93
star
65

cyphr

:shipit: Humane encryption
R
92
star
66

av

Working with Video in R
C
88
star
67

mapscanner

R package to print maps, draw on them, and scan them back in
R
87
star
68

opencage

🌐 R package for the OpenCage API -- both forward and reverse geocoding 🌐
R
87
star
69

GSODR

API Client for Global Surface Summary of the Day ('GSOD') Weather Data Client in R
R
85
star
70

tidync

NetCDF exploration and data extraction
R
85
star
71

openalexR

Getting bibliographic records from OpenAlex
R
85
star
72

rzmq

R package for ZMQ
C++
82
star
73

gittargets

Data version control for reproducible analysis pipelines in R with {targets}.
R
81
star
74

rebird

Wrapper to the eBird API
R
79
star
75

bikedata

🚲 Extract data from public hire bicycle systems
R
79
star
76

vcr

Record HTTP calls and replay them
R
78
star
77

historydata

Datasets for Historians
R
78
star
78

dittodb

dittodb: A Test Environment for DB Queries in R
R
78
star
79

arkdb

Archive and unarchive databases as flat text files
R
78
star
80

fingertipsR

R package to interact with Public Health England’s Fingertips data tool
R
78
star
81

CoordinateCleaner

Automated flagging of common spatial and temporal errors in biological and palaeontological collection data, for the use in conservation, ecology and palaeontology.
HTML
77
star
82

smapr

An R package for acquisition and processing of NASA SMAP data
R
77
star
83

nodbi

Document DBI connector for R
R
76
star
84

opentripplanner

An R package to set up and use OpenTripPlanner (OTP) as a local or remote multimodal trip planner.
R
73
star
85

nlrx

nlrx NetLogo R
R
71
star
86

rb3

A bunch of downloaders and parsers for data delivered from B3
R
69
star
87

tidyhydat

An R package to import Water Survey of Canada hydrometric data and make it tidy
R
69
star
88

robotstxt

robots.txt file parsing and checking for R
R
68
star
89

slopes

Package to calculate slopes of roads, rivers and trajectories
R
65
star
90

tradestatistics

R package to access Open Trade Statistics API
R
65
star
91

unconf17

Website for 2017 rOpenSci Unconf
JavaScript
64
star
92

roadoi

Use Unpaywall with R
R
64
star
93

codemetar

an R package for generating and working with codemeta
R
64
star
94

terrainr

Get DEMs and orthoimagery from the USGS National Map, georeference your images and merge rasters, and visualize with Unity 3D
R
64
star
95

NLMR

📦 R package to simulate neutral landscape models 🏔
R
63
star
96

parzer

Parse geographic coordinates
R
63
star
97

tiler

Generate geographic and non-geographic map tiles from R
R
63
star
98

rWBclimate

R interface for the World Bank climate data
R
62
star
99

comtradr

Functions for Interacting with the UN Comtrade API
R
60
star
100

stats19

R package for working with open road traffic casualty data from Great Britain
R
59
star