Archetypes for Targets
Function-oriented Make-like declarative pipelines for Statistics and data science are supported in the targets R package. As an extension to targets, the tarchetypes package provides convenient user-side functions to make targets easier to use. By establishing reusable archetypes for common kinds of targets and pipelines, these functions help express complicated reproducible pipelines concisely and compactly. The methods in this package were influenced by the drake R package by Will Landau (2018) doi:10.21105/joss.00550.
View DocumentationrOpenSci Review Roclets
Dynamic Function-Oriented Make-Like Declarative Pipelines
Pipeline tools coordinate the pieces of computationally demanding analysis projects. The targets package is a Make-like pipeline tool for statistics and data science in R. The package skips costly runtime for tasks that are already up to date, orchestrates the necessary computation with implicit parallel computing, and abstracts files as R objects. If all the current output matches the current upstream code and data, then the whole pipeline is up to date, and the results are more trustworthy than otherwise. The methodology in this package borrows from GNU Make (2015, ISBN:978-9881443519) and drake (2018, doi:10.21105/joss.00550).
View DocumentationSimple Git Client for R
Simple git client for R based on libgit2 https://libgit2.org with support for SSH and HTTPS remotes. All functions in gert use basic R data types (such as vectors and data-frames) for their arguments and return values. User credentials are shared with command line git through the git-credential store and ssh keys stored on disk or ssh-agent.
View DocumentationKeep a Collection of Sparkly Data Resources
Tools to get and maintain a data repository from third-party data providers.
View DocumentationRenders a Multilingual Quarto Book
Automate rendering and cross-linking of Quarto books following a prescribed structure.
View DocumentationConstruct Reproducible Analytic Data Sets as R Packages
A framework to help construct R data packages in a reproducible manner. Potentially time consuming processing of raw data sets into analysis ready data sets is done in a reproducible manner and decoupled from the usual R CMD build process so that data sets can be processed into R objects in the data package and the data package can then be shared, built, and installed by others without the need to repeat computationally costly data processing. The package maintains data provenance by turning the data processing scripts into package vignettes, as well as enforcing documentation and version checking of included data objects. Data packages can be version controlled on GitHub, and used to share data for manuscripts, collaboration and reproducible research.
View DocumentationPlumber API to report package structure and function
Simple Jenkins Client for R
Manage jobs and builds on your Jenkins CI server https://jenkins.io/. Create and edit projects, schedule builds, manage the queue, download build logs, and much more.
View Documentation//r-universe.dev>
Utilities to interact with the R-universe platform. Includes functions to manage local package repositories, as well as API wrappers for retrieving data and metadata about packages in r-universe.
View DocumentationTools for Managing SSH and Git Credentials
Setup and retrieve HTTPS and SSH credentials for use with git and other services. For HTTPS remotes the package interfaces the git-credential utility which git uses to store HTTP usernames and passwords. For SSH remotes we provide convenient functions to find or generate appropriate SSH keys. The package both helps the user to setup a local git installation, and also provides a back-end for git/ssh client libraries to authenticate with existing user credentials.
View DocumentationR Bindings for ZeroMQ
Interface to the ZeroMQ lightweight messaging kernel (see https://zeromq.org/ for more information).
View DocumentationA Modern and Easy-to-Use Crypto Library
Bindings to libsodium https://doc.libsodium.org/: a modern, easy-to-use software library for encryption, decryption, signatures, password hashing and more. Sodium uses curve25519, a state-of-the-art Diffie-Hellman function by Daniel Bernstein, which has become very popular after it was discovered that the NSA had backdoored Dual EC DRBG.
View DocumentationMinimal and Uncluttered Package Documentation
Generates simple and beautiful one-page HTML reference manuals with package documentation. Math rendering and syntax highlighting are done server-side in R such that no JavaScript libraries are needed in the browser, which makes the documentation portable and fast to load.
View DocumentationServer-Side Syntax Highlighting
Prism https://prismjs.com/ is a lightweight, extensible syntax highlighter, built with modern web standards in mind. This package provides server-side rendering in R using V8 such that no JavaScript library is required in the resulting HTML documents. Over 400 languages are supported.
View DocumentationTargets for JAGS Pipelines
Bayesian data analysis usually incurs long runtimes and cumbersome custom code. A pipeline toolkit tailored to Bayesian statisticians, the jagstargets R package is leverages targets and R2jags to ease this burden. jagstargets makes it super easy to set up scalable JAGS pipelines that automatically parallelize the computation and skip expensive steps when the results are already up to date. Minimal custom code is required, and there is no need to manually configure branching, so usage is much easier than targets alone. For the underlying methodology, please refer to the documentation of targets doi:10.21105/joss.02959 and JAGS (Plummer 2003) https://www.r-project.org/conferences/DSC-2003/Proceedings/Plummer.pdf.
View DocumentationWork with BEAST2 Packages
BEAST2 (https://www.beast2.org) is a widely used Bayesian phylogenetic tool, that uses DNA/RNA/protein data and many model priors to create a posterior of jointly estimated phylogenies and parameters. BEAST2 is commonly accompanied by BEAUti 2 (https://www.beast2.org), which, among others, allows one to install BEAST2 package. This package allows to work with BEAST2 packages from R.
View DocumentationCall BEAST2
BEAST2 (https://www.beast2.org) is a widely used Bayesian phylogenetic tool, that uses DNA/RNA/protein data and many model priors to create a posterior of jointly estimated phylogenies and parameters. BEAST2 is a command-line tool. This package provides a way to call BEAST2 from an R function call.
View DocumentationSetup, Run and Analyze NetLogo Model Simulations from R via XML
Setup, run and analyze NetLogo (https://ccl.northwestern.edu/netlogo/) model simulations in R. nlrx experiments use a similar structure as NetLogos Behavior Space experiments. However, nlrx offers more flexibility and additional tools for running and analyzing complex simulation designs and sensitivity analyses. The user defines all information that is needed in an intuitive framework, using class objects. Experiments are submitted from R to NetLogo via XML files that are dynamically written, based on specifications defined by the user. By nesting model calls in future environments, large simulation design with many runs can be executed in parallel. This also enables simulating NetLogo experiments on remote high performance computing machines. In order to use this package, Java and NetLogo (>= 5.3.1) need to be available on the executing system.
View Documentationpkgdown template and utilities for rOpenSci docs
This is a private template for use by rOpenSci packages. Please don’t use it for your own non-rOpenSci package.
View DocumentationrOpenSci Package Checks
Check whether a package is ready for submission to rOpenSci’s peer review system.
View DocumentationMetrics of R Packages
Static code analyses for R packages using the external code-tagging libraries ctags and gtags. Static analyses enable packages to be analysed very quickly, generally a couple of seconds at most. The package also provides access to a database generating by applying the main function to the full CRAN archive, enabling the statistical properties of any package to be compared with all other CRAN packages.
View DocumentationR Client Package for Circle CI
Tools for interacting with the Circle CI API (https://circleci.com/docs/api/v2/). Besides executing common tasks such as querying build logs and restarting builds, this package also helps setting up permissions to deploy from builds.
View DocumentationModel Comparison Using babette
BEAST2 (https://www.beast2.org) is a widely used Bayesian phylogenetic tool, that uses DNA/RNA/protein data and many model priors to create a posterior of jointly estimated phylogenies and parameters. mcbette allows to do a Bayesian model comparison over some site and clock models, using babette (https://github.com/ropensci/babette/).
View DocumentationCreate Useful .gitignore Files for your Project
Simple interface to query gitignore.io to fetch gitignore templates that can be included in the .gitignore file. More than 450 templates are currently available.
View DocumentationAn R Client for the ODK Central API
Access and tidy up data from the ODK Central API. ODK Central is a clearinghouse for digitally captured data https://docs.getodk.org/central-intro/. ODK Central and its API are documented at https://docs.getodk.org/.
View DocumentationA Pipeline Toolkit for Reproducible Computation at Scale
A general-purpose computational engine for data analysis, drake rebuilds intermediate data objects when their dependencies change, and it skips work when the results are already up to date. Not every execution starts from scratch, there is native support for parallel and distributed computing, and completed projects have tangible evidence that they are reproducible. Extensive documentation, from beginner-friendly tutorials to practical examples and more, is available at the reference website https://docs.ropensci.org/drake/ and the online manual https://books.ropensci.org/drake/.
View DocumentationTargets for Stan Workflows
Bayesian data analysis usually incurs long runtimes and cumbersome custom code. A pipeline toolkit tailored to Bayesian statisticians, the stantargets R package leverages targets and cmdstanr to ease these burdens. stantargets makes it super easy to set up scalable Stan pipelines that automatically parallelize the computation and skip expensive steps when the results are already up to date. Minimal custom code is required, and there is no need to manually configure branching, so usage is much easier than targets alone. stantargets can access all of cmdstanrs major algorithms (MCMC, variational Bayes, and optimization) and it supports both single-fit workflows and multi-rep simulation studies. For the statistical methodology, please refer to Stan’ documentation (Stan Development Team 2020) https://mc-stan.org/.
View DocumentationData Version Control for the Targets Package
In computationally demanding data analysis pipelines, the targets R package (2021, doi:10.21105/joss.02959) maintains an up-to-date set of results while skipping tasks that do not need to rerun. This process increases speed and increases trust in the final end product. However, it also overwrites old output with new output, and past results disappear by default. To preserve historical output, the gittargets package captures version-controlled snapshots of the data store, and each snapshot links to the underlying commit of the source code. That way, when the user rolls back the code to a previous branch or commit, gittargets can recover the data contemporaneous with that commit so that all targets remain up to date.
View DocumentationCI-Agnostic Workflow Definitions
Provides a way to describe common build and deployment workflows for R-based projects: packages, websites (e.g. blogdown, pkgdown), or data processing (e.g. research compendia). The recipe is described independent of the continuous integration tool used for processing the workflow (e.g. GitHub Actions or Circle CI). This package has been peer-reviewed by rOpenSci (v0.3.0.9004).
View DocumentationAcknowledge all Contributors to a Project
Acknowledge all contributors to a project via a single function call. The function appends to a README or other specified file(s) a table with names of all individuals who contributed via code or repository issues. The package also includes several additional functions to extract and quantify contributions to any repository.
View DocumentationAutomatic Package Testing
BEAUti from R
BEAST2 (https://www.beast2.org) is a widely used Bayesian phylogenetic tool, that uses DNA/RNA/protein data and many model priors to create a posterior of jointly estimated phylogenies and parameters. BEAUti 2 (which is part of BEAST2) is a GUI tool that allows users to specify the many possible setups and generates the XML file BEAST2 needs to run. This package provides a way to create BEAST2 input files without active user input, but using R function calls instead.
View DocumentationInterface to the Open Science Framework (OSF)
An interface for interacting with OSF (https://osf.io). osfr enables you to access open research materials and data, or create and manage your own private or public projects.
View DocumentationCast (R)Markdown Files to XML and Back Again
Parsing (R)Markdown files with numerous regular expressions can be fraught with peril, but it does not have to be this way. Converting (R)Markdown files to XML using the commonmark package allows in-memory editing via of markdown elements via XPath through the extensible R6 class called yarn. These modified XML representations can be written to (R)Markdown documents via an xslt stylesheet which implements an extended version of GitHub-flavoured markdown so that you can tinker to your hearts content.
View DocumentationControl BEAST2
BEAST2 (https://www.beast2.org) is a widely used Bayesian phylogenetic tool, that uses DNA/RNA/protein data and many model priors to create a posterior of jointly estimated phylogenies and parameters. BEAST2 is commonly accompanied by BEAUti 2, Tracer and DensiTree. babette provides for an alternative workflow of using all these tools separately. This allows doing complex Bayesian phylogenetics easily and reproducibly from R.
View DocumentationConvert R Package Documentation to a readthedocs Website
Provides Access to Git Repositories
Interface to the libgit2 library, which is a pure C implementation of the Git core methods. Provides access to Git repositories to extract data and running some basic Git commands.
View DocumentationAdvice on R Package Building
Give advice about good practices when building R packages. Advice includes functions and syntax to avoid, package structure, code complexity, code formatting, etc.
View DocumentationHelpers for Automatic Translation of Markdown-based Content
Provide workflows and guidance for automatic translation of Markdown-based R content using DeepL API.
View DocumentationHelps Download Archives of GitHub Repositories
Provide functionality to download archives (backups) for all repositories in a GitHub organization (useful for backups!).
View DocumentationEstimate, report and combine felling dates of historical tree-ring series
fellingdater is an R package that aims to facilitate the analysis and interpretation of tree-ring data from wooden cultural heritage objects and structures. The package standardizes the process of computing and combining felling date estimates, both for individual and groups of related tree-ring series.
View DocumentationCreate Networks of Dendrochronological Series using Pairwise Similarity
Creating dendrochronological networks based on the similarity between tree-ring series or chronologies. The package includes various functions to compare tree-ring curves building upon the dplR package. The networks can be used to visualise and understand the relations between tree-ring curves. These networks are also very useful to estimate the provenance of wood as described in Visser (2021) DOI:10.5334/jcaa.79 or wood-use within a structure/context/site as described in Visser and Vorst (2022) DOI:10.1163/27723194-bja10014.
View DocumentationMake Fake Data
Make fake data that looks realistic, supporting addresses, person names, dates, times, colors, coordinates, currencies, digital object identifiers (DOIs), jobs, phone numbers, DNA sequences, doubles and integers from distributions and within a range.
View DocumentationAssertive Programming for R Analysis Pipelines
Provides functionality to assert conditions that have to be met so that errors in data used in analysis pipelines can fail quickly. Similar to stopifnot() but more powerful, friendly, and easier for use in pipelines.
View DocumentationFast, Consistent Tokenization of Natural Language Text
Convert natural language text into tokens. Includes tokenizers for shingled n-grams, skip n-grams, words, word stems, sentences, paragraphs, characters, shingled characters, lines, Penn Treebank, regular expressions, as well as functions for counting characters, words, and sentences, and a function for splitting longer texts into separate documents, each with the same number of words. The tokenizers have a consistent interface, and the package is built on the stringi and Rcpp packages for fast yet correct tokenization in UTF-8.
View DocumentationValidate JSON Schema
Uses the node library is-my-json-valid or ajv to validate JSON against a JSON schema. Drafts 04, 06 and 07 of JSON schema are supported.
View DocumentationLabel Creation for Tracking and Collecting Data from Biological Samples
Tools to generate unique identifier codes and printable barcoded labels for the management of biological samples. The creation of unique ID codes and printable PDF files can be initiated by standard commands, user prompts, or through a GUI addin for R Studio. Biologically informative codes can be included for hierarchically structured sampling designs.
View DocumentationManage Cached Files
Suite of tools for managing cached files, targeting use in other R packages. Uses rappdirs for cross-platform paths. Provides utilities to manage cache directories, including targeting files by path or by key; cached directories can be compressed and uncompressed easily to save disk space.
View DocumentationWorking with Sets the Tidy Way
Implements a class and methods to work with sets, doing intersection, union, complementary sets, power sets, cartesian product and other set operations in a “tidy” way. These set operations are available for both classical sets and fuzzy sets. Import sets from several formats or from other several data structures.
View DocumentationEstimate Avian Body Size Distributions
Generate estimated body size distributions for populations or communities of birds, given either species ID or species’ mean body size. Designed to work naturally with the North American Breeding Bird Survey, or with any dataset of bird species, abundance, and/or mean size data.
View DocumentationTracer from R
BEAST2 (https://www.beast2.org) is a widely used Bayesian phylogenetic tool, that uses DNA/RNA/protein data and many model priors to create a posterior of jointly estimated phylogenies and parameters. Tracer (https://github.com/beast-dev/tracer/) is a GUI tool to parse and analyze the files generated by BEAST2. This package provides a way to parse and analyze BEAST2 input files without active user input, but using R function calls instead.
View DocumentationrOpenSci package review project template
Creates files and collects materials necessary to complete an rOpenSci package review. Review files are prepopulated with review package specific metadata. Review package source code is also cloned for local testing and inspection.
View DocumentationStraightforward BibTeX and BibLaTeX Bibliography Management
Provides tools for importing and working with bibliographic references. It greatly enhances the bibentry class by providing a class BibEntry which stores BibTeX and BibLaTeX references, supports UTF-8 encoding, and can be easily searched by any field, by date ranges, and by various formats for name lists (author by last names, translator by full names, etc.). Entries can be updated, combined, sorted, printed in a number of styles, and exported. BibTeX and BibLaTeX .bib files can be read into R and converted to BibEntry objects. Interfaces to NCBI Entrez, CrossRef, and Zotero are provided for importing references and references can be created from locally stored PDF files using Poppler. Includes functions for citing and generating a bibliography with hyperlinks for documents prepared with RMarkdown or RHTML.
View DocumentationAvoid the Typical Working Directory Pain When Using knitr
An extension of knitr that adds flexibility in several ways. One common source of frustration with knitr is that it assumes the directory where the source file lives should be the working directory, which is often not true. ezknitr addresses this problem by giving you complete control over where all the inputs and outputs are, and adds several other convenient features to make rendering markdown/HTML documents easier.
View DocumentationTransform Google Docs into Quarto Books
Automate the Transformation of a Google Document into a Quarto Book source.
View DocumentationCompact and Flexible Summaries of Data
A simple to use summary function that can be used with pipes and displays nicely in the console. The default summary statistics may be modified by the user as can the default formatting. Support for data frames and vectors is included, and users can implement their own skim methods for specific object types as described in a vignette. Default summaries include support for inline spark graphs. Instructions for managing these on specific operating systems are given in the “Using skimr” vignette and the README.
View DocumentationWork with GitHub Gists
Work with GitHub gists from R (e.g., https://en.wikipedia.org/wiki/GitHub#Gist, https://docs.github.com/en/github/writing-on-github/creating-gists/). A gist is simply one or more files with code/text/images/etc. This package allows the user to create new gists, update gists with new files, rename files, delete files, get and delete gists, star and un-star gists, fork gists, open a gist in your default browser, get embed code for a gist, list gist commits, and get rate limit information when authenticated. Some requests require authentication and some do not. Gists website: https://gist.github.com/.
View DocumentationrOpenSci's blog guidance
It provides templates for roweb2 blogging and help for a GitHub forking workflow.
View DocumentationA Flexible Container to Transport and Manipulate Data and Associated Resources
Provides a flexible container to transport and manipulate complex sets of data. These data may consist of multiple data files and associated meta data and ancillary files. Individual data objects have associated system level meta data, and data files are linked together using the OAI-ORE standard resource map which describes the relationships between the files. The OAI- ORE standard is described at https://www.openarchives.org/ore/. Data packages can be serialized and transported as structured files that have been created following the BagIt specification. The BagIt specification is described at https://tools.ietf.org/html/draft-kunze-bagit-08.
View DocumentationSnapshot Current Versions of CRAN-like Repositories
Read, Tidy, and Display Data from Microtiter Plates
Tools for interacting with data from experiments done in microtiter plates. Easily read in plate-shaped data and convert it to tidy format, combine plate-shaped data with tidy data, and view tidy data in plate shape.
View DocumentationConvert Antipsychotic Doses to Chlorpromazine Equivalents
As different antipsychotic medications have different potencies, the doses of different medications cannot be directly compared. Various strategies are used to convert doses into a common reference so that comparison is meaningful. Chlorpromazine (CPZ) has historically been used as a reference medication into which other antipsychotic doses can be converted, as “chlorpromazine-equivalent doses”. Using conversion keys generated from widely-cited scientific papers, e.g. Gardner et. al 2010 doi:10.1176/appi.ajp.2009.09060802 and Leucht et al. 2016 doi:10.1093/schbul/sbv167, antipsychotic doses are converted to CPZ (or any specified antipsychotic) equivalents. The use of the package is described in the included vignette. Not for clinical use.
View DocumentationAAPOR Survey Outcome Rates
Standardized survey outcome rate functions, including the response rate, contact rate, cooperation rate, and refusal rate. These outcome rates allow survey researchers to measure the quality of survey data using definitions published by the American Association of Public Opinion Research (AAPOR). For details on these standards, see AAPOR (2016) https://www.aapor.org/Standards-Ethics/Standard-Definitions-(1).aspx.
View Documentation