16

This is a follow up to a previous question about learning multiple models.

The use case is that I have multiple observations for each subject, and I want to train a model for each of them. See Hadley's excellent presentation on how to do this.

In short, this is possible to do using dplyr and purrr like so:

library(purrr)
library(dplyr)
library(fitdistrplus)
dt %>% 
    split(dt$subject_id) %>%
    map( ~ fitdist(.$observation, "norm")) 

So since the model building is an embarrassingly parallel task, I was wondering if dplyr, purrr have an easy to use parallelization mechanism for such tasks (like a parallel map).

If these libraries don't provide easy parallelization could it be done using the classic R parallelization libraries (parallel, foreach etc)?

Axeman
  • 32,068
  • 8
  • 81
  • 94
Bar
  • 2,736
  • 3
  • 33
  • 41

2 Answers2

13

Just adding an answer for completeness here, you will need to install multidplyr from Hadley's repo to run this, more info in the vignette:

library(dplyr)
library(multidplyr)
library(purrr)

cluster <- create_cluster(4)
set_default_cluster(cluster)
cluster_library(cluster, "fitdistrplus")

# dt is a dataframe, subject_id identifies observations from each subject
by_subject <- partition(dt, subject_id)

fits <- by_subject %>% 
    do(fit = fitdist(.$observation, "norm")))

collected_fits <- collect(fits)$fit
collected_summaries <- collected_fits %>% map(summary)
Bar
  • 2,736
  • 3
  • 33
  • 41
  • 3
    multidplyr hasn't been developed for 2 years (as of Aug 2018) so something like furrr might be better. – xiaodai Aug 15 '18 at 04:43
13

There is the furrr package now, for example something like:

library(dplyr)
library(furrr)
plan(multisession)   # or perhaps:  plan(multicore), see ?plan

dt %>% 
    split(dt$subject_id) %>%
    future_map(~fitdist(.$observation, "norm"))
Axeman
  • 32,068
  • 8
  • 81
  • 94