forked from hadley/adv-r
-
Notifications
You must be signed in to change notification settings - Fork 0
/
memory.Rmd
518 lines (372 loc) · 23 KB
/
memory.Rmd
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
# Memory {#memory}
```{r setup, include = FALSE}
source("common.R")
library(pryr)
knitr::read_chunk("memory-read-delim.r")
options(scipen = 9)
```
A solid understanding of R's memory management will help you predict how much memory you'll need for a given task and help you to make the most of the memory you have. It can even help you write faster code because accidental copies are a major cause of slow code. The goal of this chapter is to help you understand the basics of memory management in R, moving from individual objects to functions to larger blocks of code. Along the way, you'll learn about some common myths, such as that you need to call `gc()` to free up memory, or that `for` loops are always slow. \index{memory}
##### Outline
* [Object size](#object-size) shows you how to use `object_size()` to see
how much memory an object occupies, and uses that as a launching point to
improve your understanding of how R objects are stored in memory.
* [Memory usage and garbage collection](#gc) introduces you to the `mem_used()`
and `mem_change()` functions that will help you understand how R allocates
and frees memory.
* [Memory profiling with lineprof](#memory-profiling) shows you how to use
the lineprof package to understand how memory is allocated and released in
larger code blocks.
* [Modification in place](#modification) introduces you to the `address()` and
`refs()` functions so that you can understand when R modifies in place and
when R modifies a copy. Understanding when objects are copied is very
important for writing efficient R code.
##### Prerequisites
In this chapter, we'll use tools from the pryr and lineprof packages to understand memory usage, and a sample dataset from ggplot2. If you don't already have them, run this code to get the packages you need:
```{r, eval = FALSE}
install.packages("ggplot2")
install.packages("pryr")
install.packages("devtools")
devtools::install_github("hadley/lineprof")
```
##### Sources
<!-- FIXME: cite R-exts and R-ints formally -->
The details of R's memory management are not documented in a single place. Most of the information in this chapter was gleaned from a close reading of the documentation (particularly `?Memory` and `?gc`), the [memory profiling](http://cran.r-project.org/doc/manuals/R-exts.html#Profiling-R-code-for-memory-use) section of R-exts, and the [SEXPs](http://cran.r-project.org/doc/manuals/R-ints.html#SEXPs) section of R-ints. The rest I figured out by reading the C source code, performing small experiments, and asking questions on R-devel. Any mistakes are entirely mine.
## Object size {#object-size}
<!-- deletion size vs. serialization size -->
To understand memory usage in R, we will start with `pryr::object_size()`. This function tells you how many bytes of memory an object occupies: \index{object\_size()}
```{r}
library(pryr)
object_size(1:10)
object_size(mean)
object_size(mtcars)
```
(This function is better than the built-in `object.size()` because it accounts for shared elements within an object and includes the size of environments.)
Something interesting occurs if we use `object_size()` to systematically explore the size of an integer vector. The code below computes and plots the memory usage of integer vectors ranging in length from 0 to 50 elements. You might expect that the size of an empty vector would be zero and that memory usage would grow proportionately with length. Neither of those things are true! \index{vectors!size of}
```{r size-q, fig.height = 2.5, fig.width = 4, small_mar = TRUE}
sizes <- sapply(0:50, function(n) object_size(seq_len(n)))
plot(0:50, sizes, xlab = "Length", ylab = "Size (bytes)",
type = "s")
```
This isn't just an artefact of integer vectors. Every length 0 vector occupies 40 bytes of memory:
```{r}
object_size(numeric())
object_size(logical())
object_size(raw())
object_size(list())
```
Those 40 bytes are used to store four components possessed by every object in R:
* Object metadata (4 bytes). These metadata store the base type (e.g. integer)
and information used for debugging and memory management.
* Two pointers: one to the next object in memory and one to the previous
object (2 * 8 bytes). This doubly-linked list makes it easy for internal
R code to loop through every object in memory.
* A pointer to the attributes (8 bytes).
All vectors have three additional components: \indexc{SEXP}
* The length of the vector (4 bytes). By using only 4 bytes, you might expect
that R could only support vectors up to $2 ^ {4 \times 8 - 1}$ ($2 ^ {31}$, about
two billion) elements. But in R 3.0.0 and later, you can actually have
vectors up to $2 ^ {52}$ elements. [Read R-internals][long-vectors] to see how
support for long vectors was added without having to change the size of this
field. \index{long vectors} \index{atomic vectors!long}
* The "true" length of the vector (4 bytes). This is basically never used,
except when the object is the hash table used for an environment. In that
case, the true length represents the allocated space, and the length
represents the space currently used.
* The data (?? bytes). An empty vector has 0 bytes of data. Numeric vectors occupy 8 bytes for
every element, integer vectors 4, and complex vectors 16.
If you're keeping count you'll notice that this only adds up to 36 bytes. The remaining 4 bytes are used for padding so that each component starts on an 8 byte (= 64-bit) boundary. Most cpu architectures require pointers to be aligned in this way, and even if they don't require it, accessing non-aligned pointers tends to be rather slow. (If you're interested, you can read more about it in [C structure packing](http://www.catb.org/esr/structure-packing/).)
This explains the intercept on the graph. But why does the memory size grow irregularly? To understand why, you need to know a little bit about how R requests memory from the operating system. Requesting memory (with `malloc()`) is a relatively expensive operation. Having to request memory every time a small vector is created would slow R down considerably. Instead, R asks for a big block of memory and then manages that block itself. This block is called the small vector pool and is used for vectors less than 128 bytes long. For efficiency and simplicity, it only allocates vectors that are 8, 16, 32, 48, 64, or 128 bytes long. If we adjust our previous plot to remove the 40 bytes of overhead, we can see that those values correspond to the jumps in memory use.
```{r size-a, fig.height = 2.5, fig.width = 4, small_mar = TRUE}
plot(0:50, sizes - 40, xlab = "Length",
ylab = "Bytes excluding overhead", type = "n")
abline(h = 0, col = "grey80")
abline(h = c(8, 16, 32, 48, 64, 128), col = "grey80")
abline(a = 0, b = 4, col = "grey90", lwd = 4)
lines(sizes - 40, type = "s")
```
Beyond 128 bytes, it no longer makes sense for R to manage vectors. After all, allocating big chunks of memory is something that operating systems are very good at. Beyond 128 bytes, R will ask for memory in multiples of 8 bytes. This ensures good alignment.
A subtlety of the size of an object is that components can be shared across multiple objects. For example, look at the following code:
```{r}
x <- 1:1e6
object_size(x)
y <- list(x, x, x)
object_size(y)
```
`y` isn't three times as big as `x` because R is smart enough to not copy `x` three times; instead it just points to the existing `x`.
It's misleading to look at the sizes of `x` and `y` individually. If you want to know how much space they take up together, you have to supply them to the same `object_size()` call:
```{r}
object_size(x, y)
```
In this case, `x` and `y` together take up the same amount of space as `y` alone. This is not always the case. If there are no shared components, as in the following example, then you can add up the sizes of individual components to find out the total size:
```{r}
x1 <- 1:1e6
y1 <- list(1:1e6, 1:1e6, 1:1e6)
object_size(x1)
object_size(y1)
object_size(x1, y1)
object_size(x1) + object_size(y1) == object_size(x1, y1)
```
The same issue also comes up with strings, because R has a global string pool. This means that each unique string is only stored in one place, and therefore character vectors take up less memory than you might expect: \index{string pool}
```{r}
object_size("banana")
object_size(rep("banana", 10))
```
### Exercises
1. Repeat the analysis above for numeric, logical, and complex vectors.
1. If a data frame has one million rows, and three variables (two numeric, and
one integer), how much space will it take up? Work it out from theory,
then verify your work by creating a data frame and measuring its size.
1. Compare the sizes of the elements in the following two lists. Each
contains basically the same data, but one contains vectors of small
strings while the other contains a single long string.
```{r}
vec <- lapply(0:50, function(i) c("ba", rep("na", i)))
str <- lapply(vec, paste0, collapse = "")
```
1. Which takes up more memory: a factor (`x`) or the equivalent character
vector (`as.character(x)`)? Why?
1. Explain the difference in size between `1:5` and `list(1:5)`.
## Memory usage and garbage collection {#gc}
While `object_size()` tells you the size of a single object, `pryr::mem_used()` tells you the total size of all objects in memory: \indexc{mem\_used()}
```{r}
library(pryr)
mem_used()
```
This number won't agree with the amount of memory reported by your operating system for a number of reasons:
1. It only includes objects created by R, not the R interpreter itself.
1. Both R and the operating system are lazy: they won't reclaim memory
until it's actually needed. R might be holding on to memory because
the OS hasn't yet asked for it back.
1. R counts the memory occupied by objects but there may be gaps due to
deleted objects. This problem is known as memory fragmentation.
`mem_change()` builds on top of `mem_used()` to tell you how memory changes during code execution. Positive numbers represent an increase in the memory used by R, and negative numbers represent a decrease. \indexc{mem\_change()}
```{r}
# Need about 4 mb to store 1 million integers
mem_change(x <- 1:1e6)
# We get that memory back when we delete it
mem_change(rm(x))
```
Even operations that don't do anything use up a little memory. This is because R is tracking the history of everything you do. You can ignore anything smaller than a couple kB.
```{r}
mem_change(NULL)
mem_change(NULL)
```
In some languages, you have to explicitly delete unused objects for their memory to be returned. R uses an alternative approach: garbage collection (or GC for short). GC automatically releases memory when an object is no longer used. It does this by tracking how many names point to each object, and when there are no names pointing to an object, it deletes that object. \index{garbage collection}
```{r, echo = FALSE}
rm(y)
```
```{r}
# Create a big object
mem_change(x <- 1:1e6)
# Also point to 1:1e6 from y
mem_change(y <- x)
# Remove x, no memory freed because y is still pointing to it
mem_change(rm(x))
# Now nothing points to it and the memory can be freed
mem_change(rm(y))
```
Despite what you might have read elsewhere, there's never any need to call `gc()` yourself. R will automatically run garbage collection whenever it needs more space; if you want to see when that is, call `gcinfo(TRUE)`. The only reason you _might_ want to call `gc()` is to ask R to return memory to the operating system. However, even that might not have any effect: older versions of Windows had no way for a program to return memory to the OS. \indexc{gc()}
GC takes care of releasing objects that are no longer used. However, you do need to be aware of possible memory leaks. A memory leak occurs when you keep pointing to an object without realising it. In R, the two main causes of memory leaks are formulas and closures because they both capture the enclosing environment. The following code illustrates the problem. In `f1()`, `1:1e6` is only referenced inside the function, so when the function completes the memory is returned and the net memory change is 0. `f2()` and `f3()` both return objects that capture environments, so that `x` is not freed when the function completes. \index{memory!leaks}
```{r}
f1 <- function() {
x <- 1:1e6
10
}
mem_change(x <- f1())
object_size(x)
f2 <- function() {
x <- 1:1e6
a ~ b
}
mem_change(y <- f2())
object_size(y)
f3 <- function() {
x <- 1:1e6
function() 10
}
mem_change(z <- f3())
object_size(z)
```
```{r, echo = FALSE}
rm(y, z)
```
## Memory profiling with lineprof {#memory-profiling}
`mem_change()` captures the net change in memory when running a block of code. Sometimes, however, we may want to measure incremental change. One way to do this is to use memory profiling to capture usage every few milliseconds. This functionality is provided by `utils::Rprof()` but it doesn't provide a very useful display of the results. Instead we'll use the [lineprof](https://github.com/hadley/lineprof) package. It is powered by `Rprof()`, but displays the results in a more informative manner. \index{memory!profiling}
To demonstrate `lineprof`, we're going to explore a bare-bones implementation of `read.delim()` with only three arguments: \indexc{read\_delim()}
```{r read_delim}
```
We'll also create a sample csv file:
```{r}
library(ggplot2)
write.csv(diamonds, "diamonds.csv", row.names = FALSE)
```
Using lineprof is straightforward. `source()` the code, apply `lineprof()` to an expression, then use `shine()` to view the results. Note that you _must_ use `source()` to load the code. This is because lineprof uses srcrefs to match up the code and run times. The needed srcrefs are only created when you load code from disk.
```{r, eval = FALSE}
library(lineprof)
source("code/read-delim.R")
prof <- lineprof(read_delim("diamonds.csv"))
shine(prof)
```
```{r, echo = FALSE}
knitr::include_graphics("screenshots/memory-lineprof.png", dpi = 220)
```
`shine()` will also open a new web page (or if you're using RStudio, a new pane) that shows your source code annotated with information about memory usage. `shine()` starts a shiny app which will "block" your R session. To exit, press escape or ctrl + break.
Next to the source code, four columns provide details about the performance of the code:
* `t`, the time (in seconds) spent on that line of code (explained in
[measuring performance](#measure-perf)).
* `a`, the memory (in megabytes) allocated by that line of code.
* `r`, the memory (in megabytes) released by that line of code. While memory
allocation is deterministic, memory release is stochastic: it depends on when
the GC was run. This means that memory release only tells you that the memory
released was no longer needed before this line.
* `d`, the number of vector duplications that occurred. A vector duplication
occurs when R copies a vector as a result of its copy on modify semantics.
You can hover over any of the bars to get the exact numbers. In this example, looking at the allocations tells us most of the story:
* `scan()` allocates about 2.5 MB of memory, which is very close to the 2.8 MB
of space that the file occupies on disk. You wouldn't expect the two numbers
to be identical because R doesn't need to store the commas and because the
global string pool will save some memory.
* Converting the columns allocates another 0.6 MB of memory. You'd also expect
this step to free some memory because we've converted string columns into
integer and numeric columns (which occupy less space), but we can't see those
releases because GC hasn't been triggered yet.
* Finally, calling `as.data.frame()` on a list allocates about 1.6 megabytes
of memory and performs over 600 duplications. This is because
`as.data.frame()` isn't terribly efficient and ends up copying the input
multiple times. We'll discuss duplication more in the next section.
There are two downsides to profiling:
1. `read_delim()` only takes around half a second, but profiling can, at best,
capture memory usage every 1 ms. This means we'll only get about 500 samples.
1. Since GC is lazy, we can never tell exactly when memory is no longer needed.
You can work around both problems by using `torture = TRUE`, which forces R to run GC after every allocation (see `gctorture()` for more details). This helps with both problems because memory is freed as soon as possible, and R runs 10--100x slower. This effectively makes the resolution of the timer greater, so that you can see smaller allocations and exactly when memory is no longer needed.
### Exercises
1. When the input is a list, we can make a more efficient `as.data.frame()`
by using special knowledge. A data frame is a list with class `data.frame`
and `row.names` attribute. `row.names` is either a character vector or
vector of sequential integers, stored in a special format created by
`.set_row_names()`. This leads to an alternative `as.data.frame()`:
```{r}
to_df <- function(x) {
class(x) <- "data.frame"
attr(x, "row.names") <- .set_row_names(length(x[[1]]))
x
}
```
What impact does this function have on `read_delim()`? What are the
downsides of this function?
1. Line profile the following function with `torture = TRUE`. What is
surprising? Read the source code of `rm()` to figure out what's going on.
```{r}
f <- function(n = 1e5) {
x <- rep(1, n)
rm(x)
}
```
## Modification in place {#modification}
What happens to `x` in the following code? \index{copy-on-modify!exceptions} \index{avoiding copies}
```{r}
x <- 1:10
x[5] <- 10L
x
```
There are two possibilities:
1. R modifies `x` in place.
2. R makes a copy of `x` to a new location, modifies the copy, and then uses
the name `x` to point to the new location.
It turns out that R can do either depending on the circumstances. In the example above, it will modify in place. But if another variable also points to `x`, then R will copy it to a new location. To explore what's going on in greater detail, we use two tools from the pryr package. Given the name of a variable, `address()` will tell us the variable's location in memory and `refs()` will tell us how many names point to that location. \indexc{address()} \indexc{refs()}
```{r, eval = FALSE}
library(pryr)
x <- 1:10
c(address(x), refs(x))
# [1] "0x103100060" "1"
y <- x
c(address(y), refs(y))
# [1] "0x103100060" "2"
```
(Note that if you're using RStudio, `refs()` will always return 2: the environment browser makes a reference to every object you create on the command line.)
`refs()` is only an estimate. It can only distinguish between one and more than one reference (future versions of R might do better). This means that `refs()` returns 2 in both of the following cases: \index{reference counting}
```{r}
x <- 1:5
y <- x
rm(y)
# Should really be 1, because we've deleted y
refs(x)
x <- 1:5
y <- x
z <- x
# Should really be 3
refs(x)
```
When `refs(x)` is 1, modification will occur in place. When `refs(x)` is 2, R will make a copy (this ensures that other pointers to the object remain unaffected). Note that in the following example, `y` keeps pointing to the same location while `x` changes.
```{r}
x <- 1:10
y <- x
c(address(x), address(y))
x[5] <- 6L
c(address(x), address(y))
```
Another useful function is `tracemem()`. It prints a message every time the traced object is copied: \indexc{tracemem()}
```{r tracemem, eval = FALSE}
x <- 1:10
# Prints the current memory location of the object
tracemem(x)
# [1] "<0x7feeaaa1c6b8>"
x[5] <- 6L
y <- x
# Prints where it has moved from and to
x[5] <- 6L
# tracemem[0x7feeaaa1c6b8 -> 0x7feeaaa1c768]:
```
For interactive use, `tracemem()` is slightly more useful than `refs()`, but because it just prints a message, it's harder to program with. I don't use it in this book because it interacts poorly with [knitr](http://yihui.name/knitr/), the tool I use to interleave text and code.
Non-primitive functions that touch the object always increment the ref count. Primitive functions usually don't. (The reasons are a little complicated, but see the R-devel thread [confused about NAMED](http://r.789695.n4.nabble.com/Confused-about-NAMED-td4103326.html).) \index{primitive functions}
```{r}
# Touching the object forces an increment
f <- function(x) x
{x <- 1:10; f(x); refs(x)}
# Sum is primitive, so no increment
{x <- 1:10; sum(x); refs(x)}
# f() and g() never evaluate x, so refs don't increment
f <- function(x) 10
g <- function(x) substitute(x)
{x <- 1:10; f(x); refs(x)}
{x <- 1:10; g(x); refs(x)}
```
Generally, provided that the object is not referred to elsewhere, any primitive replacement function will modify in place. This includes `[[<-`, `[<-`, `@<-`, `$<-`, `attr<-`, `attributes<-`, `class<-`, `dim<-`, `dimnames<-`, `names<-`, and `levels<-`. To be precise, all non-primitive functions increment refs, but a primitive function may be written in such a way that it doesn't. The rules are sufficiently complicated that there's little point in trying to memorise them. Instead, you should approach the problem practically by using `refs()` and `address()` to figure out when objects are being copied. \index{subsetting|subassignment}
While determining that copies are being made is not hard, preventing such behaviour is. If you find yourself resorting to exotic tricks to avoid copies, it may be time to rewrite your function in C++, as described in [Rcpp](#rcpp).
### Loops
For loops in R have a reputation for being slow. Often that slowness is because you're modifying a copy instead of modifying in place. Consider the following code. It subtracts the median from each column of a large data frame: \index{loops!avoiding copies}
```{r, cache = TRUE}
x <- data.frame(matrix(runif(100 * 1e4), ncol = 100))
medians <- vapply(x, median, numeric(1))
for(i in seq_along(medians)) {
x[, i] <- x[, i] - medians[i]
}
```
You may be surprised to realise that every iteration of the loop copies the data frame. We can see that more clearly by using `address()` and `refs()` for a small sample of the loop:
```{r}
for(i in 1:5) {
x[, i] <- x[, i] - medians[i]
print(c(address(x), refs(x)))
}
```
For each iteration, `x` is moved to a new location so `refs(x)` is always 2. This occurs because `[<-.data.frame` is not a primitive function, so it always increments the refs. We can make the function substantially more efficient by using a list instead of a data frame. Modifying a list uses primitive functions, so the refs are not incremented and all modifications occur in place:
```{r}
y <- as.list(x)
for(i in 1:5) {
y[[i]] <- y[[i]] - medians[i]
print(c(address(y), refs(y)))
}
```
This behaviour was substantially more problematic prior to R 3.1.0, because every copy of the data frame was a deep copy. This made the motivating example take around 5 s, compared to 0.01 s today.
### Exercises
1. The code below makes one duplication. Where does it occur and why?
(Hint: look at `refs(y)`.)
```{r}
y <- as.list(x)
for(i in seq_along(medians)) {
y[[i]] <- y[[i]] - medians[i]
}
```
1. The implementation of `as.data.frame()` in the previous section has one
big downside. What is it and how could you avoid it?
[long-vectors]: http://cran.r-project.org/doc/manuals/R-ints.html#Long-vectors