Skip to content

Commit

Permalink
Merge pull request #26 from bmschmidt/dev
Browse files Browse the repository at this point in the history
Dev
  • Loading branch information
bmschmidt authored Mar 13, 2017
2 parents 4ca3aeb + b67b2d5 commit 8b858fa
Show file tree
Hide file tree
Showing 59 changed files with 2,832 additions and 485 deletions.
1 change: 1 addition & 0 deletions .Rbuildignore
Original file line number Diff line number Diff line change
@@ -1,2 +1,3 @@
^.*\.Rproj$
^\.Rproj\.user$
^CONDUCT\.md$
14 changes: 13 additions & 1 deletion .gitignore
Original file line number Diff line number Diff line change
Expand Up @@ -6,6 +6,18 @@ inst/extdata/
hum2vec.Rproj
src/*.o
src/*.so
cookbooks cookbooks.txt cookbooks.vectors cookbooks.zip
cookbooks
cookbooks.txt
cookbooks.vectors
cookbooks.zip
cookbooks*
etc
cookbook_vectors.bin
tests/testthat/binary.bin
tests/testthat/input.txt
tests/testthat/tmp.txt
tests/testthat/binary.bin
tests/testthat/tmp.bin
vignettes/*.R
vignettes/*.html
vignettes/*_files
5 changes: 3 additions & 2 deletions .travis.yml
Original file line number Diff line number Diff line change
@@ -1,7 +1,8 @@
language: r
cache: packages
warnings_are_errors: false

warnings_are_errors: true
r_build_args: --no-build-vignettes --no-manual --no-resave-data
r_check_args: --no-build-vignettes --no-manual
r:
- release
- devel
Expand Down
25 changes: 25 additions & 0 deletions CONDUCT.md
Original file line number Diff line number Diff line change
@@ -0,0 +1,25 @@
# Contributor Code of Conduct

As contributors and maintainers of this project, we pledge to respect all people who
contribute through reporting issues, posting feature requests, updating documentation,
submitting pull requests or patches, and other activities.

We are committed to making participation in this project a harassment-free experience for
everyone, regardless of level of experience, gender, gender identity and expression,
sexual orientation, disability, personal appearance, body size, race, ethnicity, age, or religion.

Examples of unacceptable behavior by participants include the use of sexual language or
imagery, derogatory comments or personal attacks, trolling, public or private harassment,
insults, or other unprofessional conduct.

Project maintainers have the right and responsibility to remove, edit, or reject comments,
commits, code, wiki edits, issues, and other contributions that are not aligned to this
Code of Conduct. Project maintainers who do not follow the Code of Conduct may be removed
from the project team.

Instances of abusive, harassing, or otherwise unacceptable behavior may be reported by
opening an issue or contacting one or more of the project maintainers.

This Code of Conduct is adapted from the Contributor Covenant
(http:contributor-covenant.org), version 1.0.0, available at
http://contributor-covenant.org/version/1/0/0/
35 changes: 25 additions & 10 deletions DESCRIPTION
Original file line number Diff line number Diff line change
@@ -1,23 +1,38 @@
Package: wordVectors
Type: Package
Title: Tools for creating and analyzing vector-space models of texts
Version: 1.3
Date: 2015-09-10
Version: 2.0
Author: Ben Schmidt, Jian Li
Maintainer: Ben Schmidt <bmschmidt@gmail.com>
Description: wordVectors wraps Google's word2vec code for creating vector-space
models of texts, and defines a new class "VectorSpaceModel" (extending the native matrix class)
with a number of functions that make it easier to perform useful operations in a
word-vector space.
Description:
wordVectors wraps Google's implementation in C for training word2vec models,
and provides several R functions for exploratory data analysis of word2vec
and other related models. These include import-export from the binary format,
some useful linear algebra operations missing from R, and a streamlined
syntax for working with models and performing vector arithmetic that make it
easier to perform useful operations in a word-vector space.
License: Apache License (== 2.0)
URL: http://github.com/bmschmidt/wordVectors
BugReports: https://github.com/bmschmidt/wordVectors/issues
Depends:
R (>= 2.14.0)
LazyData: TRUE
Imports:
magrittr,
graphics,
methods,
utils
utils,
stats,
readr,
stringr,
stringi
Suggests:
stringi,
tsne
RoxygenNote: 5.0.1
tsne,
testthat,
ggplot2,
knitr,
dplyr,
rmarkdown,
devtools
RoxygenNote: 6.0.1
VignetteBuilder: knitr
5 changes: 5 additions & 0 deletions NAMESPACE
Original file line number Diff line number Diff line change
@@ -1,9 +1,13 @@
# Generated by roxygen2: do not edit by hand

export("%>%")
export(as.VectorSpaceModel)
export(closest_to)
export(cosineDist)
export(cosineSimilarity)
export(distend)
export(filter_to_rownames)
export(improve_vectorspace)
export(magnitudes)
export(nearest_to)
export(normalize_lengths)
Expand All @@ -17,4 +21,5 @@ export(word2phrase)
export(write.binary.word2vec)
exportClasses(VectorSpaceModel)
exportMethods(plot)
importFrom(magrittr,"%>%")
useDynLib(wordVectors)
137 changes: 137 additions & 0 deletions NEWS.md
Original file line number Diff line number Diff line change
@@ -0,0 +1,137 @@
# VERSION 2.0

Upgrade focusing on ease of use and CRAN-ability. Bumping major version because of a breaking change in the behavior of `closest_to`, which now returns a data.frame.

# Changes

## New default function: closest_to.

`nearest_to` was previously the easiest way to interact with cosine similarity functions. That's been deprecated
in favor of a new function, `closest_to`. (I would have changed the name but for back-compatibility reasons.)
The data.frame columns have elaborate names so they can easily be manipulated with dplyr, and/or plotted with ggplot.
`nearest_to` is now just a wrapped version of the new function.

## New syntax for vector addition.

This package now allows formula scoping for the most common operations, and string inputs to access in the context of a particular matrix. This makes this much nicer for handling the bread and butter word2vec operations.

For instance, instead of writing
```R
vectors %>% closest_to(vectors[rownames(vectors)=="king",] - vectors[rownames(vectors)=="man",] + vectors[rownames(vectors)=="woman",])
```

(whew!), you can write

```R
vectors %>% closest_to(~"king" - "man" + "woman")
```


## Reading tweaks.

In keeping with the goal of allowing manipulation of models in low-memory environments, it's now possible to read only rows with words matching certain criteria by passing an argument to read.binary.vectors(); either `rowname_list` for a fixed list, or `rowname_regexp` for a regular expression. (You could, say, read only the gerunds from a file by entering `rowname_regexp = "*.ing"`).

## Test Suite

The package now includes a test suite.

## Other materials for rOpenScience and JOSS.

This package has enough users it might be nice to get it on CRAN. I'm trying doing so through rOpenSci. That requires a lot of small files scattered throughout.


# VERSION 1.3

Two significant performance improvements.
1. Row magnitudes for a `VectorSpaceModel` object are now **cached** in an environment that allows some pass-by-reference editing. This means that the most time-consuming part of any comparison query is only done once for any given vector set; subsequent queries are at least an order of magnitude (10-20x)? faster.

Although this is a big performance improvement, certain edge cases might not wipe the cache clear. **In particular, assignment inside a VSM object might cause incorrect calculations.** I can't see why anyone would be in the habit of manually tweaking a row or block (rather than a whole matrix).
1. Access to rows in a `VectorSpaceModel` object is now handled through callNextMethod() rather than accessing the element's .Data slot. For reasons opaque to me, hitting the .Data slot seems to internally require copying the whole huge matrix internally. Now that no longer happens.


# VERSION 1.2

This release implements a number of incremental improvements and clears up some errors.
- The package is now able to read and write in the binary word2vec format; since this is faster and takes much less hard drive space (down by about 66%) than writing out floats as text, it does so internally.
- Several improvements to the C codebase to avoid warnings by @mukul13, described [here](https://github.com/bmschmidt/wordVectors/pull/9). (Does this address the `long long` issue?)
- Subsetting with `[[` now takes an argument `average`; if false, rather than collapse a matrix down to a single row, it just extracts the elements that correspond to the words.
- Added sample data in the object `demo_vectors`: the 999 words from the most common vectors.
- Began adding examples to the codebase.
- Tracking build success using Travis.
- Dodging most warnings from R CMD check.

Bug fixes
- If the `dir.exists` function is undefined, it creates one for you. This should allow installation on R 3.1 and some lower versions.
- `reject` and `project` are better about returning VSM objects, rather than dropping back into a matrix.

# VERSION 1.1

A few changes, primarily to the functions for _training_ vector spaces to produce higher quality models. A number of these changes are merged back in from the fork of this repo by github user @sunecasp . Thanks!

## Some bug fixes

Filenames can now be up to 1024 characters. Some parameters on alpha decay may be fixed; I'm not entirely sure what sunecasp's changes do.

## Changes to default number of iterations.

Models now default to 5 iterations through the text rather than 1. That means training may take 5 times as long; but particularly for small corpora, the vectors should be of higher quality. See below for an example.

## More training arguments

You can now specify more flags to the word2vec code. `?train_word2vec` gives a full list, but particularly useful are:
1. `window` now accurately sets the window size.
2. `iter` sets the number of iterations. For very large corpora, `iter=1` will train most quickly; for very small corpora, `iter=15` will give substantially better vectors. (See below). You should set this as high as you can stand within reason (Setting `iter` to a number higher than `window` is probably not that useful). But more text is better than more iterations.
3. `min_count` gives a cutoff for vocabulary size. Tokens occurring fewer than `min_count` times will be dropped from the model. Setting this high can be useful. (But note that a trained model is sorted in order of frequency, so if you have the RAM to train a big model you can reduce it in size for analysis by just subsetting to the first 10,000 or whatever rows).

## Example of vectors

Here's an example of training on a small set (c. 1000 speeches on the floor of the house of commons from the early 19th century).

> proc.time({one = train_word2vec("~/tmp2.txt","~/1_iter.vectors",iter = 1)})
> Error in train_word2vec("~/tmp2.txt", "~/1_iter.vectors", iter = 1) :
> The output file '~/1_iter.vectors' already exists: delete or give a new destination.
> proc.time({one = train_word2vec("~/tmp2.txt","~/1_iter.vectors",iter = 1)})
> Starting training using file /Users/bschmidt/tmp2.txt
> Vocab size: 4469
> Words in train file: 407583
> Alpha: 0.000711 Progress: 99.86% Words/thread/sec: 67.51k
> Error in proc.time({ : 1 argument passed to 'proc.time' which requires 0
> ?proc.time
> system.time({one = train_word2vec("~/tmp2.txt","~/1_iter.vectors",iter = 1)})
> Starting training using file /Users/bschmidt/tmp2.txt
> Vocab size: 4469
> Words in train file: 407583
> Alpha: 0.000711 Progress: 99.86% Words/thread/sec: 66.93k user system elapsed
> 6.753 0.055 6.796
> system.time({two = train_word2vec("~/tmp2.txt","~/2_iter.vectors",iter = 3)})
> Starting training using file /Users/bschmidt/tmp2.txt
> Vocab size: 4469
> Words in train file: 407583
> Alpha: 0.000237 Progress: 99.95% Words/thread/sec: 67.15k user system elapsed
> 18.846 0.085 18.896
>
> two %>% nearest_to(two["debt"]) %>% round(3)
> debt remainder Jan including drawback manufactures prisoners mercantile subsisting
> 0.000 0.234 0.256 0.281 0.291 0.293 0.297 0.314 0.314
> Dec
> 0.318
> one %>% nearest_to(one[["debt"]]) %>% round(3)
> debt Christmas exception preventing Indies import remainder eye eighteen labouring
> 0.000 0.150 0.210 0.214 0.215 0.220 0.221 0.223 0.225 0.227
>
> system.time({ten = train_word2vec("~/tmp2.txt","~/10_iter.vectors",iter = 10)})
> Starting training using file /Users/bschmidt/tmp2.txt
> Vocab size: 4469
> Words in train file: 407583
> Alpha: 0.000071 Progress: 99.98% Words/thread/sec: 66.13k user system elapsed
> 62.070 0.296 62.333
>
> ten %>% nearest_to(ten[["debt"]]) %>% round(3)
> debt surplus Dec remainder manufacturing grants Jan drawback prisoners
> 0.000 0.497 0.504 0.510 0.519 0.520 0.533 0.536 0.546
> compelling
> 0.553
```
```

2 changes: 1 addition & 1 deletion R/data.R
Original file line number Diff line number Diff line change
Expand Up @@ -3,7 +3,7 @@
#' A sample VectorSpaceModel object trained on about 15 million
#' teaching evaluations, limited to the 999 most common words.
#' Included for demonstration purposes only: there's only so much you can
#' do with a 999 dimension vocabulary.
#' do with a 999 length vocabulary.
#'
#' You're best off downloading a real model to work with,
#' such as the precompiled vectors distributed by Google
Expand Down
5 changes: 0 additions & 5 deletions R/functions.R

This file was deleted.

Loading

0 comments on commit 8b858fa

Please sign in to comment.