Core ML community tools contains all supporting tools for Core ML model conversion, editing and validation. This includes deep learning frameworks like TensorFlow, Keras, Caffe as well as classical machine learning frameworks like LIBSVB, scikit-learn, and XGBoost.
To get the latest version of coremltools:
pip install --upgrade coremltools
For the latest changes please see the release notes.
- Neural network conversion
- Core ML specification
- coremltools user guide and examples
- Installation from Source
Link to the detailed NN conversion guide.
There are several converters
available to translate neural networks trained in various frameworks into the Core ML model format.
Following formats can be converted to the Core ML .mlmodel
format through the coremltools python package (this repo):
- Caffe V1 (
.prototxt
,.caffemodel
format) - Keras API (2.2+) (
.h5
format) - TensorFlow 1 (1.13+) (
.pb
frozen graph def format) - TensorFlow 2 (
.h5
andSavedModel
formats)
In addition, there are two more neural network converters build on top of coremltools
:
- onnx-coreml: to convert
.onnx
model format. Several frameworks such as PyTorch, MXNet, CaffeV2 etc provide native export to the ONNX format. - tfcoreml: to convert TensorFlow models. For producing Core ML models targeting iOS 13 or later, tfcoreml defers to the TensorFlow converter implemented inside coremltools. For iOS 12 or earlier, the code path is different and lives entirely in the tfcoreml package.
To get an overview on how to use the converters and features such as post-training quantization using coremltools, please see the neural network guide.
- Core ML specification is fully described in a set of protobuf files.
They are all located in the folder
mlmodel/format/
- For an overview of the Core ML framework API, see here.
- To find the list of model types supported by Core ML, see this
portion of the
model.proto
file. - To find the list of neural network layer types supported see this
portion of the
NeuralNetwork.proto
file. - Auto-generated documentation for all the protobuf files can be found at this link
- API documentation
- Updatable models
- Neural network inference examples
- Neural network guide
- Miscellaneous How-to code snippets
We recommend using virtualenv to use, install, or build coremltools. Be sure to install virtualenv using your system pip.
pip install virtualenv
The method for installing coremltools follows the
standard python package installation steps.
To create a Python virtual environment called pythonenv
follow these steps:
# Create a folder for your virtualenv
mkdir mlvirtualenv
cd mlvirtualenv
# Create a Python virtual environment for your Core ML project
virtualenv pythonenv
To activate your new virtual environment and install coremltools
in this environment, follow these steps:
# Active your virtual environment
source pythonenv/bin/activate
# Install coremltools in the new virtual environment, pythonenv
(pythonenv) pip install -U coremltools
The package documentation contains more details on how to use coremltools.
coremltools has the following dependencies:
- numpy (1.10.0+)
- protobuf (3.1.0+)
In addition, it has the following soft dependencies that are only needed when you are converting models of these formats:
- Keras (1.2.2, 2.0.4+) with corresponding TensorFlow version
- XGBoost (0.7+)
- scikit-learn (0.17+)
- LIBSVM
To build the project, you need CMake to configure the project
mkdir build
cd build
cmake ../
When several python virtual environments are installed, it may be useful to use the following command instead, to point to the correct intended version of python:
cmake \
-DPYTHON_EXECUTABLE:FILEPATH=/Library/Frameworks/Python.framework/Versions/3.7/bin/python \
-DPYTHON_INCLUDE_DIR=/Library/Frameworks/Python.framework/Versions/3.7/include/python3.7m/ \
-DPYTHON_LIBRARY=/Library/Frameworks/Python.framework/Versions/3.7/lib/ \
../
after which you can use make to build the project
make
To make a wheel/egg that you can distribute, you can do the following
make dist
In order to run unit tests, you need pytest
, pandas
, and h5py
.
pip install pytest pandas h5py
To add a new unit test, add it to the coremltools/test
folder. Make sure you
name the file with a 'test' as the prefix.
Additionally, running unit-tests would require more packages (like LIBSVM)
pip install -r test_requirements.pip
To install LIBSVM
git clone https://github.com/cjlin1/libsvm.git
cd libsvm/
make
cd python/
make
To make sure you can run LIBSVM python bindings everywhere, you need the
following command, replacing <LIBSVM_PATH>
with the path to the root of
your repository.
export PYTHONPATH=${PYTHONPATH}:<LIBSVM_PATH>/python
To install XGBoost
git clone --recursive https://github.com/dmlc/xgboost
cd xgboost
git checkout v0.90
git submodule update
make config=make/config.mk -j8
cd python-package; python setup.py develop
To install Keras (Version >= 2.0)
pip install keras tensorflow
If you'd like to use the old Keras version, you can:
pip install keras==1.2.2 tensorflow
Finally, to run the most important unit tests, you can use:
pytest -rs
some tests are marked as slow because they test a lot of combinations. If you want to run, all tests, you can use:
pytest
First install all external dependencies.
pip install Sphinx==1.8.5 sphinx-rtd-theme==0.4.3 numpydoc==0.9.1
pip install -e git+git://github.com/michaeljones/sphinx-to-github.git#egg=sphinx-to-github
You also must have the coremltools package install, see the Building section.
Then from the root of the repository:
cd docs
make html
open _build/html/index.html