The docker container has the latest version of OpenML-Tensorflow downloaded and pre-installed. It can be used to run TensorFlow Deep Learning analysis on OpenML datasets. This document contains information about:
Usage: how to use the image
Using Locally Stored Datasets: mounting datasets from the local cache
Environment Variables: setting the cache directory path
These are the steps to use the image:
- Pull the docker image
docker pull openml/openml-tensorflow:latest
- If you want to run a local script, it needs to be mounted first. Mount it into the 'app' folder:
docker run -it -v PATH/TO/CODE_FOLDER:/app openml/openml-tensorflow /bin/bash
You can also mount multiple directories into the container (such as your code file directory and dataset directory ) using:
docker run -t -i -v PATH/TO/CODE_FOLDER:/app -v PATH/TO/DATASET_FOLDER:/app/dataset openml/openml-tensorflow /bin/bash
- Please make sure to give the correct path to the dataset. For example,
openml_tensorflow.config.dir = 'dataset/Images'
- Run your code scripts using, for example:
python docs/Examples/tf_image_classification.py
If you don't want to download the dataset each time you run your script, you can mount your dataset saved in your local cache directory to the container.
- Mount the dataset to the 'app/dataset' folder
docker run -t -i -v PATH/TO/CODE_FOLDER:/app -v PATH/TO/DATASET_FOLDER:/app/dataset openml/openml-tensorflow /bin/bash
- Set correct path to the dataset.
openml_tensorflow.config.dir = '/app/dataset/Images'
You can configure the cache directory to control where 'OpenML' datasets are downloaded and cached.
cache_dir = "/app/.openml"
openml.config.set_root_cache_directory(cache_dir)