These are instructions on how to run the experiments in the article "CoLLIE: Continual Learning of Language Grounding from Language-Image Embeddings" in the Journal of Artificial Intelligence Research (JAIR) 2022.
The code is written in Python and should be possible to run on both cpu and gpu.
We suggest you use Conda package manager and install everything within a Conda environment. On a CUDA GPU machine, the following will install everything you should need:
$ conda install --yes -c pytorch pytorch=1.7.1 torchvision cudatoolkit=11.0
$ pip install ftfy regex tqdm sklearn matplotlib pandas seaborn
$ pip install git+https://github.com/openai/CLIP.git
Replace cudatoolkit=11.0
above with the appropriate CUDA version on your machine or cpuonly
when installing on a machine without a GPU.
- Go to https://paperswithcode.com/dataset/lad and follow the links to download the images from the LAD dataset.
- Unzip the file to a directory of your choice.
- Open
runLAD.py
and setdataDir
to that directory.
- Run
runLAD.py
. If you want, you can open the file and reduce the number ofiterations
(default is 50). This will save you time, but result in a less smooth performance curve with larger CI. This will generate a file calledresult-lad.tsv
. - Run
plotLAD.py
to plot the results. - Run
resultsLAD.py
to print a summary of the results in table format. a summary of the results in table format.
- Run
runTangrams.py
. If you want, you can open the file and reduce the number ofiterations
(default is 3000). This will save you time, but result in a less smooth performance curve with larger CI. This will generate a file calledresult-tangrams.tsv
. - Run
plotTangrams.py
to plot the results. - Run
resultsTangrams.py
to print a summary of the results in table format. a summary of the results in table format.