DeepMind Lab is a 3D learning environment based on id Software's Quake III Arena via ioquake3 and other open source software.
DeepMind Lab provides a suite of challenging 3D navigation and puzzle-solving tasks for learning agents. Its primary purpose is to act as a testbed for research in artificial intelligence, especially deep reinforcement learning.
About
Disclaimer: This is not an official Google product.
If you use DeepMind Lab in your research and would like to cite the DeepMind Lab environment, we suggest you cite the DeepMind Lab paper.
You can reach us at [email protected].
Getting started on Linux
-
Get Bazel from bazel.io.
-
Clone DeepMind Lab, e.g. by running
$ git clone https://github.com/deepmind/lab
$ cd lab
For a live example of a random agent, run
lab$ bazel run :python_random_agent --define graphics=sdl -- \
--length=10000 --width=640 --height=480
Here is some more detailed build documentation, including how to install dependencies if you don't have them.
To enable compiler optimizations, pass the flag --compilation_mode=opt
, or
-c opt
for short, to each bazel build
, bazel test
and bazel run
command.
The flag is omitted from the examples here for brevity, but it should be used
for real training and evaluation where performance matters.
Play as a human
To test the game using human input controls, run
lab$ bazel run :game -- --level_script=tests/empty_room_test --level_setting=logToStdErr=true
# or:
lab$ bazel run :game -- -l tests/empty_room_test -s logToStdErr=true
Leave the logToStdErr
setting off to disable most log output.
The values of observations that the environment exposes can be printed at every
step by adding a flag --observation OBSERVATION_NAME
for each observation of
interest.
lab$ bazel run :game -- --level_script=lt_chasm --observation VEL.TRANS --observation VEL.ROT
Train an agent
DeepMind Lab ships with an example random agent in
python/random_agent.py
which can be used as a starting point for implementing a learning agent. To let
this agent interact with DeepMind Lab for training, run
lab$ bazel run :python_random_agent
The Python API is used for agent-environment interactions. We also provide bindings to DeepMind's "dm_env" general API for reinforcement learning, as well as a way to build a self-contained PIP package; see the separate documentation for details.
DeepMind Lab ships with different levels implementing different tasks. These tasks can be configured using Lua scripts, as described in the Lua API.
Upstream sources
DeepMind Lab is built from the ioquake3 game engine, and it uses the tools q3map2 and bspc for map creation. Bug fixes and cleanups that originate with those projects are best fixed upstream and then merged into DeepMind Lab.
-
bspc is taken from github.com/TTimo/bspc, revision d9a372db3fb6163bc49ead41c76c801a3d14cf80. There are virtually no local modifications, although we integrate this code with the main ioq3 code and do not use their copy in the
deps
directory. We expect this code to be stable. -
q3map2 is taken from github.com/TTimo/GtkRadiant, revision d3d00345c542c8d7cc74e2e8a577bdf76f79c701. A few minor local modifications add synchronization. We also expect this code to be stable.
-
ioquake3 is taken from github.com/ioquake/ioq3, revision 29db64070aa0bae49953bddbedbed5e317af48ba. The code contains extensive modifications and additions. We aim to merge upstream changes occasionally.
We are very grateful to the maintainers of these repositories for all their hard work on maintaining high-quality code bases.
External dependencies, prerequisites and porting notes
DeepMind Lab currently ships as source code only. It depends on a few external software libraries, which we ship in several different ways:
-
The
zlib
,glib
,libxml2
,jpeg
andpng
libraries are referenced as external Bazel sources, and Bazel BUILD files are provided. The dependent code itself should be fairly portable, but the BUILD rules we ship are specific to Linux on x86. To build on a different platform you will most likely have to edit those BUILD files. -
Message digest algorithms are included in this package (in
//third_party/md
), taken from the reference implementations of their respective RFCs. A "generic reinforcement learning API" is included in//third_party/rl_api
, which has also been created by the DeepMind Lab authors. This code is portable. -
EGL headers are included in this package (in
//third_party/GL/{
EGL
,
KHR
}
), taken from the Khronos OpenGL/OpenGL ES XML API Registry at www.khronos.org/registry/EGL. The headers have been modified slightly to remove the dependency of EGL on X. -
Several additional libraries are required but are not shipped in any form; they must be present on your system:
- SDL 2
- gettext (required by
glib
) - OpenGL: A hardware driver and library are needed for hardware-accelerated
human play. The headless library that machine learning agents will want to
use can use either hardware-accelerated rendering via EGL or GLX or
software rendering via OSMesa, depending on the
--define headless=...
build setting. - Python 2.7 (other versions might work, too) with NumPy, PIL (a few tests require a NumPy version of at least 1.8), or Python 3 (at least 3.5) with NumPy and Pillow.
The build rules are using a few compiler settings that are specific to GCC. If some flags are not recognized by your compiler (typically those would be specific warning suppressions), you may have to edit those flags. The warnings should be noisy but harmless.