AI2 Tango replaces messy directories and spreadsheets full of file versions by organizing experiments into discrete steps that can be cached and reused throughout the lifetime of a research project.
Quick links
In this README
Quick start
Create a Tango step:
# hello.py
from tango import step
@step()
def hello(name: str) -> str:
message = f"Hello, {name}!"
print(message)
return message
And create a corresponding experiment configuration file:
// hello.jsonnet
{
steps: {
hello: {
type: "hello",
name: "World",
}
}
}
Then run the experiment using a local workspace to cache the result:
tango run hello.jsonnet -w /tmp/workspace
You'll see something like this in the output:
Starting new run expert-llama
β Starting step "hello"...
Hello, World!
β Finished step "hello"
β Finished run expert-llama
If you run this a second time the output will now look like this:
Starting new run open-crab
β Found output for step "hello" in cache...
β Finished run open-crab
You won't see "Hello, World!" this time because the result of the step was found in the cache, so it wasn't run again.
For a more detailed introduction check out the First Steps walk-through.
Installation
ai2-tango requires Python 3.8 or later.
pip
Installing with ai2-tango is available on PyPI. Just run
pip install ai2-tango
To install with a specific integration, such as torch
for example, run
pip install 'ai2-tango[torch]'
To install with all integrations, run
pip install 'ai2-tango[all]'
conda
Installing with ai2-tango is available on conda-forge. You can install just the base package with
conda install tango -c conda-forge
You can pick and choose from the integrations with one of these:
conda install tango-datasets -c conda-forge
conda install tango-torch -c conda-forge
conda install tango-wandb -c conda-forge
You can also install everything:
conda install tango-all -c conda-forge
Even though ai2-tango itself is quite small, installing everything will pull in a lot of dependencies. Don't be surprised if this takes a while!
Installing from source
To install ai2-tango from source, first clone the repository:
git clone https://github.com/allenai/tango.git
cd tango
Then run
pip install -e '.[all]'
To install with only a specific integration, such as torch
for example, run
pip install -e '.[torch]'
Or to install just the base tango library, you can run
pip install -e .
Checking your installation
Run
tango info
to check your installation.
Docker image
You can build a Docker image suitable for tango projects by using the official Dockerfile as a starting point for your own Dockerfile, or you can simply use one of our prebuilt images as a base image in your Dockerfile. For example:
# Start from a prebuilt tango base image.
# You can choose the right tag from the available options here:
# https://github.com/allenai/tango/pkgs/container/tango/versions
FROM ghcr.io/allenai/tango:cuda11.3
# Install your project's additional requirements.
COPY requirements.txt .
RUN /opt/conda/bin/pip install --no-cache-dir -r requirements.txt
# Install source code.
# This instruction copies EVERYTHING in the current directory (build context),
# which may not be what you want. Consider using a ".dockerignore" file to
# exclude files and directories that you don't want on the image.
COPY . .
Make sure to choose the right base image for your use case depending on the version of tango you're using and the CUDA version that your host machine supports. You can see a list of all available image tags on GitHub.
FAQ
Why is the library named Tango?
The motivation behind this library is that we can make research easier by composing it into well-defined steps. What happens when you choreograph a number of steps together? Well, you get a dance. And since our team's leader is part of a tango band, "AI2 Tango" was an obvious choice!
How can I debug my steps through the Tango CLI?
You can run the tango
command through pdb. For example:
python -m pdb -m tango run config.jsonnet
Metaflow, Airflow, or redun?
How is Tango different fromWe've found that existing DAG execution engines like these tools are great for production workflows but not as well suited for messy, collaborative research projects where code is changing constantly. AI2 Tango was built specifically for these kinds of research projects.
How does Tango's caching mechanism work?
AI2 Tango caches the results of steps based on the unique_id
of the step. The unique_id
is essentially a hash of all of the inputs to the step along with:
- the step class's fully qualified name, and
- the step class's
VERSION
class variable (an arbitrary string).
Unlike other workflow engines like redun, Tango does not take into account the source code of the class itself (other than its fully qualified name) because we've found that using a hash of the source code bytes is way too sensitive and less transparent for users.
When you change the source code of your step in a meaningful way you can just manually change the VERSION
class variable to indicate to Tango
that the step has been updated.
Team
ai2-tango is developed and maintained by the AllenNLP team, backed by the Allen Institute for Artificial Intelligence (AI2). AI2 is a non-profit institute with the mission to contribute to humanity through high-impact AI research and engineering. To learn more about who specifically contributed to this codebase, see our contributors page.
License
ai2-tango is licensed under Apache 2.0. A full copy of the license can be found on GitHub.