• Stars
    star
    990
  • Rank 46,251 (Top 1.0 %)
  • Language
    Python
  • License
    MIT License
  • Created over 6 years ago
  • Updated 3 months ago

Reviews

There are no reviews yet. Be the first to send feedback to the community and the maintainers!

Repository Details

PyTorch Extension Library of Optimized Autograd Sparse Matrix Operations

PyTorch Sparse

PyPI Version Testing Status Linting Status Code Coverage


This package consists of a small extension library of optimized sparse matrix operations with autograd support. This package currently consists of the following methods:

All included operations work on varying data types and are implemented both for CPU and GPU. To avoid the hazzle of creating torch.sparse_coo_tensor, this package defines operations on sparse tensors by simply passing index and value tensors as arguments (with same shapes as defined in PyTorch). Note that only value comes with autograd support, as index is discrete and therefore not differentiable.

Installation

Anaconda

Update: You can now install pytorch-sparse via Anaconda for all major OS/PyTorch/CUDA combinations πŸ€— Given that you have pytorch >= 1.8.0 installed, simply run

conda install pytorch-sparse -c pyg

Binaries

We alternatively provide pip wheels for all major OS/PyTorch/CUDA combinations, see here.

PyTorch 2.0

To install the binaries for PyTorch 2.0.0, simply run

pip install torch-scatter torch-sparse -f https://data.pyg.org/whl/torch-2.0.0+${CUDA}.html

where ${CUDA} should be replaced by either cpu, cu117, or cu118 depending on your PyTorch installation.

cpu cu117 cu118
Linux βœ… βœ… βœ…
Windows βœ… βœ… βœ…
macOS βœ…

PyTorch 1.13

To install the binaries for PyTorch 1.13.0, simply run

pip install torch-scatter torch-sparse -f https://data.pyg.org/whl/torch-1.13.0+${CUDA}.html

where ${CUDA} should be replaced by either cpu, cu116, or cu117 depending on your PyTorch installation.

cpu cu116 cu117
Linux βœ… βœ… βœ…
Windows βœ… βœ… βœ…
macOS βœ…

Note: Binaries of older versions are also provided for PyTorch 1.4.0, PyTorch 1.5.0, PyTorch 1.6.0, PyTorch 1.7.0/1.7.1, PyTorch 1.8.0/1.8.1, PyTorch 1.9.0, PyTorch 1.10.0/1.10.1/1.10.2, PyTorch 1.11.0 and PyTorch 1.12.0/1.12.1 (following the same procedure). For older versions, you need to explicitly specify the latest supported version number or install via pip install --no-index in order to prevent a manual installation from source. You can look up the latest supported version number here.

From source

Ensure that at least PyTorch 1.7.0 is installed and verify that cuda/bin and cuda/include are in your $PATH and $CPATH respectively, e.g.:

$ python -c "import torch; print(torch.__version__)"
>>> 1.7.0

$ echo $PATH
>>> /usr/local/cuda/bin:...

$ echo $CPATH
>>> /usr/local/cuda/include:...

If you want to additionally build torch-sparse with METIS support, e.g. for partioning, please download and install the METIS library by following the instructions in the Install.txt file. Note that METIS needs to be installed with 64 bit IDXTYPEWIDTH by changing include/metis.h. Afterwards, set the environment variable WITH_METIS=1.

Then run:

pip install torch-scatter torch-sparse

When running in a docker container without NVIDIA driver, PyTorch needs to evaluate the compute capabilities and may fail. In this case, ensure that the compute capabilities are set via TORCH_CUDA_ARCH_LIST, e.g.:

export TORCH_CUDA_ARCH_LIST="6.0 6.1 7.2+PTX 7.5+PTX"

Functions

Coalesce

torch_sparse.coalesce(index, value, m, n, op="add") -> (torch.LongTensor, torch.Tensor)

Row-wise sorts index and removes duplicate entries. Duplicate entries are removed by scattering them together. For scattering, any operation of torch_scatter can be used.

Parameters

  • index (LongTensor) - The index tensor of sparse matrix.
  • value (Tensor) - The value tensor of sparse matrix.
  • m (int) - The first dimension of sparse matrix.
  • n (int) - The second dimension of sparse matrix.
  • op (string, optional) - The scatter operation to use. (default: "add")

Returns

  • index (LongTensor) - The coalesced index tensor of sparse matrix.
  • value (Tensor) - The coalesced value tensor of sparse matrix.

Example

import torch
from torch_sparse import coalesce

index = torch.tensor([[1, 0, 1, 0, 2, 1],
                      [0, 1, 1, 1, 0, 0]])
value = torch.Tensor([[1, 2], [2, 3], [3, 4], [4, 5], [5, 6], [6, 7]])

index, value = coalesce(index, value, m=3, n=2)
print(index)
tensor([[0, 1, 1, 2],
        [1, 0, 1, 0]])
print(value)
tensor([[6.0, 8.0],
        [7.0, 9.0],
        [3.0, 4.0],
        [5.0, 6.0]])

Transpose

torch_sparse.transpose(index, value, m, n) -> (torch.LongTensor, torch.Tensor)

Transposes dimensions 0 and 1 of a sparse matrix.

Parameters

  • index (LongTensor) - The index tensor of sparse matrix.
  • value (Tensor) - The value tensor of sparse matrix.
  • m (int) - The first dimension of sparse matrix.
  • n (int) - The second dimension of sparse matrix.
  • coalesced (bool, optional) - If set to False, will not coalesce the output. (default: True)

Returns

  • index (LongTensor) - The transposed index tensor of sparse matrix.
  • value (Tensor) - The transposed value tensor of sparse matrix.

Example

import torch
from torch_sparse import transpose

index = torch.tensor([[1, 0, 1, 0, 2, 1],
                      [0, 1, 1, 1, 0, 0]])
value = torch.Tensor([[1, 2], [2, 3], [3, 4], [4, 5], [5, 6], [6, 7]])

index, value = transpose(index, value, 3, 2)
print(index)
tensor([[0, 0, 1, 1],
        [1, 2, 0, 1]])
print(value)
tensor([[7.0, 9.0],
        [5.0, 6.0],
        [6.0, 8.0],
        [3.0, 4.0]])

Sparse Dense Matrix Multiplication

torch_sparse.spmm(index, value, m, n, matrix) -> torch.Tensor

Matrix product of a sparse matrix with a dense matrix.

Parameters

  • index (LongTensor) - The index tensor of sparse matrix.
  • value (Tensor) - The value tensor of sparse matrix.
  • m (int) - The first dimension of sparse matrix.
  • n (int) - The second dimension of sparse matrix.
  • matrix (Tensor) - The dense matrix.

Returns

  • out (Tensor) - The dense output matrix.

Example

import torch
from torch_sparse import spmm

index = torch.tensor([[0, 0, 1, 2, 2],
                      [0, 2, 1, 0, 1]])
value = torch.Tensor([1, 2, 4, 1, 3])
matrix = torch.Tensor([[1, 4], [2, 5], [3, 6]])

out = spmm(index, value, 3, 3, matrix)
print(out)
tensor([[7.0, 16.0],
        [8.0, 20.0],
        [7.0, 19.0]])

Sparse Sparse Matrix Multiplication

torch_sparse.spspmm(indexA, valueA, indexB, valueB, m, k, n) -> (torch.LongTensor, torch.Tensor)

Matrix product of two sparse tensors. Both input sparse matrices need to be coalesced (use the coalesced attribute to force).

Parameters

  • indexA (LongTensor) - The index tensor of first sparse matrix.
  • valueA (Tensor) - The value tensor of first sparse matrix.
  • indexB (LongTensor) - The index tensor of second sparse matrix.
  • valueB (Tensor) - The value tensor of second sparse matrix.
  • m (int) - The first dimension of first sparse matrix.
  • k (int) - The second dimension of first sparse matrix and first dimension of second sparse matrix.
  • n (int) - The second dimension of second sparse matrix.
  • coalesced (bool, optional): If set to True, will coalesce both input sparse matrices. (default: False)

Returns

  • index (LongTensor) - The output index tensor of sparse matrix.
  • value (Tensor) - The output value tensor of sparse matrix.

Example

import torch
from torch_sparse import spspmm

indexA = torch.tensor([[0, 0, 1, 2, 2], [1, 2, 0, 0, 1]])
valueA = torch.Tensor([1, 2, 3, 4, 5])

indexB = torch.tensor([[0, 2], [1, 0]])
valueB = torch.Tensor([2, 4])

indexC, valueC = spspmm(indexA, valueA, indexB, valueB, 3, 3, 2)
print(indexC)
tensor([[0, 1, 2],
        [0, 1, 1]])
print(valueC)
tensor([8.0, 6.0, 8.0])

Running tests

pytest

C++ API

torch-sparse also offers a C++ API that contains C++ equivalent of python models. For this, we need to add TorchLib to the -DCMAKE_PREFIX_PATH (e.g., it may exists in {CONDA}/lib/python{X.X}/site-packages/torch if installed via conda):

mkdir build
cd build
# Add -DWITH_CUDA=on support for CUDA support
cmake -DCMAKE_PREFIX_PATH="..." ..
make
make install

More Repositories

1

pytorch_scatter

PyTorch Extension Library of Optimized Scatter Operations
Python
1,531
star
2

pytorch_cluster

PyTorch Extension Library of Optimized Graph Cluster Algorithms
C++
815
star
3

deep-graph-matching-consensus

Implementation of "Deep Graph Matching Consensus" in PyTorch
Python
256
star
4

pytorch_spline_conv

Implementation of the Spline-Based Convolution Operator of SplineCNN in PyTorch
C++
170
star
5

pyg_autoscale

Implementation of "GNNAutoScale: Scalable and Expressive Graph Neural Networks via Historical Embeddings" in PyTorch
Python
157
star
6

table2excel

Convert and download html tables to a xlsx-file that can be opened in Microsoft Excel
JavaScript
112
star
7

deep-learning-cheatsheet

TeX
92
star
8

embedded_gcnn

Embedded Graph Convolutional Neural Networks (EGCNN) in TensorFlow
Jupyter Notebook
78
star
9

himp-gnn

Hierarchical Inter-Message Passing for Learning on Molecular Graphs
Python
75
star
10

koa2-rest-api

ES6 RESTFul Koa2 API with Mongoose and OAuth2
JavaScript
75
star
11

graph-based-image-classification

Implementation of Planar Graph Convolutional Networks in TensorFlow
Python
43
star
12

pytorch_unique

PyTorch Extension Library of Optimized Unique Operation
Python
37
star
13

deep-learning-on-graphs

TeX
31
star
14

mongoose-i18n-localize

Mongoose plugin to support i18n and localization
JavaScript
22
star
15

dotfiles

Shell
18
star
16

RSClipperWrapper

A small and simple wrapper for the Clipper library to perform polygon clipping (Swift)
C++
17
star
17

RSShapeNode

A RSShapeNode object draws a shape by rendering a Core Graphics path offscreen using a disconnected CAShapeLayer and snapshots the image to a SKSpriteNode (Swift)
Swift
8
star
18

rusty1s.github.io

HTML
6
star
19

pytorch_bincount

Python
6
star
20

vim-happy-hacking

Vim Script
5
star
21

rusty1s

4
star
22

table-select

Allows you to select table row elements like in your standard finder environment
JavaScript
3
star
23

DigDeeper

the Mining / Crafting / Trading game (Swift 2.0)
C++
3
star
24

react-pattern-library

React Pattern Library for various UI components
JavaScript
3
star
25

mongoose-i18n-error

lightweight module for node.js/express.js to create beautiful mongoose i18n validation error messages
JavaScript
2
star
26

react-dev-config

Customizable Configuration for modern React apps
JavaScript
2
star
27

mongoose-integer

mongoose plugin to validate integer values within a Mongoose Schema
JavaScript
2
star
28

hyper-happy-hacking

JavaScript
1
star
29

RSRoundBorderedButton

Round bordered Button like the ones used in the Apple AppStore (Swift)
Swift
1
star
30

ComputationOffloading

Energieeffizienz durch Computation Offloading in der Cloud
1
star
31

react-documentviewer

React Documentviewer for various mimetypes
JavaScript
1
star
32

RSRandomPolygon

Swift
1
star
33

tensorflow-graph-plugin

Python
1
star
34

dependent-select-boxes

Allows a child select box to change its options dependent on its parent select box
JavaScript
1
star
35

texture-synthesis

TeX
1
star
36

RSScene

An inheritance of SKScene that adds a game logic loop to the runtime of a scene (Swift)
Swift
1
star
37

OCF-andCP-Networks

Qualitative Semantiken fΓΌr DAGs - ein Vergleich von OCF- und CP-Netzwerken
1
star
38

js-dev-utils

JavaScript
1
star
39

RSContactGrid

A triangular/square/rotated square/hexagonal grid tile map with contact detection for any path (Swift 2.0)
Swift
1
star