• Stars
    star
    147
  • Rank 251,347 (Top 5 %)
  • Language
    Rust
  • License
    Other
  • Created about 4 years ago
  • Updated 7 months ago

Reviews

There are no reviews yet. Be the first to send feedback to the community and the maintainers!

Repository Details

A zenoh plug-in that allows to transparently route DDS data. This plugin can be used by DDS applications, such as ROS2 robotic applications and others, to leverage the zenoh for geographical routing or for better scaling discovery.

Discussion Discord License License

Eclipse Zenoh

The Eclipse Zenoh: Zero Overhead Pub/sub, Store/Query and Compute.

Zenoh (pronounce /zeno/) unifies data in motion, data at rest and computations. It carefully blends traditional pub/sub with geo-distributed storages, queries and computations, while retaining a level of time and space efficiency that is well beyond any of the mainstream stacks.

Check the website zenoh.io and the roadmap for more detailed information.


DDS plugin and standalone zenoh-bridge-dds

👉 Install latest release: see below

👉 Docker image: see below

👉 Build "master" branch: see below

Background

The Data Distribution Service (DDS) is a standard for data-centric publish subscribe. Whilst DDS has been around for quite some time and has a long history of deployments in various industries, it has recently gained quite a bit of attentions thanks to its adoption by the Robotic Operating System (ROS2) -- where it is used for communication between ROS2 nodes.

Robot Swarms and Edge Robotics

As mentioned above, ROS2 has adopted DDS as the mechanism to exchange data between nodes within and potentially across a robot. That said, due to some of the very core assumptions at the foundations of the DDS wire-protocol, beside the fact that it leverages UDP/IP multicast for communication, it is not so straightforward to scale DDS communication over a WAN or across multiple LANs. Zenoh, on the other hand was designed since its inception to operate at Internet Scale.

zenoh-plugin-dds

Thus, the main motivations to have a DDS plugin for Eclipse zenoh are:

  • Facilitate the interconnection of robot swarms.
  • Support use cases of edge robotics.
  • Give the possibility to use zenoh's geo-distributed storage and query system to better manage robot's data.

As any plugin for Eclipse zenoh, it can be dynamically loaded by a zenoh router, at startup or at runtime.
In addition, this project also provides a standalone version of this plugin as an executable binary named zenoh-bridge-dds.

How to install it

To install the latest release of either the DDS plugin for the Zenoh router, either the zenoh-bridge-dds standalone executable, you can do as follows:

Manual installation (all platforms)

All release packages can be downloaded from:

Each subdirectory has the name of the Rust target. See the platforms each target corresponds to on https://doc.rust-lang.org/stable/rustc/platform-support.html

Choose your platform and download:

  • the zenoh-plugin-dds-<version>-<platform>.zip file for the plugin.
    Then unzip it in the same directory than zenohd or to any directory where it can find the plugin library (e.g. /usr/lib)
  • the zenoh-bridge-dds-<version>-<platform>.zip file for the standalone executable.
    Then unzip it where you want, and run the extracted zenoh-bridge-dds binary.

Linux Debian

Add Eclipse Zenoh private repository to the sources list:

echo "deb [trusted=yes] https://download.eclipse.org/zenoh/debian-repo/ /" | sudo tee -a /etc/apt/sources.list > /dev/null
sudo apt update

Then either:

  • install the plugin with: sudo apt install zenoh-plugin-dds.
  • install the standalone executable with: sudo apt install zenoh-bridge-dds.

How to build it

⚠️ WARNING ⚠️ : Zenoh and its ecosystem are under active development. When you build from git, make sure you also build from git any other Zenoh repository you plan to use (e.g. binding, plugin, backend, etc.). It may happen that some changes in git are not compatible with the most recent packaged Zenoh release (e.g. deb, docker, pip). We put particular effort in mantaining compatibility between the various git repositories in the Zenoh project.

⚠️ WARNING ⚠️ : As Rust doesn't have a stable ABI, the plugins should be built with the exact same Rust version than zenohd, and using for zenoh dependency the same version (or commit number) than 'zenohd'. Otherwise, incompatibilities in memory mapping of shared types between zenohd and the library can lead to a "SIGSEV" crash.

In order to build the zenoh bridge for DDS you need first to install the following dependencies:

  • Rust. If you already have the Rust toolchain installed, make sure it is up-to-date with:

    $ rustup update
  • On Linux, make sure the llvm and clang development packages are installed:

    • on Debians do: sudo apt install llvm-dev libclang-dev
    • on CentOS or RHEL do: sudo yum install llvm-devel clang-devel
    • on Alpine do: apk install llvm11-dev clang-dev
  • CMake (to build CycloneDDS which is a native dependency)

Once these dependencies are in place, you may clone the repository on your machine:

$ git clone https://github.com/eclipse-zenoh/zenoh-plugin-dds.git
$ cd zenoh-plugin-dds

⚠️ WARNING ⚠️ : On Linux, don't use cargo build command without specifying a package with -p. Building both zenoh-plugin-dds (plugin library) and zenoh-bridge-dds (standalone executable) together will lead to a multiple definition of load_plugin'` error at link time. See #117 for explanations.

You can then choose between building the zenoh bridge for DDS:

  • as a plugin library that can be dynamically loaded by the zenoh router (zenohd):
$ cargo build --release -p zenoh-plugin-dds

The plugin shared library (*.so on Linux, *.dylib on Mac OS, *.dll on Windows) will be generated in the target/release subdirectory.

  • or as a standalone executable binary:
$ cargo build --release -p zenoh-bridge-dds

The zenoh-bridge-dds binary will be generated in the target/release sub-directory.

Enabling Cyclone DDS Shared Memory Support

Cyclone DDS Shared memory support is provided by the Iceoryx library. Iceoryx introduces additional system requirements which are documented here.

To build the zenoh bridge for DDS with support for shared memory the dds_shm optional feature must be enabled during the build process as follows:

  • plugin library:
$ cargo build --release -p zenoh-plugin-dds --features dds_shm
  • standalone executable binary:
$ cargo build --release -p zenoh-bridge-dds --features dds_shm

Note: Iceoryx does not need to be installed to build the bridge when the dds_shm feature is enabled. Iceoryx will be automatically downloaded, compiled, and statically linked into the zenoh bridge as part of the cargo build process.

When the zenoh bridge is configured to use DDS shared memory (see Configuration) the Iceoryx RouDi daemon (iox-roudi) must be running in order for the bridge to start successfully. If not started the bridge will wait for a period of time for the daemon to become available before timing out and terminating.

When building the zenoh bridge with the dds_shm feature enabled the iox-roudi daemon is also built for convenience. The daemon can be found under target/debug|release/build/cyclors-<hash>/out/iceoryx-build/bin/iox-roudi.

See here for more details of shared memory support in Cyclone DDS.

ROS2 package

If you're a ROS2 user, you can also build zenoh-bridge-dds as a ROS package running:

rosdep install --from-paths . --ignore-src -r -y
colcon build --packages-select zenoh_bridge_dds --cmake-args -DCMAKE_BUILD_TYPE=Release

The rosdep command will automatically install Rust and clang as build dependencies.

Docker image

The zenoh-bridge-dds standalone executable is also available as a Docker images for both amd64 and arm64. To get it, do:

  • docker pull eclipse/zenoh-bridge-dds:latest for the latest release
  • docker pull eclipse/zenoh-bridge-dds:master for the master branch version (nightly build)

⚠️ However, notice that it's usage is limited to Docker on Linux and using the --net host option.
The cause being that DDS uses UDP multicast and Docker doesn't support UDP multicast between a container and its host (see cases moby/moby#23659, moby/libnetwork#2397 or moby/libnetwork#552). The only known way to make it work is to use the --net host option that is only supported on Linux hosts.

Usage: docker run --init --net host eclipse/zenoh-bridge-dds
It supports the same command line arguments than the zenoh-bridge-dds (see below or check with -h argument).

For a quick test with ROS2 turtlesim

Prerequisites:

1 host, 2 ROS domains

For a quick test on a single host, you can run the turtlesim_node and the turtle_teleop_key on distinct ROS domains. As soon as you run 2 zenoh-bridge-dds (1 per domain) the turtle_teleop_key can drive the turtlesim_node.
Here are the commands to run:

  • ROS_DOMAIN_ID=1 ros2 run turtlesim turtlesim_node
  • ROS_DOMAIN_ID=2 ros2 run turtlesim turtle_teleop_key
  • ./target/release/zenoh-bridge-dds -d 1
  • ./target/release/zenoh-bridge-dds -d 2

Notice that by default the 2 bridges will discover each other using UDP multicast.

2 hosts, avoiding UDP multicast communication

By default DDS (and thus ROS2) uses UDP multicast for discovery and publications. But on some networks, UDP multicast is not or badly supported.
In such cases, deploying the zenoh-bridge-dds on both hosts will make it to:

  • limit the DDS discovery traffic, as detailled in this blog
  • route all the DDS publications made on UDP multicast by each node through the zenoh protocol that by default uses TCP.

Here are the commands to test this configuration with turtlesim:

  • on host 1:
    • ROS_DOMAIN_ID=1 ros2 run turtlesim turtlesim_node
    • ./target/release/zenoh-bridge-dds -d 1 -l tcp/0.0.0.0:7447
  • on host 2:
    • ROS_DOMAIN_ID=2 ros2 run turtlesim turtle_teleop_key
    • ./target/release/zenoh-bridge-dds -d 2 -e tcp/<host-1-ip>:7447 - where <host-1-ip> is the IP of host 1

Notice that to avoid unwanted direct DDS communication, 2 disctinct ROS domains are still used.

2 hosts, with an intermediate zenoh router in the cloud

In case your 2 hosts can't have a point-to-point communication, you could leverage a zenoh router deployed in a cloud instance (any Linux VM will do the job). You just need to configure your cloud instanse with a public IP and authorize the TCP port 7447.

⚠️ the zenoh protocol is still under development leading to possible incompatibilities between the bridge and the router if their zenoh version differ. Please make sure you use a zenoh router built from a recent commit id from its master branch.

Here are the commands to test this configuration with turtlesim:

  • on cloud VM:
    • zenohd
  • on host 1:
    • ros2 run turtlesim turtlesim_node
    • ./target/release/zenoh-bridge-dds -e tcp/<cloud-ip>:7447
      where <cloud-ip> is the IP of your cloud instance
  • on host 2:
    • ros2 run turtlesim turtle_teleop_key
    • ./target/release/zenoh-bridge-dds -e tcp/<cloud-ip>:7447
      where <cloud-ip> is the IP of your cloud instance

Notice that there is no need to use distinct ROS domain here, since the 2 hosts are not supposed to directly communicate with each other.

More advanced usage for ROS2

Full support of ROS graph and topic lists via the forward discovery mode

By default the bridge doesn't route throught zenoh the DDS discovery traffic to the remote bridges.
Meaning that, in case you use 2 zenoh-bridge-dds to interconnect 2 DDS domains, the DDS entities discovered in one domain won't be advertised in the other domain. Thus, the DDS data will be routed between the 2 domains only if matching readers and writers are declared in the 2 domains independently.

This default behaviour has an impact on ROS2 behaviour: on one side of the bridge the ROS graph might not reflect all the nodes from the other side of the bridge. The ros2 topic list command might not list all the topics declared on the other side. And the ROS graph is limited to the nodes in each domain.

But using the --fwd-discovery (or -f) option for all bridges make them behave differently:

  • each bridge will forward via zenoh the local DDS discovery data to the remote bridges (in a more compact way than the original DDS discovery traffic)
  • each bridge receiving DDS discovery data via zenoh will create a replica of the DDS reader or writer, with similar QoS. Those replicas will serve the route to/from zenoh, and will be discovered by the ROS2 nodes.
  • each bridge will forward the ros_discovery_info data (in a less intensive way than the original publications) to the remote bridges. On reception, the remote bridges will convert the original entities' GIDs into the GIDs of the corresponding replicas, and re-publish on DDS the ros_discovery_info. The full ROS graph can then be discovered by the ROS2 nodes on each host.

Limiting the ROS2 topics, services, parameters or actions to be routed

By default 2 zenoh bridges will route all ROS2 topics and services for which they detect a Writer on one side and a Reader on the other side. But you might want to avoid some topics and services to be routed by the bridge.

Starting zenoh-bridge-dds you can use the --allow argument to specify the subset of topics and services that will be routed by the bridge. This argument accepts a string wich is a regular expression that must match a substring of an allowed zenoh key (see details of mapping of ROS2 names to zenoh keys).

Here are some examples of usage:

--allow value allowed ROS2 communication
/rosout /rosout
/rosout|/turtle1/cmd_vel|/turtle1/rotate_absolute /rosout
/turtle1/cmd_vel
/turtle1/rotate_absolute
/rosout|/turtle1/ /rosout and all /turtle1 topics, services, parameters and actions
/turtle1/.* all topics and services with name containing /turtle1/
/turtle1/ same: all topics, services, parameters and actions with name containing /turtle1/
rt/turtle1 all topics with name containing /turtle1 (no services, parameters or actions)
rq/turtle1|/rr/turtle1 all services and parameters with name containing /turtle1 (no topics or actions)
rq/turtlesim/.*parameter|/rr/turtlesim/.*parameter all parameters with name containing /turtlesim (no topics, services or actions)
rq/turtle1/.*/_action|/rr/turtle1/.*/_action all actions with name containing /turtle1 (no topics, services or parameters)

Running several robots without changing the ROS2 configuration

If you run similar robots in the same network, they will by default all us the same DDS topics, leading to interferences in their operations.
A simple way to address this issue using the zenoh bridge is to:

  • deploy 1 zenoh bridge per robot
  • have each bridge started with the --scope "/<id>" argument, each robot having its own id.
  • make sure each robot cannot directly communicate via DDS with another robot by setting a distinct domain per robot, or configuring its network interface to not route UDP multicast outside the host.

Using the --scope option, a prefix is added to each zenoh key published/subscribed by the bridge (more details in mapping of ROS2 names to zenoh keys). To interact with a robot, a remote ROS2 application must use a zenoh bridge configured with the same scope than the robot.

Closer integration of ROS2 with zenoh

As you understood, using the zenoh bridge, each ROS2 publications and subscriptions are mapped to a zenoh key. Therefore, its relatively easy to develop an application using one of the zenoh APIs to interact with one or more robot at the same time.

See in details how to achieve that in this blog.

Configuration

zenoh-bridge-dds can be configured via a JSON5 file passed via the -cargument. You can see a commented example of such configuration file: DEFAULT_CONFIG.json5.

The "dds" part of this same configuration file can also be used in the configuration file for the zenoh router (within its "plugins" part). The router will automatically try to load the plugin library (zenoh-plugin_dds) at startup and apply its configuration.

zenoh-bridge-dds also accepts the following arguments. If set, each argument will override the similar setting from the configuration file:

  • zenoh-related arguments:
    • -c, --config <FILE> : a config file
    • -m, --mode <MODE> : The zenoh session mode. Default: peer Possible values: peer or client.
      See zenoh documentation for more details.
    • -l, --listen <LOCATOR> : A locator on which this router will listen for incoming sessions. Repeat this option to open several listeners. Example of locator: tcp/localhost:7447.
    • -e, --peer <LOCATOR> : A peer locator this router will try to connect to (typically another bridge or a zenoh router). Repeat this option to connect to several peers. Example of locator: tcp/<ip-address>:7447.
    • --no-multicast-scouting : disable the zenoh scouting protocol that allows automatic discovery of zenoh peers and routers.
    • -i, --id <hex_string> : The identifier (as an hexadecimal string - e.g.: 0A0B23...) that the zenoh bridge must use. WARNING: this identifier must be unique in the system! If not set, a random UUIDv4 will be used.
    • --group-member-id <ID> : The bridges are supervising each other via zenoh liveliness tokens. This option allows to set a custom identifier for the bridge, that will be used the liveliness token key (if not specified, the zenoh UUID is used).
    • --rest-http-port <rest-http-port> : set the REST API http port (default: 8000)
  • DDS-related arguments:
    • -d, --domain <ID> : The DDS Domain ID. By default set to 0, or to "$ROS_DOMAIN_ID" is this environment variable is defined.

    • --dds-localhost-only : If set, the DDS discovery and traffic will occur only on the localhost interface (127.0.0.1). By default set to false, unless the "ROS_LOCALHOST_ONLY=1" environment variable is defined.

    • --dds-enable-shm : If set, DDS will be configured to use shared memory. Requires the bridge to be built with the 'dds_shm' feature for this option to valid. By default set to false.

    • -f, --fwd-discovery : When set, rather than creating a local route when discovering a local DDS entity, this discovery info is forwarded to the remote plugins/bridges. Those will create the routes, including a replica of the discovered entity. More details here

    • -s, --scope <String> : A string used as prefix to scope DDS traffic when mapped to zenoh keys.

    • -a, --allow <String> : A regular expression matching the set of 'partition/topic-name' that must be routed via zenoh. By default, all partitions and topics are allowed.
      If both 'allow' and 'deny' are set a partition and/or topic will be allowed if it matches only the 'allow' expression.
      Repeat this option to configure several topic expressions. These expressions are concatenated with '|'. Examples of expressions:

      • .*/TopicA will allow only the TopicA to be routed, whatever the partition.
      • PartitionX/.* will allow all the topics to be routed, but only on PartitionX.
      • cmd_vel|rosout will allow only the topics containing cmd_vel or rosout in their name or partition name to be routed.
    • --deny <String> : A regular expression matching the set of 'partition/topic-name' that must NOT be routed via zenoh. By default, no partitions and no topics are denied.
      If both 'allow' and 'deny' are set a partition and/or topic will be allowed if it matches only the 'allow' expression.
      Repeat this option to configure several topic expressions. These expressions are concatenated with '|'.

    • --max-frequency <String>... : specifies a maximum frequency of data routing over zenoh per-topic. The string must have the format "regex=float" where:

      • "regex" is a regular expression matching the set of 'partition/topic-name' for which the data (per DDS instance) must be routedat no higher rate than associated max frequency (same syntax than --allow option).
      • "float" is the maximum frequency in Hertz; if publication rate is higher, downsampling will occur when routing.

      (usable multiple times)

    • --queries-timeout <Duration>: A duration in seconds (default: 5.0 sec) that will be used as a timeout when the bridge queries any other remote bridge for discovery information and for historical data for TRANSIENT_LOCAL DDS Readers it serves (i.e. if the query to the remote bridge exceed the timeout, some historical samples might be not routed to the Readers, but the route will not be blocked forever).

    • -w, --generalise-pub <String> : A list of key expressions to use for generalising the declaration of the zenoh publications, and thus minimizing the discovery traffic (usable multiple times). See this blog for more details.

    • -r, --generalise-sub <String> : A list of key expressions to use for generalising the declaration of the zenoh subscriptions, and thus minimizing the discovery traffic (usable multiple times). See this blog for more details.

Admin space

The zenoh bridge for DDS exposes an administration space allowing to browse the DDS entities that have been discovered (with their QoS), and the routes that have been established between DDS and zenoh. This administration space is accessible via any zenoh API, including the REST API that you can activate at zenoh-bridge-dds startup using the --rest-http-port argument.

The zenoh-bridge-dds exposes this administration space with paths prefixed by @/service/<uuid>/dds (where <uuid> is the unique identifier of the bridge instance). The informations are then organized with such paths:

  • @/service/<uuid>/dds/version : the bridge version
  • @/service/<uuid>/dds/config : the bridge configuration
  • @/service/<uuid>/dds/participant/<gid>/reader/<gid>/<topic> : a discovered DDS reader on <topic>
  • @/service/<uuid>/dds/participant/<gid>/writer/<gid>/<topic> : a discovered DDS reader on <topic>
  • @/service/<uuid>/dds/route/from_dds/<zenoh-resource> : a route established from a DDS writer to a zenoh key named <zenoh-resource> (see mapping rules).
  • @/service/<uuid>/dds/route/to_dds/<zenoh-resource> : a route established from a zenoh key named <zenoh-resource> (see mapping rules)..

Example of queries on administration space using the REST API with the curl command line tool (don't forget to activate the REST API with --rest-http-port 8000 argument):

  • List all the DDS entities that have been discovered:
    curl http://localhost:8000/@/service/**/participant/**
  • List all established routes:
    curl http://localhost:8000/@/service/**/route/**
  • List all discovered DDS entities and established route for topic cmd_vel:
    curl http://localhost:8000/@/service/**/cmd_vel

Pro tip: pipe the result into jq command for JSON pretty print or transformation.

Architecture details

Whether it's built as a library or as a standalone executable, the zenoh bridge for DDS do the same things:

  • in default mode:

    • it discovers the DDS readers and writers declared by any DDS application, via the standard DDS discovery protocol (that uses UDP multicast)
    • it creates a mirror DDS writer or reader for each discovered reader or writer (using the same QoS)
    • if maps the discovered DDS topics and partitions to zenoh keys (see mapping details below)
    • it forwards user's data from a DDS topic to the corresponding zenoh key, and vice versa
    • it does not forward to the remote bridge any DDS discovery information
  • in "forward discovery" mode

    • it behaves as described here

Mapping of DDS topics to zenoh keys

The mapping between DDS and zenoh is rather straightforward: given a DDS Reader/Writer for topic A without the partition QoS set, then the equivalent zenoh key will have the same name: A. If a partition QoS P is defined, the equivalent zenoh key will be named as P/A.

Optionally, the bridge can be configured with a scope that will be used as a prefix to each zenoh key. That is, for scope S the equivalent zenoh key will be:

  • S/A for a topic A without partition
  • S/P/A for a topic A and a partition P

Mapping ROS2 names to zenoh keys

The mapping from ROS2 topics and services name to DDS topics is specified here. Notice that ROS2 does not use the DDS partitions.
As a consequence of this mapping and of the DDS to zenoh mapping specified above, here are some examples of mapping from ROS2 names to zenoh keys:

ROS2 names DDS Topics names zenoh keys (no scope) zenoh keys (if scope="myscope")
topic: /rosout rt/rosout rt/rosout myscope/rt/rosout
topic: /turtle1/cmd_vel rt/turtle1/cmd_vel rt/turtle1/cmd_vel myscope/rt/turtle1/cmd_vel
service: /turtle1/set_pen rq/turtle1/set_penRequest
rr/turtle1/set_penReply
rq/turtle1/set_penRequest
rr/turtle1/set_penReply
myscope/rq/turtle1/set_penRequest
myscope/rr/turtle1/set_penReply
action: /turtle1/rotate_absolute rq/turtle1/rotate_absolute/_action/send_goalRequest
rr/turtle1/rotate_absolute/_action/send_goalReply
rq/turtle1/rotate_absolute/_action/cancel_goalRequest
rr/turtle1/rotate_absolute/_action/cancel_goalReply
rq/turtle1/rotate_absolute/_action/get_resultRequest
rr/turtle1/rotate_absolute/_action/get_resultReply
rt/turtle1/rotate_absolute/_action/status
rt/turtle1/rotate_absolute/_action/feedback
rq/turtle1/rotate_absolute/_action/send_goalRequest
rr/turtle1/rotate_absolute/_action/send_goalReply
rq/turtle1/rotate_absolute/_action/cancel_goalRequest
rr/turtle1/rotate_absolute/_action/cancel_goalReply
rq/turtle1/rotate_absolute/_action/get_resultRequest
rr/turtle1/rotate_absolute/_action/get_resultReply
rt/turtle1/rotate_absolute/_action/status
rt/turtle1/rotate_absolute/_action/feedback
myscope/rq/turtle1/rotate_absolute/_action/send_goalRequest
myscope/rr/turtle1/rotate_absolute/_action/send_goalReply
myscope/rq/turtle1/rotate_absolute/_action/cancel_goalRequest
myscope/rr/turtle1/rotate_absolute/_action/cancel_goalReply
myscope/rq/turtle1/rotate_absolute/_action/get_resultRequest
myscope/rr/turtle1/rotate_absolute/_action/get_resultReply
myscope/rt/turtle1/rotate_absolute/_action/status
myscope/rt/turtle1/rotate_absolute/_action/feedback
all parameters for node turtlesim rq/turtlesim/list_parametersRequest
rr/turtlesim/list_parametersReply
rq/turtlesim/describe_parametersRequest
rr/turtlesim/describe_parametersReply
rq/turtlesim/get_parametersRequest
rr/turtlesim/get_parametersReply
rr/turtlesim/get_parameter_typesReply
rq/turtlesim/get_parameter_typesRequest
rq/turtlesim/set_parametersRequest
rr/turtlesim/set_parametersReply
rq/turtlesim/set_parameters_atomicallyRequest
rr/turtlesim/set_parameters_atomicallyReply
rq/turtlesim/list_parametersRequest
rr/turtlesim/list_parametersReply
rq/turtlesim/describe_parametersRequest
rr/turtlesim/describe_parametersReply
rq/turtlesim/get_parametersRequest
rr/turtlesim/get_parametersReply
rr/turtlesim/get_parameter_typesReply
rq/turtlesim/get_parameter_typesRequest
rq/turtlesim/set_parametersRequest
rr/turtlesim/set_parametersReply
rq/turtlesim/set_parameters_atomicallyRequest
rr/turtlesim/set_parameters_atomicallyReply
myscope/rq/turtlesim/list_parametersRequest
myscope/rr/turtlesim/list_parametersReply
myscope/rq/turtlesim/describe_parametersRequest
myscope/rr/turtlesim/describe_parametersReply
myscope/rq/turtlesim/get_parametersRequest
myscope/rr/turtlesim/get_parametersReply
myscope/rr/turtlesim/get_parameter_typesReply
myscope/rq/turtlesim/get_parameter_typesRequest
myscope/rq/turtlesim/set_parametersRequest
myscope/rr/turtlesim/set_parametersReply
myscope/rq/turtlesim/set_parameters_atomicallyRequest
myscope/rr/turtlesim/set_parameters_atomicallyReply
specific ROS discovery topic ros_discovery_info ros_discovery_info myscope/ros_discovery_info

More Repositories

1

zenoh

zenoh unifies data in motion, data in-use, data at rest and computations. It carefully blends traditional pub/sub with geo-distributed storages, queries and computations, while retaining a level of time and space efficiency that is well beyond any of the mainstream stacks.
Rust
1,256
star
2

zenoh-pico

Eclipse zenoh for pico devices
C
86
star
3

zenoh-plugin-ros2dds

A Zenoh plug-in for ROS2 with a DDS RMW.
Rust
76
star
4

zenoh-flow

zenoh-flow aims at providing a zenoh-based data-flow programming framework for computations that span from the cloud to the device.
Rust
71
star
5

zenoh-c

C API for Zenoh
Rust
63
star
6

zenoh-python

Python API for zenoh
Python
58
star
7

zenoh-cpp

C++ API for zenoh
C++
32
star
8

zenoh-demos

Some demos using Eclipse zenoh
C++
29
star
9

roadmap

23
star
10

zenoh-backend-influxdb

Backend and Storages for zenoh using InfluxDB
Rust
16
star
11

zenoh-go

Go-lang API for zenoh
Go
14
star
12

zenoh-plugin-webserver

A zenoh plug-in implementing an HTTP server mapping URLs to zenoh paths. This plugin can be used to set-up a Web server where the resources are retrieved from geo-distributed zenoh storages, each leveraging various backends (file system, database, memory...)
Rust
14
star
13

zenoh-backend-filesystem

Backend and Storages for zenoh using the file system
Rust
13
star
14

zenoh-backend-s3

Backend and Storages for Zenoh using AWS S3 API
Rust
13
star
15

zenoh-java

Java APIs for zenoh
Kotlin
12
star
16

zenoh-backend-rocksdb

Backend and Storages for zenoh using RocksDB
Rust
12
star
17

zenoh-csharp

C# API for zenoh
C#
11
star
18

zenoh-plugin-ros1

Rust
11
star
19

zenoh-plugin-mqtt

A Zenoh plug-in that allows to integrate and/or route MQTT pub/sub with Eclipse Zenoh.
Rust
11
star
20

zenoh-kotlin

Kotlin
9
star
21

zenoh-rust

zenoh-rust
7
star
22

zenoh-flow-python

Zenoh Flow Python API
Rust
6
star
23

homebrew-zenoh

Homebrew tap for Eclipse zenoh formulae (MacOS)
Ruby
3
star
24

zenoh-backend-sql

Backend and Storages for zenoh using an SQL database (SQLITE3, MySQL, PostgreSQL...)
3
star
25

ci

GitHub Actions and workflows used across eclipse-zenoh
TypeScript
1
star