• Stars
    star
    36,731
  • Rank 386 (Top 0.01 %)
  • Language
    Python
  • License
    Apache License 2.0
  • Created over 1 year ago
  • Updated 15 days ago

Reviews

There are no reviews yet. Be the first to send feedback to the community and the maintainers!

Repository Details

OpenAssistant is a chat-based assistant that understands tasks, can interact with third-party systems, and retrieve information dynamically to do so.

Open-Assistant

GitHub Repo stars Docs GitHub Workflow Status GitHub Workflow Status GitHub Workflow Status GitHub Workflow Status GitHub Workflow Status GitHub Workflow Status GitHub Workflow Status GitHub Workflow Status GitHub release (latest by date) Translate

Table of Contents


What is Open Assistant?

Open Assistant is a project meant to give everyone access to a great chat based large language model.

We believe that by doing this we will create a revolution in innovation in language. In the same way that stable-diffusion helped the world make art and images in new ways we hope Open Assistant can help improve the world by improving language itself.

Useful Links

How To Try It Out

Chatting with the AI

The chat frontend is now live here. Log in and start chatting! Please try to react with a thumbs up or down for the assistant's responses when chatting.

Contributing to Data Collection

The data collection frontend is now live here. Log in and start taking on tasks! We want to collect a high volume of quality data. By submitting, ranking, and labelling model prompts and responses you will be directly helping to improve the capabilities of Open Assistant.

Running the Development Setup Locally (without chat)

You do not need to run the project locally unless you are contributing to the development process. The website link above will take you to the public website where you can use the data collection app and the chat.

If you would like to run the data collection app locally for development, you can set up an entire stack needed to run Open-Assistant, including the website, backend, and associated dependent services, with Docker.

To start the demo, run this in the root directory of the repository (check this FAQ if you have problems):

docker compose --profile ci up --build --attach-dependencies

Note: when running on MacOS with an M1 chip you have to use: DB_PLATFORM=linux/x86_64 docker compose ...

Then, navigate to http://localhost:3000 (It may take some time to boot up) and interact with the website.

Note: If an issue occurs with the build, please head to the FAQ and check out the entries about Docker.

Note: When logging in via email, navigate to http://localhost:1080 to get the magic email login link.

Note: If you would like to run this in a standardized development environment (a "devcontainer") using vscode locally or in a web browser using GitHub Codespaces, you can use the provided .devcontainer folder.

Running the Development Setup Locally for Chat

You do not need to run the project locally unless you are contributing to the development process. The website link above will take you to the public website where you can use the data collection app and the chat.

Also note that the local setup is only for development and is not meant to be used as a local chatbot, unless you know what you are doing.

If you do know what you are doing, then see the inference folder for getting the inference system up and running, or have a look at --profile inference in addition to --profile ci in the above command.

The Vision

We are not going to stop at replicating ChatGPT. We want to build the assistant of the future, able to not only write email and cover letters, but do meaningful work, use APIs, dynamically research information, and much more, with the ability to be personalized and extended by anyone. And we want to do this in a way that is open and accessible, which means we must not only build a great assistant, but also make it small and efficient enough to run on consumer hardware.

The Plan

We want to get to an initial MVP as fast as possible, by following the 3-steps outlined in the InstructGPT paper
  1. Collect high-quality human generated Instruction-Fulfillment samples (prompt + response), goal >50k. We design a crowdsourced process to collect and reviewed prompts. We do not want to train on flooding/toxic/spam/junk/personal information data. We will have a leaderboard to motivate the community that shows progress and the most active users. Swag will be given to the top-contributors.
  2. For each of the collected prompts we will sample multiple completions. Completions of one prompt will then be shown randomly to users to rank them from best to worst. Again this should happen crowd-sourced, e.g. we need to deal with unreliable potentially malicious users. At least multiple votes by independent users have to be collected to measure the overall agreement. The gathered ranking-data will be used to train a reward model.
  3. Now follows the RLHF training phase based on the prompts and the reward model.

We can then take the resulting model and continue with completion sampling step 2 for a next iteration.

Slide Decks

Vision & Roadmap

Important Data Structures

How You Can Help

All open source projects begin with people like you. Open source is the belief that if we collaborate we can together gift our knowledge and technology to the world for the benefit of humanity.

Check out our contributing guide to get started.

More Repositories

1

audio-dataset

Audio Dataset for training CLAP and other models
Python
575
star
2

CLIP_benchmark

CLIP-like model evaluation
Jupyter Notebook
497
star
3

dalle2-laion

Pretrained Dalle2 from laion
Python
489
star
4

CLAP

Contrastive Language-Audio Pretraining
Python
479
star
5

natural_voice_assistant

Python
349
star
6

laion-3d

Collect large 3d dataset and build models
238
star
7

phenaki

A phenaki reproduction using pytorch.
Python
218
star
8

aesthetic-predictor

A linear estimator on top of clip to predict the aesthetic quality of pictures
Jupyter Notebook
199
star
9

Open-Instruction-Generalist

Open Instruction Generalist is an assistant trained on massive synthetic instructions to perform many millions of tasks
Python
195
star
10

ldm-finetune

Home of `erlich` and `ongo`. Finetune latent-diffusion/glid-3-xl text2image on your own data.
Python
169
star
11

CLIP-based-NSFW-Detector

Python
135
star
12

scaling-laws-openclip

Reproducible scaling laws for contrastive language-image learning (https://arxiv.org/abs/2212.07143)
Jupyter Notebook
135
star
13

laion-datasets

Description and pointers of laion datasets
HTML
131
star
14

laion-dreams

Aim for the moon. If you miss, you may hit a star.
121
star
15

laion.ai

HTML
99
star
16

Open-GIA

O-GIA is an umbrella for research, infrastructure and projects ecosystem that should provide open source, reproducible datasets, models, applications & safety tools for Open Generalist Interactive Agents (O-GIA). O-GIA systems will act in collaboration with human or autonomously, supporting various kind of validated decision making and assistance.
88
star
17

video-clip

Let's make a video clip
85
star
18

LAION-5B-WatermarkDetection

Python
85
star
19

General-GPT

Jupyter Notebook
64
star
20

Text-to-speech

Python
59
star
21

Big-Interleaved-Dataset

Big-Interleaved-Dataset
Python
56
star
22

riverbed

Tools for content datamining and NLP at scale
Python
40
star
23

Discord-Scrapers

Implementation of a discord channel scraper to generate datasets.
Python
40
star
24

OCR-ensemble

Jupyter Notebook
37
star
25

Conditional-Pretraining-of-Large-Language-Models

Python
36
star
26

interesting-text-datasets

33
star
27

blade2blade

Adversarial Training and SFT for Bot Safety Models
Python
32
star
28

deep-image-diffusion-prior

Inverts CLIP text embeds to image embeds and visualizes with deep-image-prior.
Jupyter Notebook
31
star
29

watermark-detection

A repository containing datasets and tools to train a watermark classifier.
Python
31
star
30

temporal-embedding-aggregation

Aggregating embeddings over time
Python
30
star
31

medical

This repository will be a summary and outlook on all our open, medical, AI advancements.
Jupyter Notebook
28
star
32

Anh

Anh - LAION's multilingual assistant datasets and models
Python
27
star
33

laion50BU

Un-*** 50 billions multimodality dataset
24
star
34

conditioned-prior

(wip) Use LAION-AI's CLIP "conditoned prior" to generate CLIP image embeds from CLIP text embeds.
Python
18
star
35

LAION-SAFETY

An open toolbox for NSFW & toxicity detection
Jupyter Notebook
16
star
36

laion5B-paper

Building the laion5B paper
13
star
37

opendream

Frontend (and soon also midleware and backend) for a new, opensource image generation platform.
TypeScript
13
star
38

notebooks

A collection of generative and training notebooks getting mirrored to google colab.
Jupyter Notebook
12
star
39

laionide

This repository contains training code and checkpoitns for finetuning glide.
Python
12
star
40

laion-dedup

Python
12
star
41

super-resolution

This is the LAION repository for creating open super-resolution models with the help of LAION-5B subsets.
11
star
42

dataset-spec

Describe the format of image/text datasets
Python
10
star
43

LAION-PEOPLE

This project provides a data set with bounding boxes, body poses, 3D face meshes & captions of people from our LAION-2.2B. Additionally it provides clusters based on the poses and face meshes and pose-related captions based on these cluster assignments.
10
star
44

image-deduplication-testset

HTML
8
star
45

project-menu

Projects at LAION
8
star
46

laion-ai.github.io

laion github website
Svelte
6
star
47

dataset-usage

This repository is a summary of all systems and scientific papers that use LAION datasets.
6
star
48

repository-overview

This repository will give a quick overview of all projects and repositories from LAION.
5
star
49

KAISER

Knowledge Acquisition and Interlinking via Semantic Embeddings and Reasoning
4
star
50

LionizeR

Experiments with Summarization, Long Context and Retrieval
Python
4
star
51

safety-pipeline

A collection of safety classifiers and models to process image and texts.
Python
3
star
52

decentralized-learning

A basic setup for decentralized-learning that can be used for training future DALLE/CLIP/CLAP models.
3
star
53

diffusion-prior

DALL-E2 diffusion prior
Python
3
star
54

GIF

General / Global Inference Framework
Python
3
star
55

website

This is the development repository of the LAION-AI website.
HTML
3
star
56

lucidrains-projects

A summary of all lucidrains repositores and links to training / research approaches by LAION or other communities.
Jupyter Notebook
3
star
57

NeoGen

3
star
58

laion5b-subsets

Creating subsets from laion5b via embeddings search
Jupyter Notebook
2
star
59

human_artifacts

A repo containing images for artifact annotation.
2
star
60

public-relations

All media / publicity on LAION and related stuff!
2
star
61

dataset-inference

The new repository for the genral inference pipeline.
Python
2
star
62

public-domain-images

A collection of public domain images donated for ML training.
2
star
63

math_problems-step-by-step_solutions

Here we provide and collect many functions to generate math problem and step by step solutions for LLM training
Python
2
star
64

language-models

2
star
65

introduction-resources

Recommended intro resources
2
star
66

balanced-laion5b

This repository shall help finding a good distribution for huge datasets like LAION-5B for more efficient training.
2
star
67

hand-inference

A model to run hand inference on a cluster.
Jupyter Notebook
2
star
68

laion5b-bias

This repository is a collection of found biases in the LAION-5B dataset.
1
star
69

dataset-tasks

datasets that should be downloaded & converted to our standard training formart.
1
star