• Stars
    star
    1,699
  • Rank 27,472 (Top 0.6 %)
  • Language
    TypeScript
  • License
    MIT License
  • Created about 1 year ago
  • Updated 7 months ago

Reviews

There are no reviews yet. Be the first to send feedback to the community and the maintainers!

Repository Details

Replace Copilot local AI

Llama Coder

Llama Coder is a better and self-hosted Github Copilot replacement for VS Studio Code. Llama Coder uses Ollama and codellama to provide autocomplete that runs on your hardware. Works best with Mac M1/M2/M3 or with RTX 4090.

VS Code Plugin

Features

  • 🚀 As good as Copilot
  • ⚡️ Fast. Works well on consumer GPUs. Apple Silicon or RTX 4090 is recommended for best performance.
  • 🔐 No telemetry or tracking
  • 🔬 Works with any language coding or human one.

Recommended hardware

Minimum required RAM: 16GB is a minimum, more is better since even smallest model takes 5GB of RAM. The best way: dedicated machine with RTX 4090. Install Ollama on this machine and configure endpoint in extension settings to offload to this machine. Second best way: run on MacBook M1/M2/M3 with enougth RAM (more == better, but 10gb extra would be enougth). For windows notebooks: it runs good with decent GPU, but dedicated machine with a good GPU is recommended. Perfect if you have a dedicated gaming PC.

Local Installation

Install Ollama on local machine and then launch the extension in VSCode, everything should work as it is.

Remote Installation

Install Ollama on dedicated machine and configure endpoint to it in extension settings. Ollama usually uses port 11434 and binds to 127.0.0.1, to change it you should set OLLAMA_HOST to 0.0.0.0.

Models

Currently Llama Coder supports only Codellama. Model is quantized in different ways, but our tests shows that q4 is an optimal way to run network. When selecting model the bigger the model is, it performs better. Always pick the model with the biggest size and the biggest possible quantization for your machine. Default one is stable-code:3b-code-q4_0 and should work everywhere and outperforms most other models.

Name RAM/VRAM Notes
stable-code:3b-code-q4_0 3GB
codellama:7b-code-q4_K_M 5GB
codellama:7b-code-q6_K 6GB m
codellama:7b-code-fp16 14GB g
codellama:13b-code-q4_K_M 10GB
codellama:13b-code-q6_K 14GB m
codellama:34b-code-q4_K_M 24GB
codellama:34b-code-q6_K 32GB m
  • m - slow on MacOS
  • g - slow on older NVidia cards (pre 30xx)

Troubleshooting

Most of the problems could be seen in output of a plugin in VS Code extension output.

Changelog

[0.0.11]

  • Added Stable Code model
  • Pause download only for specific model instead of all models

[0.0.10]

  • Adding ability to pick a custom model
  • Asking user if they want to download model if it is not available

[0.0.9]

  • Adding deepseek 1b model and making it default

[0.0.8]

  • Improved DeepSeek support and language detection

[0.0.7]

  • Added DeepSeek support
  • Ability to change temperature and top p
  • Fixed some bugs

[0.0.6]

  • Fix ollama links
  • Added more models

[0.0.4]

  • Initial release of Llama Coder

More Repositories

1

supervoice-voicebox

VoiceBox neural network implementation
Jupyter Notebook
91
star
2

telegram-proxy

Simplest Docker image for starting proxy for Telegram
Go
89
star
3

supervoice-vall-e-2

VALL-E 2 reproduction
Jupyter Notebook
74
star
4

supervoice-gpt

GPT-style network for phonemization with durations of text
Jupyter Notebook
61
star
5

supervoice-hybrid

My hybrid TTS network that combines, VALL-E, VoiceBox, SpeechFlow, Seamless and TortoiseTTS into one
Jupyter Notebook
27
star
6

supervoice-enhance

Supervoice diffusion enhance
Jupyter Notebook
24
star
7

opentunnel

Encrypted tunnels to localhost
TypeScript
23
star
8

supervoice-librilight-preprocessed

60k hours of phoneme-aligned audio from audio books
Python
18
star
9

supervoice-flow

SpeechFlow neural network implementation
Jupyter Notebook
17
star
10

supervoice-gpt-facodec

GPT for FACodec
Jupyter Notebook
13
star
11

supervoice-separate

Supervoice Speaker Separation Network
Jupyter Notebook
13
star
12

bubble-firmware

Open Firmware for AI Wearables
C
11
star
13

supervoice-vocoder

Production-ready vocoder using BigVSAN
Jupyter Notebook
11
star
14

glassium

Mobile App for AI Wearables
C
9
star
15

srp6a

Secure Remote Protocol (SRP) implementation that is complaint with RFC 5054 and HomeKit Accessory Protocol
TypeScript
8
star
16

facodec

FACodec implementation, extracted from original paper and repository
Jupyter Notebook
8
star
17

supervoice-libriheavy-encodec

Compressed using encodec librilight datasets
Jupyter Notebook
7
star
18

llama-app

Accelerating AI research
TypeScript
7
star
19

superwhisper

Python
6
star
20

datasets

Declare your datasets and download them using a simple tool
Go
5
star
21

supervad

Open, performant and robust voice activity detector
Jupyter Notebook
4
star
22

transformers

🤖 My sandbox of efficient attention implementations
3
star
23

telegram-ui

Telegram UI libraries
3
star
24

supervoice-libriheavy-aligned

Aligned libriheavy using MFA
Python
3
star
25

hobby-cnc

TypeScript
2
star
26

hobby-cnc-controller

TypeScript
2
star
27

supervad-typescript

Typescript library for using SuperVAD
TypeScript
2
star
28

supervoice-librilight-hd

Re-downloading librilight in 24khz instead of default 16khz
2
star
29

react-native-fast-pbkdf2

React Native fast PBKDF2 implementation
Java
1
star
30

universal-secure-random

Universal Secure Random
TypeScript
1
star
31

teslabot

My various personal tools for Typescript development
TypeScript
1
star
32

botmate

TypeScript
1
star
33

supervoice-phonemizer

Wrapped Deep Phonemizer with support of torch hub
Jupyter Notebook
1
star
34

ton-api

TON HTTP API server
TypeScript
1
star
35

supervoice-vocoder-vocos

Yet another attempt to train a nice vocoder
Jupyter Notebook
1
star
36

superclap

CLAP tailored for voice embeddings
Jupyter Notebook
1
star
37

stevie-backend

TypeScript
1
star
38

gitchats-app

TypeScript
1
star