Skip to content

Commit

Permalink
Merge branch 'master' into feat-request-middleware
Browse files Browse the repository at this point in the history
  • Loading branch information
dave-gray101 authored Dec 19, 2024
2 parents 4c3cad7 + 3be9a08 commit 095dd60
Show file tree
Hide file tree
Showing 22 changed files with 112 additions and 43 deletions.
9 changes: 9 additions & 0 deletions .env
Original file line number Diff line number Diff line change
Expand Up @@ -82,6 +82,15 @@
# Enable to allow p2p mode
# LOCALAI_P2P=true

# Enable to use federated mode
# LOCALAI_FEDERATED=true

# Enable to start federation server
# FEDERATED_SERVER=true

# Define to use federation token
# TOKEN=""

### Watchdog settings
###
# Enables watchdog to kill backends that are inactive for too much time
Expand Down
2 changes: 1 addition & 1 deletion Makefile
Original file line number Diff line number Diff line change
Expand Up @@ -8,7 +8,7 @@ DETECT_LIBS?=true
# llama.cpp versions
GOLLAMA_REPO?=https://github.com/go-skynet/go-llama.cpp
GOLLAMA_VERSION?=2b57a8ae43e4699d3dc5d1496a1ccd42922993be
CPPLLAMA_VERSION?=081b29bd2a3d91e7772e3910ce223dd63b8d7d26
CPPLLAMA_VERSION?=cd920d0ac38ec243605a5a57c50941140a193f9e

# whisper.cpp version
WHISPER_REPO?=https://github.com/ggerganov/whisper.cpp
Expand Down
5 changes: 3 additions & 2 deletions backend/python/autogptq/requirements-intel.txt
Original file line number Diff line number Diff line change
@@ -1,5 +1,6 @@
--extra-index-url https://pytorch-extension.intel.com/release-whl/stable/xpu/us/
intel-extension-for-pytorch
torch
intel-extension-for-pytorch==2.3.110+xpu
torch==2.3.1+cxx11.abi
oneccl_bind_pt==2.3.100+xpu
optimum[openvino]
setuptools
7 changes: 4 additions & 3 deletions backend/python/bark/requirements-intel.txt
Original file line number Diff line number Diff line change
@@ -1,7 +1,8 @@
--extra-index-url https://pytorch-extension.intel.com/release-whl/stable/xpu/us/
intel-extension-for-pytorch
torch
torchaudio
intel-extension-for-pytorch==2.3.110+xpu
torch==2.3.1+cxx11.abi
torchaudio==2.3.1+cxx11.abi
oneccl_bind_pt==2.3.100+xpu
optimum[openvino]
setuptools
transformers
Expand Down
5 changes: 3 additions & 2 deletions backend/python/common/template/requirements-intel.txt
Original file line number Diff line number Diff line change
@@ -1,4 +1,5 @@
--extra-index-url https://pytorch-extension.intel.com/release-whl/stable/xpu/us/
intel-extension-for-pytorch
torch
intel-extension-for-pytorch==2.3.110+xpu
torch==2.3.1+cxx11.abi
oneccl_bind_pt==2.3.100+xpu
optimum[openvino]
7 changes: 4 additions & 3 deletions backend/python/coqui/requirements-intel.txt
Original file line number Diff line number Diff line change
@@ -1,7 +1,8 @@
--extra-index-url https://pytorch-extension.intel.com/release-whl/stable/xpu/us/
intel-extension-for-pytorch
torch
torchaudio
intel-extension-for-pytorch==2.3.110+xpu
torch==2.3.1+cxx11.abi
torchaudio==2.3.1+cxx11.abi
oneccl_bind_pt==2.3.100+xpu
optimum[openvino]
setuptools
transformers
Expand Down
7 changes: 4 additions & 3 deletions backend/python/diffusers/requirements-intel.txt
Original file line number Diff line number Diff line change
@@ -1,7 +1,8 @@
--extra-index-url https://pytorch-extension.intel.com/release-whl/stable/xpu/us/
intel-extension-for-pytorch
torch
torchvision
intel-extension-for-pytorch==2.3.110+xpu
torch==2.3.1+cxx11.abi
torchvision==0.18.1+cxx11.abi
oneccl_bind_pt==2.3.100+xpu
optimum[openvino]
setuptools
diffusers
Expand Down
7 changes: 6 additions & 1 deletion backend/python/openvoice/requirements-cpu.txt
Original file line number Diff line number Diff line change
@@ -1,3 +1,8 @@
torch==2.4.1
git+https://github.com/myshell-ai/MeloTTS.git
git+https://github.com/myshell-ai/OpenVoice.git
git+https://github.com/myshell-ai/OpenVoice.git
whisper-timestamped
numpy==1.22.0
pydub==0.25.1
wavmark==0.0.3
eng_to_ipa==0.0.2
7 changes: 6 additions & 1 deletion backend/python/openvoice/requirements-cublas11.txt
Original file line number Diff line number Diff line change
@@ -1,4 +1,9 @@
--extra-index-url https://download.pytorch.org/whl/cu118
torch==2.4.1+cu118
git+https://github.com/myshell-ai/MeloTTS.git
git+https://github.com/myshell-ai/OpenVoice.git
git+https://github.com/myshell-ai/OpenVoice.git
whisper-timestamped
numpy==1.22.0
pydub==0.25.1
wavmark==0.0.3
eng_to_ipa==0.0.2
7 changes: 6 additions & 1 deletion backend/python/openvoice/requirements-cublas12.txt
Original file line number Diff line number Diff line change
@@ -1,3 +1,8 @@
torch==2.4.1
git+https://github.com/myshell-ai/MeloTTS.git
git+https://github.com/myshell-ai/OpenVoice.git
git+https://github.com/myshell-ai/OpenVoice.git
whisper-timestamped
numpy==1.22.0
pydub==0.25.1
wavmark==0.0.3
eng_to_ipa==0.0.2
7 changes: 6 additions & 1 deletion backend/python/openvoice/requirements-hipblas.txt
Original file line number Diff line number Diff line change
@@ -1,4 +1,9 @@
--extra-index-url https://download.pytorch.org/whl/rocm6.0
torch==2.4.1+rocm6.0
git+https://github.com/myshell-ai/MeloTTS.git
git+https://github.com/myshell-ai/OpenVoice.git
git+https://github.com/myshell-ai/OpenVoice.git
whisper-timestamped
numpy==1.22.0
pydub==0.25.1
wavmark==0.0.3
eng_to_ipa==0.0.2
6 changes: 4 additions & 2 deletions backend/python/openvoice/requirements-intel.txt
Original file line number Diff line number Diff line change
@@ -1,6 +1,8 @@
--extra-index-url https://pytorch-extension.intel.com/release-whl/stable/xpu/us/
intel-extension-for-pytorch
torch
intel-extension-for-pytorch==2.3.110+xpu
torch==2.3.1+cxx11.abi
torchaudio==2.3.1+cxx11.abi
oneccl_bind_pt==2.3.100+xpu
optimum[openvino]
grpcio==1.68.1
protobuf
Expand Down
5 changes: 0 additions & 5 deletions backend/python/openvoice/requirements.txt
Original file line number Diff line number Diff line change
Expand Up @@ -2,13 +2,8 @@ grpcio==1.68.1
protobuf
librosa
faster-whisper
pydub==0.25.1
wavmark==0.0.3
numpy==1.22.0
eng_to_ipa==0.0.2
inflect
unidecode
whisper-timestamped
openai
python-dotenv
pypinyin
Expand Down
7 changes: 4 additions & 3 deletions backend/python/parler-tts/requirements-intel.txt
Original file line number Diff line number Diff line change
@@ -1,7 +1,8 @@
--extra-index-url https://pytorch-extension.intel.com/release-whl/stable/xpu/us/
intel-extension-for-pytorch
torch
torchaudio
intel-extension-for-pytorch==2.3.110+xpu
torch==2.3.1+cxx11.abi
torchaudio==2.3.1+cxx11.abi
oneccl_bind_pt==2.3.100+xpu
optimum[openvino]
transformers
accelerate
5 changes: 3 additions & 2 deletions backend/python/rerankers/requirements-intel.txt
Original file line number Diff line number Diff line change
@@ -1,8 +1,9 @@
--extra-index-url https://pytorch-extension.intel.com/release-whl/stable/xpu/us/
intel-extension-for-pytorch
intel-extension-for-pytorch==2.3.110+xpu
transformers
accelerate
torch
torch==2.3.1+cxx11.abi
oneccl_bind_pt==2.3.100+xpu
rerankers[transformers]
optimum[openvino]
setuptools
5 changes: 3 additions & 2 deletions backend/python/sentencetransformers/requirements-intel.txt
Original file line number Diff line number Diff line change
@@ -1,6 +1,7 @@
--extra-index-url https://pytorch-extension.intel.com/release-whl/stable/xpu/us/
intel-extension-for-pytorch
torch
intel-extension-for-pytorch==2.3.110+xpu
torch==2.3.1+cxx11.abi
oneccl_bind_pt==2.3.100+xpu
optimum[openvino]
setuptools
accelerate
Expand Down
5 changes: 3 additions & 2 deletions backend/python/transformers-musicgen/requirements-intel.txt
Original file line number Diff line number Diff line change
@@ -1,7 +1,8 @@
--extra-index-url https://pytorch-extension.intel.com/release-whl/stable/xpu/us/
intel-extension-for-pytorch
intel-extension-for-pytorch==2.3.110+xpu
transformers
oneccl_bind_pt==2.3.100+xpu
accelerate
torch
torch==2.3.1+cxx11.abi
optimum[openvino]
setuptools
5 changes: 3 additions & 2 deletions backend/python/transformers/requirements-intel.txt
Original file line number Diff line number Diff line change
@@ -1,6 +1,7 @@
--extra-index-url https://pytorch-extension.intel.com/release-whl/stable/xpu/us/
intel-extension-for-pytorch
torch
intel-extension-for-pytorch==2.3.110+xpu
torch==2.3.1+cxx11.abi
oneccl_bind_pt==2.3.100+xpu
optimum[openvino]
intel-extension-for-transformers
bitsandbytes
9 changes: 5 additions & 4 deletions backend/python/vall-e-x/requirements-intel.txt
Original file line number Diff line number Diff line change
@@ -1,6 +1,7 @@
--extra-index-url https://pytorch-extension.intel.com/release-whl/stable/xpu/us/
intel-extension-for-pytorch
intel-extension-for-pytorch==2.3.110+xpu
accelerate
torch
torchaudio
optimum[openvino]
torch==2.3.1+cxx11.abi
torchaudio==2.3.1+cxx11.abi
optimum[openvino]
oneccl_bind_pt==2.3.100+xpu
7 changes: 4 additions & 3 deletions backend/python/vllm/requirements-intel.txt
Original file line number Diff line number Diff line change
@@ -1,8 +1,9 @@
--extra-index-url https://pytorch-extension.intel.com/release-whl/stable/xpu/us/
intel-extension-for-pytorch
intel-extension-for-pytorch==2.3.110+xpu
accelerate
torch
torch==2.3.1+cxx11.abi
transformers
optimum[openvino]
setuptools
bitsandbytes
bitsandbytes
oneccl_bind_pt==2.3.100+xpu
3 changes: 3 additions & 0 deletions docs/content/docs/features/distributed_inferencing.md
Original file line number Diff line number Diff line change
Expand Up @@ -129,6 +129,9 @@ There are options that can be tweaked or parameters that can be set using enviro

| Environment Variable | Description |
|----------------------|-------------|
| **LOCALAI_P2P** | Set to "true" to enable p2p |
| **LOCALAI_FEDERATED** | Set to "true" to enable federated mode |
| **FEDERATED_SERVER** | Set to "true" to enable federated server |
| **LOCALAI_P2P_DISABLE_DHT** | Set to "true" to disable DHT and enable p2p layer to be local only (mDNS) |
| **LOCALAI_P2P_ENABLE_LIMITS** | Set to "true" to enable connection limits and resources management (useful when running with poor connectivity or want to limit resources consumption) |
| **LOCALAI_P2P_LISTEN_MADDRS** | Set to comma separated list of multiaddresses to override default libp2p 0.0.0.0 multiaddresses |
Expand Down
28 changes: 28 additions & 0 deletions gallery/index.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -869,6 +869,34 @@
- filename: FuseChat-Llama-3.2-3B-Instruct-Q4_K_M.gguf
sha256: a4f0e9a905b74886b79b72622c06a3219d6812818a564a53c39fc49032d7f842
uri: huggingface://bartowski/FuseChat-Llama-3.2-3B-Instruct-GGUF/FuseChat-Llama-3.2-3B-Instruct-Q4_K_M.gguf
- !!merge <<: *llama32
name: "llama-song-stream-3b-instruct"
urls:
- https://huggingface.co/prithivMLmods/Llama-Song-Stream-3B-Instruct
- https://huggingface.co/bartowski/Llama-Song-Stream-3B-Instruct-GGUF
description: |
The Llama-Song-Stream-3B-Instruct is a fine-tuned language model specializing in generating music-related text, such as song lyrics, compositions, and musical thoughts. Built upon the meta-llama/Llama-3.2-3B-Instruct base, it has been trained with a custom dataset focused on song lyrics and music compositions to produce context-aware, creative, and stylized music output.
overrides:
parameters:
model: Llama-Song-Stream-3B-Instruct-Q4_K_M.gguf
files:
- filename: Llama-Song-Stream-3B-Instruct-Q4_K_M.gguf
sha256: 1a0f0aef16e5be46be827c430cbe5ce4b2915b2f4a1dd60b98792004d39b9f52
uri: huggingface://bartowski/Llama-Song-Stream-3B-Instruct-GGUF/Llama-Song-Stream-3B-Instruct-Q4_K_M.gguf
- !!merge <<: *llama32
name: "llama-chat-summary-3.2-3b"
urls:
- https://huggingface.co/prithivMLmods/Llama-Chat-Summary-3.2-3B
- https://huggingface.co/bartowski/Llama-Chat-Summary-3.2-3B-GGUF
description: |
Llama-Chat-Summary-3.2-3B is a fine-tuned model designed for generating context-aware summaries of long conversational or text-based inputs. Built on the meta-llama/Llama-3.2-3B-Instruct foundation, this model is optimized to process structured and unstructured conversational data for summarization tasks.
overrides:
parameters:
model: Llama-Chat-Summary-3.2-3B-Q4_K_M.gguf
files:
- filename: Llama-Chat-Summary-3.2-3B-Q4_K_M.gguf
sha256: ed1be20d2374aa6db9940923f41fa229bd7ebe13d41b1ff1ff18a6f87e99df79
uri: huggingface://bartowski/Llama-Chat-Summary-3.2-3B-GGUF/Llama-Chat-Summary-3.2-3B-Q4_K_M.gguf
- &qwen25
## Qwen2.5
name: "qwen2.5-14b-instruct"
Expand Down

0 comments on commit 095dd60

Please sign in to comment.