A Streamlit app for interacting with LLAMACPP-based GPT models locally in llama_cpp .
Download or convert safetensors
language models from hugging face to gguf format.
Tested Phi3 model on i5 8th gen CPU with output of 12 token/s.
LocalGPT is a Streamlit app that enables users to interact with LLAMACPP GPT models locally, without the need for an internet connection or remote server. It provides a user-friendly interface for generating text (streaming), and exploring the capabilities of the model.
To install the LocalGPT Streamlit app, follow these steps:
- Clone the repository:
git clone https://github.com/wambugu71/OfflineGPT-
- Navigate to the project directory:
cd OfflineGPT-
3.Install the required packages:
pip install -r requirements.txt
- (Optional) If you plan to use a GPU for acceleration (nvidia,intel,amd) gpu's install the GPU-specific packages:
- Run your model as a server from the terminal.
python -m llama_cpp.server --model <llama_cpp model name>.gguf
To run the LocalGPT Streamlit app, simply execute the following command in the project directory:
cd src
streamlit run app.py
This will launch the app in your default web browser. You can then interact with the app (after running the server) by providing input text, and generating text.
LocalGPT allows for customization of the app to suit your specific needs. You can modify the app.py
file to adjust the generation settings, or add additional functionality. Refer to the Streamlit documentation for more information on customizing Streamlit apps.
Contributions are welcome! If you have suggestions, improvements, or bug fixes, feel free to create a pull request or open an issue.
LocalGPT is licensed under the MIT License. See the LICENSE file for more information.