Skip to content

An API deployed using fastapi running LlamaCpp as the backend to do LLM inference.

License

Notifications You must be signed in to change notification settings

perpendicularai/APIinaShell

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

32 Commits
 
 
 
 
 
 
 
 

Repository files navigation

🐚 APIinaShell :octocat:

An API deployed using fastapi running LlamaCpp as the backend to do LLM inference. The purpose of the script is to abstract the complexity of deploying a LlamaCpp API instance on a global scale.

📦 Dependencies

  • llama_cpp_python
  • fastapi
  • requests
  • uvicorn

Install with pip.

🗺️ How to

To start using the api, you need to ensure that LlamaCpp is installed.

  • Download a fresh GGUF converted model from Huggingface and provide the path to it in the script.
  • In a command prompt, run:
uvicorn apiinashell:app --reload --host <ip address> --port <port number> # This should start the api server at the configured address.
  • Open a browser and browse to the address displayed in the command prompt by appending 📰 /docs. You should see the :dependabot: api interface.
  • You can then enter a string to test it out. See video :

🎥 Short Films

Llama3_test.mp4

About

An API deployed using fastapi running LlamaCpp as the backend to do LLM inference.

Topics

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages