Dockerized huggingface model with HTTP API
Go to file Use this template
2023-09-18 12:59:41 -05:00
__main__.py feat: initial commit 2023-09-18 12:59:41 -05:00
.dockerignore feat: initial commit 2023-09-18 12:59:41 -05:00
.gitignore feat: initial commit 2023-09-18 12:59:41 -05:00
.tool-versions feat: initial commit 2023-09-18 12:59:41 -05:00
Dockerfile feat: initial commit 2023-09-18 12:59:41 -05:00
README.md feat: initial commit 2023-09-18 12:59:41 -05:00
requirements.txt feat: initial commit 2023-09-18 12:59:41 -05:00

Open LLaMA

Run OpenLLaMA in a GPU environment with a single command. 📡

Speed Run

  1. Signup for Beam
  2. Download the CLI and Python SDK
  3. Clone this template locally: beam create-app openllama
  4. Spin up a GPU environment to run inference: beam start app.py
  5. Deploy the app as a web API: beam deploy app.py