pytorch-http/README.md

395 B

Open LLaMA

Run OpenLLaMA in a GPU environment with a single command. 📡

Speed Run

  1. Signup for Beam
  2. Download the CLI and Python SDK
  3. Clone this template locally: beam create-app openllama
  4. Spin up a GPU environment to run inference: beam start app.py
  5. Deploy the app as a web API: beam deploy app.py