Open LLaMA
Run OpenLLaMA in a GPU environment with a single command. 📡
Speed Run
- Signup for Beam
- Download the CLI and Python SDK
- Clone this template locally:
beam create-app openllama
- Spin up a GPU environment to run inference:
beam start app.py
- Deploy the app as a web API:
beam deploy app.py