Skip to content

S3 Bucket to host llama prebuilts for RPi #7312

@psiddh

Description

@psiddh

Context

S3 Bucket for hosting prebuilts for NEW Raspberry Pi ExecutorTorch Tutorials

We're developing comprehensive Raspberry Pi tutorials for deploying Llama models using ExecutorTorch. Our goal is to provide friction-free deployment by offering prebuilt ARM64 binaries that work out-of-the-box on Raspberry Pi 4/5.

This issue tracks the creation of a dedicated S3 bucket for hosting Raspberry Pi specific ExecutorTorch artifacts.

Install Journey (Quick Deploy)

On RPi devices:

# One-line install
curl -sSL https://download.pytorch.org/executorch/rpi/v1/install_rpi.sh | bash -s pi5
# Run model immediately and infer
./llama_main --model_path model.pte --tokenizer_path tokenizer.model --prompt "Hello"

Ideal S3 Bucket Structure

s3://executorch-artifacts/rpi/v1/install_rpi.sh
s3://executorch-artifacts/rpi/pi4/v1/llama_main
s3://executorch-artifacts/rpi/pi4/v1/libllama_runner.so
s3://executorch-artifacts/rpi/pi4/v1/libextension_module.so
s3://executorch-artifacts/rpi/pi5/v1/llama_main
s3://executorch-artifacts/rpi/pi5/v1/libllama_runner.so
s3://executorch-artifacts/rpi/pi5/v1/libextension_module.so

Public Access Requirements

The S3 bucket should be configured for public read access to enable direct wget/curl downloads, following the same pattern as existing PyTorch artifacts on download.pytorch.org.

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions