Quickstart > Inference This folder contains scripts to get you started with inference on Meta Llama models. contains scripts for tasks relating to code generation using CodeLlama contsin scripts to do memory efficient inference on servers and local machines has scripts using MLC to serve Llama on Android (h/t to OctoAI for the contribution!)