Skip to content

UFResearchComputing/Llama3_on_HPG

Repository files navigation

Llama3_on_HPG

This tutorial is adopted from github repositories meta-llama/llama and meta-llama/llama-recipes.

Llama3 Introduction

Llama 3, Meta’s latest family of open-source large language models (LLMs). Llama 3 comes in two sizes: 8B and 70B parameters. Each size has both base (pre-trained) and instruct-tuned versions. The context length for all variants is 8K tokens.

  • Base Models:
  • Meta-Llama-3-8b: The base 8B model.
  • Meta-Llama-3-70b: The base 70B model.
  • Fine-Tuned Versions:
  • Meta-Llama-3-8b-instruct: Instruct fine-tuned version of the base 8B model.
  • Meta-Llama-3-70b-instruct: Instruct fine-tuned version of the base 70B model.
  • Llama Guard 2:
  • Llama Guard 2, designed for production use cases, classifies LLM inputs and responses to detect unsafe content. It was fine-tuned on Llama 3 8B.

Llama 3 uses a new tokenizer with an expanded vocabulary size of 128,256 (compared to 32K tokens in Llama 2). This larger vocabulary improves text encoding efficiency and potentially enhances multilingualism. Grouped-Query Attention (GQA): The 8B version of Llama 3 now uses GQA, an efficient representation that helps with longer contexts. Availability: Llama 3 models are freely available for research and commercial purposes.

For more details, you can check out the official Llama 3 blog post or this article.

Download Llama 3

In order to download the model weights and tokenizer, please visit the Meta Llama website and accept our License.

Once your request is approved, you will receive a signed URL over email. Then run the download.sh script, passing the URL provided when prompted to start the download.

Pre-requisites: Make sure you have wget and md5sum installed. Then run the script: ./download.sh.

Keep in mind that the links expire after 24 hours and a certain amount of downloads. If you start seeing errors such as 403: Forbidden, you can always re-request a link.

Access to Hugging Face

We are also providing downloads on Hugging Face, in both transformers and native llama3 formats. To download the weights from Hugging Face, please follow these steps:

  • Visit one of the repos, for example meta-llama/Meta-Llama-3-8B-Instruct.
  • Read and accept the license. Once your request is approved, you'll be granted access to all the Llama 3 models. Note that requests use to take up to one hour to get processed.
  • To download the original native weights to use with this repo, click on the "Files and versions" tab and download the contents of the original folder. You can also download them from the command line if you pip install huggingface-hub:
huggingface-cli download meta-llama/Meta-Llama-3-8B-Instruct --include "original/*" --local-dir meta-llama/Meta-Llama-3-8B-Instruct
  • To use with transformers, the following pipeline snippet will download and cache the weights:

    import transformers
    import torch
    
    model_id = "meta-llama/Meta-Llama-3-8B-Instruct"
    
    pipeline = transformers.pipeline(
      "text-generation",
      model="meta-llama/Meta-Llama-3-8B-Instruct",
      model_kwargs={"torch_dtype": torch.bfloat16},
      device="cuda",
    )

Quick Start Llama3 on HiPerGator (HPG)

You can follow the steps below to quickly get up and running with Llama 3 models. These steps will let you run quick inference on HiPerGator. All the LLama3 models have already been downloaded to /data/ai/models/nlp/llama/models_llama3. You can run the model on HPG via the command line using ml nlp/1.3, or in the Jupyter Notebook with the nlp-1.3 kernels.

Finetuning and Inference

  • If you insteaed in finetune Llama 3 on single-GPU and multi-GPU setups, you can find recipes at finetuning.
  • If you want to deploy Llama3 for inference locally and using model servers, you can find recipes at inference.

For more examples, see the Llama recipes repository.

License

All rights are reserved by the Meta Llama team Please refer to the Meta Llama License file and the Acceptable Use Policy.

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published