meta / meta-llama-3-70b-instruct

A 70 billion parameter language model from Meta, fine tuned for chat completions

ℹ️
Language model training is a beta feature.
We’re still working out the kinks. If you run into any issues, please hop in our Discord and let us know. Keep in mind that we might make breaking changes to the API as we improve the training experience.

If you haven’t yet trained a model on Replicate, we recommend you read one of the following guides.

Pricing

Trainings for this model run on 8x Nvidia H100 GPU hardware.

Create a training

Install the Python library:

pip install replicate

Then, run this to create a training with replicate-internal/llama-3-70b-instruct-fp16-8xh100-triton:aa825bb1 as the base model:

import replicate

training = replicate.trainings.create(
  version="replicate-internal/llama-3-70b-instruct-fp16-8xh100-triton:aa825bb175b682c7e92c7299ec319f6cdae0cbd7803dd74a6e4f31d71512ab54",
  input={
    ...
  },
  destination=f"{username}/<destination-model-name>"
)

print(training)
curl -s -X POST \
-d '{"destination": "{username}/<destination-model-name>", "input": {...}}' \
  -H "Authorization: Bearer $REPLICATE_API_TOKEN" \
  https://api.replicate.com/v1/models/meta/meta-llama-3-70b-instruct/versions/aa825bb175b682c7e92c7299ec319f6cdae0cbd7803dd74a6e4f31d71512ab54/trainings

The API response will look like this:

{
  "id": "zz4ibbonubfz7carwiefibzgga",
  "version": "aa825bb175b682c7e92c7299ec319f6cdae0cbd7803dd74a6e4f31d71512ab54",
  "status": "starting",
  "input": {
    "data": "..."
  },
  "output": null,
  "error": null,
  "logs": null,
  "started_at": null,
  "created_at": "2023-03-28T21:47:58.566434Z",
  "completed_at": null
}

Note that before you can create a training, you’ll need to create a model and use its name as the value for the destination field.