You're looking at a specific version of this model. Jump to the model overview.

nwhitehead /llama2-7b-chat-gptq:8c1f632f

Input schema

The fields you can use to run this model with an API. If you don’t give a value for a field its default value will be used.

Field Type Default value Description
prompt
string
USER: Hello, who are you? ASSISTANT:
Text prompt for the model
temperature
number
0.5

Min: 0.01

Max: 2

Temperature of the output, it's best to keep it below 1
top_p
number
1

Min: 0.01

Max: 1

Top cumulative probability to filter candidates
top_k
integer
20

Min: 1

Max: 100

Number of top candidates to keep
repetition_penalty
number
1

Min: 1

Max: 1.5

Penalty for repeated tokens in the model's output
max_tokens
integer
50

Min: 1

Max: 4096

Maximum tokens to generate
min_tokens
integer
1

Max: 4096

Minimum tokens to generate
seed
integer
-1

Min: -2147483648

Max: 2147483647

Seed for reproducibility, -1 for random seed

Output schema

The shape of the response you’ll get when you run this model with an API.

Schema
{'items': {'type': 'string'},
 'title': 'Output',
 'type': 'array',
 'x-cog-array-display': 'concatenate',
 'x-cog-array-type': 'iterator'}