system
user
✕
Add message
Settings
0 tokens · Latency 0ms
View Code
Run
⌘
↵
Models
25
meta-llama
Meta-Llama-3.1-8B-Instruct
Model page
Temperature
Max Tokens
Top-P
Streaming
API Quota
Free
76%
Give feedback