system
user
✕
Add message
Settings
0 tokens · Latency 0ms
View Code
Run
⌘
↵
Models
19
meta-llama
Meta-Llama-3.1-70B-Instruct
Temperature
Max Tokens
Top-P
Streaming
API Quota
Free
76%