meta-llama

Llama 3.1 8B Instruct

meta-llama/llama-3.1-8b-instruct

Input$0.060 /1M tokens
·
Output$0.060 /1M tokens

Code Examples

import requests

response = requests.post(
    "https://neurongate.net/v1/chat/completions",
    headers={
        "Authorization": "Bearer ng-your-api-key",
        "Content-Type": "application/json"
    },
    json={
        "model": "meta-llama/llama-3.1-8b-instruct",
        "messages": [
            {"role": "user", "content": "Hello!"}
        ]
    }
)

print(response.json()["choices"][0]["message"]["content"])

Pricing Details

ExampleCost
1K input tokens (short prompt)< $0.0001
1K in + 500 out (typical response)< $0.0001
10K in + 2K out (document analysis)$0.00072
100K in + 10K out (large context)$0.00660

Prices in USD. Billed per actual token usage. Prepay with USDT, USDC, ETH, or BTC.

Frequently Asked Questions

Capabilities

Streaming
Yes
Function Calling / Tools
Yes
Vision (Image Input)
No
Audio Processing
No

Context Window

131K tokens

~98K words of text

Max output
4K tokens

Ready to build?

Pay with crypto. No subscriptions.

Get Started Free

Similar Models