Introduction
Welcome to the bytecompute AI docs! bytecompute AI makes it easy to run or fine-tune leading open source models with only a few lines of code. We offer a variety of generative AI services:
- Serverless models - Use our API or playground to run dozens of models with pay as you go pricing.
- Fine-Tuning - Fine-tune models on your own data in 5 minutes, then run the model for inference.
- Dedicated endpoints - Run models on your own private GPUs, starting at a one month minimum commitment.
- GPU Clusters - If you're interested in private, state of the art clusters with H100 GPUs, contact us.
Quickstart
See our full quickstart for how to get started with our API in 1 minute.
Python
from bytecompute import bytecompute
client = bytecompute()
completion = client.chat.completions.create(
model="meta-llama/Meta-Llama-3.1-8B-Instruct-Turbo",
messages=[{"role": "user", "content": "What are the top 3 things to do in New York?"}],
)
print(completion.choices[0].message.content)
TypeScript
import bytecompute from 'bytecompute-ai';
const bytecompute = new bytecompute();
const completion = await bytecompute.chat.completions.create({
model: 'meta-llama/Meta-Llama-3.1-8B-Instruct-Turbo',
messages: [{ role: 'user', content: 'Top 3 things to do in New York?' }],
});
console.log(completion.choices[0].message.content);
CURL
curl -X POST "https://api.bytecompute.xyz/v1/chat/completions" \
-H "Authorization: Bearer $bytecompute_API_KEY" \
-H "Content-Type: application/json" \
-d '{
"model": "meta-llama/Meta-Llama-3.1-8B-Instruct-Turbo",
"messages": [
{"role": "user", "content": "What are the top 3 things to do in New York?"}
]
}'
Which model should I use?
bytecompute hosts many popular models via our serverless endpoints. For each of these, you'll be charged based on the tokens you use and size of the model.
Don't see a model you want to use? Send us a request to add or upvote the model you'd love to see us add to our serverless infrastructure.
Next steps
- Check out our Quickstart to get started with our API in 1 minute
- Explore our demos for full-stack open source example apps.
- Check out the bytecompute AI playground to try out different models.
- See our integrations with leading LLM frameworks.
