One-stop service platform
designed for large-scale model fine-tuning and deploying. Through the platform, users can quickly and seamlessly deploy custom models as services and fine-tune them based on the data uploaded.
Easy to use
from openai import OpenAI
client = OpenAI(api_key="YOUR_API_KEY", base_url="https://api.siliconflow.cn/v1")
response = client.chat.completions.create(
model='alibaba/Qwen1.5-110B-Chat',
messages=[
{'role': 'user', 'content': "抛砖引玉是什么意思呀"}
],
stream=True
)
for chunk in response:
print(chunk.choices[0].delta.content)
Model Inference
With just a single line of code, developers can seamlessly integrate the fastest model services from SiliconCloud.
Model Deploy
·
Upload your workflow and Download the callable Model Service API.
·
Reduce the chances of application downtime with auto scaling.
·
Accelerate your workflow as needed.