Text generation is supported on the network. As explained in the architecture page, inference takes place in virtual machines.
We are providing multiple VMs, with an inference stack that can change with time. It means the API is subject to change on newer models.
|API Base Url
|Mixtral Instruct 8x7B MoE
|ChatML or Alpaca Instruct
|DeepSeek Coder 6.7B
Please see the according API documentation based on the model of your choice:
Each mode has its own formatting. Knowing which format you should provide for a specific model will help getting better results out of it. Please refer to the available models table to know which format is the best for your model.