Machine Learning as a Service
We run several public workers that provide machine learning inference/training as a service. Under the hood each worker may run different backend, to support different types of machine learning models. We currently run the following backend:
- Triteia (for large generative transformer models.
- DeepSpeed-MII (for text-to-image and some other models).
- Inferencia (for other HuggingFace models unsupported).
The endpoint for our public workers starts with https://api.research.computer/
. For example, the endpoint for Triteia is https://api.research.computer/triteia/
.
Triteia
Triteia is an inference engine that supports OpenAI-compatible apis for large generative transformer models. Simply replace the endpoint with https://api.research.computer/triteia/
to use Triteia.
Inferencia
The expected output is