Vultr Serverless Inference allows you to run inference workloads for large language models such as Mixtral 8x7B, Mistral 7B, Meta Llama 2 70B, and more. Using Vultr Serverless Inference, you can run inference workloads without having to worry about the infrastructure, and you only pay for the input and output tokens. This guide demonstrates step-by-step process to start using Vultr Serverless Inference in Node.js. Before you begin, you must: * [Create a Vultr Serverless Inference Subscription](h......