Using Groq
Learn how to obtain a GroqCloud API Key and configure Groq in LobeVidol to experience the powerful performance of Groq.
Using Groq in LobeVidol
Groq’s LPU inference engine has excelled in the latest independent large language model (LLM) benchmark tests, redefining the standards for AI solutions with its astonishing speed and efficiency. With the integration of LobeVidol and Groq Cloud, you can now easily leverage Groq’s technology to accelerate the performance of large language models in LobeVidol.
The Groq LPU inference engine has consistently achieved a speed of 300 tokens per second in internal benchmark tests, confirmed by benchmarks from ArtificialAnalysis.ai, showing that Groq outperforms other providers in throughput (241 tokens per second) and total time to receive 100 output tokens (0.8 seconds).
This document will guide you on how to use Groq in LobeVidol:
Step 1: Obtain a GroqCloud API Key
First, you need to obtain an API Key from the GroqCloud Console.
Create an API Key in the API Keys
menu of the console.
Safeguard the key shown in the pop-up; it will only appear once. If you accidentally lose it, you will need to create a new key.
Step 2: Configure Groq in LobeVidol
You can find the configuration options for Groq in Settings
-> Language Models
, where you can enter the API Key you just obtained.
Next, select a Groq-supported model in the assistant’s model options, and you can experience the powerful performance of Groq in LobeVidol.