If the AI is busy, your prompt will processed when a slot frees up.
Oobabooga WebUI: {{wait_time}}
Model: {{model_name}}
Serina, Anytime, Anywhere!
Server settings : {{server_settings}}
{
"uptime": {{uptime}},
"endpoints": {
"Oobabooga WebUI": "{{server_link}}"
},
"proompts": "{{proompts}}",
"input tookens": "{{input_tookens}}",
"output tookens": "{{output_tookens}}",
"proomptersNow": molu,
"openaiKeys": 0,
"anthropicKeys": 0,
"openaiOrgs": 0,
"Oobabooga WebUI": {
"usage": "{{tookens}} tokens",
"activeKeys": 1,
"trialKeys": 0,
"revokedKeys": 0,
"overQuotaKeys": 0,
"proomptersInQueue": 0,
"estimatedQueueTime": "molu"
},
"config": {
"gatekeeper": "none",
"modelRateLimit": "none",
"maxContextTokensOobabooga": "8192",
"maxContextTokensOpenAI": "0",
"maxContextTokensAnthropic": "0",
"maxOutputTokensOobabooga": "0",
"maxOutputTokensOpenAI": "1000",
"maxOutputTokensAnthropic": "400",
"allowedModelFamilies": "Oobabooga",
"rejectDisallowed": "false",
"rejectMessage": "This content violates Arcalive AiChat Channel's acceptable use policy.",
"promptLogging": "true",
"tokenQuota": {
"Oobabooga WebUI": "1",
"turbo": "0",
"gpt4": "0",
"gpt4-32k": "0",
"claude": "0"
}
},
"build": "675d4b (main@ihbs02/oai-reverse-proxy)"
}