System Initialization

System Initialization #

Initialization API #

Initialize the Coco Server with an admin account and optional LLM configuration. This endpoint can only be called once — subsequent calls will return an error.

Parameters #

FieldTypeRequiredDescription
namestringYesServer name (becomes “{name}’s Coco Server”).
emailstringYesAdmin user email address.
passwordstringYesAdmin user password.
llm.typestringNoLLM provider type, e.g., openai, ollama.
llm.endpointstringNoLLM API endpoint URL.
llm.default_modelstringNoDefault model name, e.g., deepseek-r1.
llm.tokenstringNoAPI token for the LLM provider.
llm.reasoningbooleanNoEnable reasoning mode for the model.
languagestringNoLanguage setting: en-US or zh-CN (default: zh-CN).
//request
curl -H 'Content-Type: application/json' -XPOST http://localhost:9000/setup/_initialize -d'{
  "name": "Coco",
  "email": "hello@coco.rs",
  "password": "mypassword",
  "llm": {
    "type": "ollama",
    "endpoint": "http://localhost:11434",
    "default_model": "deepseek-r1"
  },
  "language": "en-US"
}'

//response
{
  "status": "ok"
}

Initialization UI Management #

When entering the coco server, it will check whether it has been initialized. If not, it will enter the initialization page.

Create a user account #

Set up a new user account to manage access and permissions.

initialization step 1

Connect to a Large Model #

After integrating a large model, you will unlock the AI chat feature, providing intelligent search and an efficient work assistant.

initialization step 2

You can also click Set Up Later to skip and configure it later after entering the system.

Edit Edit this page