Skip to content

SeaTable AI Integration

6.0

SeaTable AI is a SeaTable extension that integrates AI functionality into SeaTable. Deploying SeaTable AI allows users to execute AI-based automation steps within SeaTable.

At the time of writing, the following types of automation steps are supported:

  • Summarize
  • Classify
  • OCR (Optical character recognition)
  • Extract
  • Custom for individual use cases

Deployment

SeaTable AI requires SeaTable 6.0

The easiest way to deploy SeaTable AI is to deploy it on the same host as SeaTable Server. A standalone deployment (on a separate host or virtual machine) is explained here.

Amend the .env file

To install SeaTable AI, include seatable-ai.yml in the COMPOSE_FILE variable within your .env file. This instructs Docker-Compose to include the seatable-ai service.

Simply copy and paste () the following code into your command line:

sed -i "s/COMPOSE_FILE='\(.*\)'/COMPOSE_FILE='\1,seatable-ai.yml'/" /opt/seatable-compose/.env

Then add SeaTable AI server configurations in .env:

ENABLE_SEATABLE_AI=true
SEATABLE_AI_SERVER_URL=http://seatable-ai:8888

LLM Provider Configuration

SeaTable AI will use AI functions in conjunction with a Large Language Model (LLM) service.

Supported LLM Providers

SeaTable AI supports a wide variety of LLM providers through LiteLLM as well as any LLM services with OpenAI-compatible endpoints. Please refer to LiteLLM's documentation in case you run into issues while trying to use a specific provider.

Model Selection

In order to ensure the efficient use of SeaTable AI features, you need to select a large, multimodal model. This requires the chosen model to support image input and recognition (e.g. for running OCR as part of automations).

The following section showcases the required configuration settings for the most popular hosted LLM services. These must be configured inside your .env file:

SEATABLE_AI_LLM_TYPE=openai
SEATABLE_AI_LLM_KEY=<your openai LLM access key>
SEATABLE_AI_LLM_MODEL=gpt-4o-mini # recommended
SEATABLE_AI_LLM_TYPE=deepseek
SEATABLE_AI_LLM_KEY=<your LLM access key>
SEATABLE_AI_LLM_MODEL=deepseek-chat # recommended
SEATABLE_AI_LLM_TYPE=azure
SEATABLE_AI_LLM_URL= # your deployment url, leave blank to use default endpoint
SEATABLE_AI_LLM_KEY=<your API key>
SEATABLE_AI_LLM_MODEL=<your deployment name>
SEATABLE_AI_LLM_TYPE=ollama_chat
SEATABLE_AI_LLM_URL=<your LLM endpoint>
SEATABLE_AI_LLM_KEY=<your LLM access key>
SEATABLE_AI_LLM_MODEL=<your model-id>
SEATABLE_AI_LLM_TYPE=huggingface
SEATABLE_AI_LLM_URL=<your huggingface API endpoint>
SEATABLE_AI_LLM_KEY=<your huggingface API key>
SEATABLE_AI_LLM_MODEL=<model provider>/<model-id>
SEATABLE_AI_LLM_TYPE=proxy
SEATABLE_AI_LLM_URL=<your proxy url>
SEATABLE_AI_LLM_KEY=<your proxy virtual key> # optional
SEATABLE_AI_LLM_MODEL=<model-id>

If you are using an LLM service with OpenAI-compatible endpoints, you should set SEATABLE_AI_LLM_TYPE to other or openai, and set other LLM configuration settings as necessary:

SEATABLE_AI_LLM_TYPE=...
SEATABLE_AI_LLM_URL=...
SEATABLE_AI_LLM_KEY=...
SEATABLE_AI_LLM_MODEL=...

Download SeaTable AI image and restart

One more step is necessary to download the SeaTable AI image and restart the SeaTable service:

cd /opt/seatable-compose
docker compose up -d

Now SeaTable AI can be used.