Connect Slack to your preprocessing pipeline, and batch process all your documents using unstructured-ingest to store structured outputs locally on your filesystem.

First, install the Slack dependencies as shown here.

pip install "unstructured-ingest[slack]"

To authenticate the Slack source connector provide the following:

  • token: Bot token used to access Slack API, must have channels:history scope for the bot user.

Provide a comma-delimited list of Slack channel IDs to pull messages from, these can be a public or private channels. Your bot user must be in the channels for them to be ingested.

Optionally, you can specify:

  • start-date: Start date/time in formats YYYY-MM-DD, YYYY-MM-DDTHH:MM:SS, YYYY-MM-DD+HH:MM:SS, or YYYY-MM-DDTHH:MM:SStz
  • end-date: End date/time in formats YYYY-MM-DD, YYYY-MM-DDTHH:MM:SS, YYYY-MM-DD+HH:MM:SS, or YYYY-MM-DDTHH:MM:SStz

Make sure to set the --partition-by-api flag and pass in your API key with --api-key:

Additionally, if you’re using Unstructured Serverless API, your locally deployed Unstructured API, or an Unstructured API deployed on Azure or AWS, you also need to specify the API URL via the --partition-endpoint argument.