This page contains two quickstarts to help you get started with the Unstructured user interface (UI).

  • The local quickstart is an ideal starting point and requires no prior setup. It shows you how to use the Unstructured UI to process a single file that is stored on your local machine. This approach is ideal for rapid testing and prototyping of Unstructured ETL+ workflows, across the full range of Unstructured features, against a single representative file.
  • The remote quickstart takes a bit longer, but it shows you how to use the UI to set up and run Unstructured ETL+ workflows against files and data stored in remote source file and object stores, websites, databases, and vector stores. Unstructured delivers the resulting processed data to remote destination file and object stores, databases, and vector stores. This approach is ideal for production use cases, where you want to process large volumes of files and data in a scalable and efficient manner.

If you’re not sure which quickstart to use, we recommend starting with the local quickstart.

The local quickstart is a fast and easy way to learn about Unstructured. When you’re happy with the results from the workflow that you create there, it is straightforward to turn it into a production-ready workflow with all of the settings that you want already in place!

Local quickstart

This quickstart uses a no-code, point-and-click user interface (UI) in your web browser to have Unstructured process a single file that is stored on your local machine.

The file is first processed on Unstructured-hosted compute resources. The UI then shows the processed data that Unstructured generates for that file. You can download that processed data as a .json file to your local machine.

This approach enables rapid, local, run-adjust-repeat prototyping of end-to-end Unstructured ETL+ workflows with a full range of Unstructured features. After you get the results you want, you can then attach remote source and destination connectors to both ends of your existing workflow to begin processing remote files and data at scale in production.

To run this quickstart, you will need a local file with a size of 10 MB or less and one of the following file types:

File type
.bmp
.csv
.doc
.docx
.email
.epub
.heic
.html
.jpg
.md
.odt
.org
.pdf
.pot
.potm
.ppt
.pptm
.pptx
.rst
.rtf
.sgl
.tiff
.txt
.tsv
.xls
.xlsx
.xml

For processing remote files at scale in production, Unstructured supports many more files types than these. See the list of supported file types.

Unstructured also supports processing files from remote object stores, and data from remote sources in websites, web apps, databases, and vector stores. For more information, see the source connector overview and the remote quickstart for how to set up and run production-ready Unstructured ETL+ workflows at scale.

If you do not have any files available, you can use one of the sample files that Unstructured offers in the UI. Or, you can download one or more sample files from the example-docs folder in the Unstructured repo on GitHub.

1

Sign up and sign in

Go to https://platform.unstructured.io and use your email address, Google account, or GitHub account to sign up for an Unstructured account (if you do not already have one) and sign into the account at the same time. The Unstructured user interface (UI) appears, and you can start using it right away.

By following the preceding instructions, you are signed up for a Developer pay per page account by default.

To save money, consider switching to a Subscribe & Save account instead. To save even more money, consider switching to an Enterprise account instead.

2

Create a workflow

  1. In the Unstructured UI, on the sidebar, click Workflows.

  2. Click New Workflow.

  3. Select Build it Myself, if it is not already selected.

  4. Click Continue. The visual workflow editor appears.

    The workflow is represented visually as a series of directed acyclic graph (DAG) nodes. Each node represents a step in the workflow. The workflow proceeds end to end from left to right. By default, the workflow starts with three nodes:

    • Source: This node represents the location where you have your files or data for Unstructured to process. For this quickstart, this node represents a single file on your local machine. After you get the results you want, you can update this node to represent files or data in a remote location at scale in production.
    • Partitioner: This node represents the partitioning step, which extracts content from unstructured files and data and outputs it as structured document elements for consistent representation across varying kinds of file and data types. For this quickstart, this node extracts the contents of a single file on your local machine and outputs it as a series of structured document elements in JSON format.
    • Destination: This node represents the location where you want Unstructured to put the processed files or data. After you get the results you want, you can update this node to have Unstructured put the processed files or data into a remote location at scale in production.
3

Process a local file

  1. Drag the file that you want Unstructured to process from your local machine’s file browser app and drop it into the Source node’s Drop file to test area. The file must have a size of 10 MB or less and one of the file types listed at the beginning of this quickstart.

    If you are not able to drag and drop the file, you can click Drop file to test and then browse to and select the file instead.

    Alternatively, you can use a sample file that Unstructured offers. To do this, click the Source node, and then in the Source pane, with Details selected, on the Local file tab, click one of the files under Or use a provided sample file. To view the file’s contents before you select it, click the eyes button next to the file.

  2. Above the Source node, click Test.

    Unstructured displays a visual representation of the file and begins processing its contents, sending it through each of the workflow’s nodes in sequence. Depending on the file’s size and the workflow’s complexity, this processing could take several minutes.

    After Unstructured has finished its processing, the processed data appears in the Test output pane, as a series of structured elements in JSON format.

  3. In the Test output pane, you can:

    • Search through the processed, JSON-formatted representation of the file by using the Search JSON box.
    • Download the full JSON as a .json file to your local machine by clicking Download full JSON.
  4. When you are done, click the Close button in the Test output pane.

4

Add more nodes to the workflow

  1. You can now add more nodes to the workflow to do further testing of various Unstructured features and with the option of eventually moving the workflow into production. For example, you can:

    • Add a Chunker node after the Partitioner node, to chunk the partitioned data into smaller pieces for your retrieval augmented generation (RAG) applications. To do this, click the add (+) button to the right of the Partitioner node, and then click Enrich > Chunker. Click the new Chunker node and specify its settings. For help, click the FAQ button in the Chunker node’s pane. Learn more about chunking and chunker settings.
    • Add an Enrichment node after the Chunker node, to apply enrichments to the chunked data such as image summaries, table summaries, table-to-HTML transforms, and named entity recognition (NER). To do this, click the add (+) button to the right of the Chunker node, and then click Enrich > Enrichment. Click the new Enrichment node and specify its settings. For help, click the FAQ button in the Enrichment node’s pane. Learn more about enrichments and enrichment settings.
    • Add an Embedder node after the Enrichment node, to generate vector embeddings for performing vector-based searches. To do this, click the add (+) button to the right of the Enrichment node, and then click Transform > Embedder. Click the new Embedder node and specify its settings. For help, click the FAQ button in the Embedder node’s pane. Learn more about embedding and embedding settings.
  2. Each time you add a node or change its settings, you can click Test above the Source node again to test the current workflow end to end and see the results of the changes, if any.

  3. Keep repeating this step as many times as you want, until you get the results you want.

5

Next steps

After you get the results you want, you have the option of moving your workflow into production. To do this, complete the following instructions.

The following instructions have you create a new workflow that is suitable for production. This behavior is planned to be fixed in a future release, allowing you to update the workflow that you just created, rather than needing to create a new one.

  1. With your workflow remaining open in the visual workflow editor, open a new tab in your web browser, and in this new tab, sign in to your Unstructured account, at https://platform.unstructured.io.

  2. In this new tab, create a source connector for your remote source location. This is the location in production where you have files or data in a file or object store, website, database, or vector store that you want Unstructured to process.

  3. Create a destination connector for your remote destination location. This is the location in production where you want Unstructured to put the processed data as .json files in a file or object store, or as records in a database or vector store.

  4. Create a workflow: on the sidebar, click Workflows, and then click New Workflow. Select Build it Myself, and then click Continue to open the visual workflow editor.

  5. In the visual workflow editor, click Source.

  6. In the Source pane, with Details selected, on the Connectors tab, select the source connector that you just created.

  7. Click the Destination node.

  8. In the Destination pane, with Details selected, select the destination connector that you just created.

  9. Using your original workflow on the other tab as a guide, add any additional nodes to this new workflow as needed, and configure those new nodes’ settings to match the other ones.

  10. Click Save.

  11. To run the workflow:

    a. Make sure to click Save first.
    b. Click the Close button next to the workflow’s name in the top navigation bar.
    c. On the sidebar, click Workflows.
    d. In the list of available workflows, click the Run button for the workflow that you just saved.
    e. On the sidebar, click Jobs.

    f. In the list of available jobs, click the job that you just ran.
    g. After the job status shows Finished, go to the your destination location to see the processed files or data that Unstructured put there.

See also the remote quickstart for more coverage about how to set up and run production-ready Unstructured ETL+ workflows at scale.

Remote quickstart

This quickstart uses a no-code, point-and-click user interface in your web browser to get all of your data RAG-ready. Data is processed on Unstructured-hosted compute resources.

The requirements are as follows.

1

Sign up and sign in

Go to https://platform.unstructured.io and use your email address, Google account, or GitHub account to sign up for an Unstructured account (if you do not already have one) and sign into the account at the same time. The Unstructured user interface (UI) appears, and you can start using it right away.

By following the preceding instructions, you are signed up for a Developer pay per page account by default.

To save money, consider switching to a Subscribe & Save account instead. To save even more money, consider switching to an Enterprise account instead.

2

Set the source (input) location

  1. From your Unstructured dashboard, in the sidebar, click Connectors.
  2. Click Sources.
  3. Cick New or Create Connector.
  4. For Name, enter some unique name for this connector.
  5. In the Provider area, click the source location type that matches yours.
  6. Click Continue.
  7. Fill in the fields with the appropriate settings. Learn more.
  8. If a Continue button appears, click it, and fill in any additional settings fields.
  9. Click Save and Test.
3

Set the destination (output) location

  1. In the sidebar, click Connectors.
  2. Click Destinations.
  3. Cick New or Create Connector.
  4. For Name, enter some unique name for this connector.
  5. In the Provider area, click the destination location type that matches yours.
  6. Click Continue.
  7. Fill in the fields with the appropriate settings. Learn more.
  8. If a Continue button appears, click it, and fill in any additional settings fields.
  9. Click Save and Test.
4

Define the workflow

  1. In the sidebar, click Workflows.

  2. Click New Workflow.

  3. Next to Build it for Me, click Create Workflow.

    If a radio button appears instead of Build it for Me, select it, and then click Continue.
  4. For Workflow Name, enter some unique name for this workflow.

  5. In the Sources dropdown list, select your source location from Step 3.

  6. In the Destinations dropdown list, select your destination location from Step 4.

    You can select multiple source and destination locations. Files will be ingested from all of the selected source locations, and the processed data will be delivered to all of the selected destination locations.
  7. Click Continue.

  8. The Reprocess All box applies only to blob storage connectors such as the Amazon S3, Azure Blob Storage, and Google Cloud Storage connectors:

    • Checking this box reprocesses all documents in the source location on every workflow run.
    • Unchecking this box causes new documents that have been added to the source location, as well as existing documents in the source location that have had their contents or titles changed, since the last workflow run to be processed on future runs. Other previously processed documents are not processed again.
  9. Click Continue.

  10. If you want this workflow to run on a schedule, in the Repeat Run dropdown list, select one of the scheduling options, and fill in the scheduling settings. Otherwise, select Don’t repeat.

  11. Click Complete.

5

Process the documents

  1. If you did not choose to run this workflow on a schedule in Step 5, you can run the workflow now: on the sidebar, click Workflows.
  2. Next to your workflow from Step 5, click Run.
6

Monitor the processing job

  1. In the sidebar, click Jobs.
  2. In the list of jobs, wait for the job’s Status to change to Finished.
  3. Click the row for the job.
  4. After Overview displays Finished, go to the next Step.
7

View the processed data

Go to your destination location to view the processed data.