# How to Use Web Scraper (Crawl Data) Feature

## Feature Purpose and Value

As an enterprise user, you may often receive instructions from management to reference or compile regulatory information from certain public websites.

If you have a technical background, or have engineering support, you might be able to automatically extract data by writing crawler programs. However, for non-technical personnel, this usually means manually organizing page by page, which is not only time-consuming and labor-intensive but also prone to missing critical information.

In such cases, you can leverage MaiAgent's web crawler feature to quickly extract website content through a No-Code approach, automatically create structured data, significantly improve information organization efficiency, and invest your time in higher-value core business activities.

## How to Perform Web Crawling?

To create a crawler request, you can:

1. Create a Page Crawl Request

Navigate to "<mark style="color:blue;">AI Features > AI Assistant > Crawler</mark>" in the left sidebar, and click the "<mark style="color:blue;">+ Create Page Crawl Request</mark>" button in the upper right corner.

<figure><img src="https://1360999650-files.gitbook.io/~/files/v0/b/gitbook-x-prod.appspot.com/o/spaces%2F6v6TNkkOQVfRYfcNirHL%2Fuploads%2Fgit-blob-b0d71725289bf274603ea789e3c522873fe3ef6d%2F%E6%88%AA%E5%9C%96%202025-04-25%20%E4%B8%8A%E5%8D%8811.13.16.png?alt=media" alt=""><figcaption></figcaption></figure>

2. Enter URL

Enter the URL of the page you want to crawl and press the \[<mark style="color:blue;">Confirm</mark>] button.

{% hint style="warning" %}

* Please note that the URL cannot exceed 200 characters
* If the status does not change, you can click the refresh button in the upper right corner to update the page status
  {% endhint %}

<div><figure><img src="https://1360999650-files.gitbook.io/~/files/v0/b/gitbook-x-prod.appspot.com/o/spaces%2F6v6TNkkOQVfRYfcNirHL%2Fuploads%2Fgit-blob-c8baadbaff430be76f61669bb979ade37b20ebdb%2Fimage%20(178).png?alt=media" alt=""><figcaption></figcaption></figure> <figure><img src="https://1360999650-files.gitbook.io/~/files/v0/b/gitbook-x-prod.appspot.com/o/spaces%2F6v6TNkkOQVfRYfcNirHL%2Fuploads%2Fgit-blob-8dc8f4ee375fa21dda1f45d1b49ba95b409b2de0%2Fimage%20(179).png?alt=media" alt=""><figcaption></figcaption></figure></div>

3. View Crawled Data

When the status shows completed, click "Import" on the right to view the crawled data entries.

4. Select Data

Check the boxes on the left to select the data you want to import into the knowledge base. After making your selections, click the "<mark style="color:blue;">Import</mark>" button, and the data will be automatically imported into that AI assistant's knowledge base.

{% hint style="info" %}
If you want to view more data entries on the same page, you can click "<mark style="color:blue;">10 items/page</mark>" in the lower right corner to expand the viewing range.
{% endhint %}

<div><figure><img src="https://1360999650-files.gitbook.io/~/files/v0/b/gitbook-x-prod.appspot.com/o/spaces%2F6v6TNkkOQVfRYfcNirHL%2Fuploads%2Fgit-blob-752a9a976df08f04b29a5e79686730940d5c048a%2Fimage%20(180).png?alt=media" alt=""><figcaption></figcaption></figure> <figure><img src="https://1360999650-files.gitbook.io/~/files/v0/b/gitbook-x-prod.appspot.com/o/spaces%2F6v6TNkkOQVfRYfcNirHL%2Fuploads%2Fgit-blob-a6e923d56b285025bced24d2f1443565f85c0f8a%2Fimage%20(181).png?alt=media" alt=""><figcaption></figcaption></figure></div>

In the knowledge base, you can see the data presented as .md files, which can be configured with tags and metadata just like regular data.

<figure><img src="https://1360999650-files.gitbook.io/~/files/v0/b/gitbook-x-prod.appspot.com/o/spaces%2F6v6TNkkOQVfRYfcNirHL%2Fuploads%2Fgit-blob-0cd115d5d2d7f7900e35bbbf41568e1fcd5293e3%2Fimage%20(182).png?alt=media" alt=""><figcaption></figcaption></figure>

## Web Crawler Usage Notes

* Please ensure you have permission to crawl content from the target website
* It is recommended to test with a small amount of data before performing large-scale crawling
* After crawling is complete, you can verify data quality through the [search test feature](https://docs.maiagent.ai/maiagent-user-guide/maiagent-user-guide-en/km/test-search-result)
* Regularly update crawled data to maintain information timeliness
