Create and Manage Crawlers
Create crawler
In Data Catalog, a crawler scans MySQL data, extracts metadata, and automatically updates the Data Catalog to make data discovery easier. To create a crawler, you must first have a database and MySQL created.
For instructions on creating a database and MySQL, see:
To create a crawler in the Data Catalog service:
- Go to the KakaoCloud console > Analytics > Data Catalog.
- In Crawlers, click Create crawler.
- In Create crawler, enter the required information and click Create.
Field Description Database The name of the database to which tables will be added.
- Only databases in theACTIVEstate are listed.
- After selecting a database, you can check the network/subnet information.
- Crawlers are not supported for Iceberg-type catalogs.Crawler name Name of the crawler MySQL full path Select the MySQL to connect to and enter the database name of that MySQL.
- Only MySQL instances in theAVAILABLEstate are listed.MySQL account Enter the username and password configured when creating MySQL.
- Connection test: After entering the MySQL full path and account information, click Test to verify the connection.
* If the connection test does not complete successfully, you cannot create the crawler.Description (optional) Additional description for the crawler Table prefix (optional) A prefix added to created table names. Tables are created as Prefix + MySQL database name_table name.
- Allowed characters: lowercase letters, digits, and underscore (_) only (1–64 chars)Schedule Manage when the crawler runs.
- For on-demand, it runs only when triggered manually and has no schedule.
Only resources whose state is normal (ACTIVE / AVAILABLE) are listed for databases and MySQL.
Manage crawlers
This section explains how to manage crawlers in the Data Catalog service.
View crawler list
View the list of crawlers currently in use.
-
Go to the KakaoCloud console > Analytics > Data Catalog.
-
Click Crawlers to see the list.
Column Description Name The crawler name entered at creation.
- Click the crawler name to open its Details tab.Description Description entered at creation Status Crawler status Schedule The schedule on which the crawler runs Last run status Status of the most recent run Last run time Timestamp when the crawler most recently ran [More] icon - Edit: Modify the crawler description and schedule
- Run: Run the crawler manually
- Delete: Delete the crawler
* When the crawler state isCREATING/ALTERING/DELETING/RUNNING, you cannot edit, run, or delete it.
View crawler details
Check detailed information about a crawler.
- Go to the KakaoCloud console > Analytics > Data Catalog.
- Click Crawlers, then select the crawler whose details you want to view.
- Review the crawler’s details.
View crawler run history
Check the run history of a crawler.
-
Go to the KakaoCloud console > Analytics > Data Catalog.
-
Click Crawlers, then select the crawler whose history you want to view.
-
In the details page, click the Run history tab to review past runs.
infoCrawler run history is retained for up to 90 days. Records older than 90 days are deleted automatically.
Column Description Start time When the crawl started End time When the crawl finished Duration How long the crawler ran Status Status of the run
-Succeeded: The crawl finished successfully
-Running: The crawl is in progress
-Failed: The crawl failed
Delete a crawler
Delete crawlers you no longer need.
Deleted crawlers and their run histories cannot be restored. If a catalog is deleted, its crawlers are deleted automatically.
- Go to the KakaoCloud console > Analytics > Data Catalog.
- Click Crawlers, then in the list click the [More] icon for the crawler you want to remove and select Delete.
- In the Delete dialog, enter the crawler name exactly and click Delete.