Skip to content

Latest commit

 

History

History
81 lines (53 loc) · 3.23 KB

ccs-tooling.md

File metadata and controls

81 lines (53 loc) · 3.23 KB
copyright lastupdated subcollection
years
2019, 2024
2023-03-02
discovery-data

{{site.data.keyword.attribute-definition-list}}

Using a custom Cloud Pak for Data connector from the Discovery user interface

{: #ccs-tooling}

After you build and deploy a custom connector, you can configure and run it in the {{site.data.keyword.discoveryshort}} user interface to create a collection. {: shortdesc}

[IBM Cloud Pak for Data]{: tag-cp4d} {{site.data.keyword.icp4dfull_notm}} only

This information applies only to installed deployments. {: note}

You create and manage a collection as described in Creating and managing collections. You can use a successfully deployed custom connector during this process as follows. Follow these instructions to use a custom connector instead of one of the pre-built connectors that are listed in Configuring Cloud Pak for Data data sources.

  1. After you create a project, look for your custom connector to connect to a data source.

  2. Select the custom connector and then click Next.

    The Configure collection page opens.

    The following steps apply specifically to the example custom connector that is included with the custom-crawler-docs.zip file. {: note}

  3. Enter values for the following fields on the Configure collection page. If a field is already populated with a value, verify and change the value if needed. A prepopulated value indicates that a value was specified in the custom connector's template.xml or message.properties file.

    General : Complete the following fields

    -   Collection name
    -   Collection language
    -   Crawl schedule
    

    Crawler properties : Complete the following fields

    -   Crawler name
    -   Crawler description
    -   Time to wait between retrieval requests (milliseconds)
    
        The default value is `0`.
    -   Maximum number of active crawler threads
    
        The default value is `10`.
    - Maximum number of documents to crawl 
    
      The default value `2000000000`.
    - Maximum document size (KB)
    
      The default value is `32768`.
    

    Data source properties : Complete the following fields

    -   Host name
    
        The default value is `localhost`.
    -   Port
    
        The default value is `22`.
    -   User name
    -   Use key file (or input password)
    
        The default value is `On`.
    -   Key file location
    -   passphrase
    -   Password
    

    Crawl Space Properties : If the custom crawler supports document-level security and the document_level_security_supported value in the template.xml is set to true, then an Enable Document Level Security switch is displayed in a Security section of the data source connection setup page. To enable document-level security, set the Enable Document Level Security switch to On. If the switch is set to Off, then the collection that is created cannot support document-level security even if the custom crawler can support document-level security.

  4. Click Finish to create the collection.