:original_name: mrs_01_1084.html .. _mrs_01_1084: Using Loader from Scratch ========================= You can use Loader to import data from the SFTP server to HDFS. This section applies to MRS clusters earlier than 3.\ *x*. Prerequisites ------------- - You have prepared service data. - You have created an analysis cluster. Procedure --------- #. Access the Loader page. a. Access the cluster details page. - For versions earlier than MRS 1.9.2, log in to MRS Manager and choose **Services**. - For MRS 1.9.2 or later, click the cluster name on the MRS console and choose **Components**. b. Choose **Hue**. In **Hue Web UI** of **Hue Summary**, click **Hue (Active)**. The Hue web UI is displayed. c. Choose **Data Browsers** > **Sqoop**. The job management tab page is displayed by default on the Loader page. #. On the Loader page, click **Manage links**. #. .. _mrs_01_1084__li48883218306: Click **New link** and create **sftp-connector**. For details, see :ref:`File Server Link `. #. .. _mrs_01_1084__li14723052103216: Click **New link**, enter the link name, select **hdfs-connector**, and create **hdfs-connector**. #. On the Loader page, click **Manage jobs**. #. Click **New Job**. #. In **Connection**, set parameters. a. In **Name**, enter a job name. b. Select the source link created in :ref:`3 ` and the target link created in :ref:`4 `. #. In **From**, configure the job of the source link. For details, see :ref:`ftp-connector or sftp-connector `. #. In **To**, configure the job of the target link. For details, see :ref:`hdfs-connector `. #. In **Task Config**, set job running parameters. .. table:: **Table 1** Loader job running properties +--------------------------------------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------+ | Parameter | Description | +======================================+========================================================================================================================================================================+ | Extractors | Number of Map tasks | +--------------------------------------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------+ | Loaders | Number of Reduce tasks | | | | | | This parameter is displayed only when the destination field is HBase or Hive. | +--------------------------------------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------+ | Max. Error Records in a Single Shard | Error record threshold. If the number of error records of a single Map task exceeds the threshold, the task automatically stops and the obtained data is not returned. | | | | | | .. note:: | | | | | | Data is read and written in batches for **MYSQL** and **MPPDB** of **generic-jdbc-connector** by default. Errors are recorded once at most for each batch of data. | +--------------------------------------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------+ | Dirty Data Directory | Directory for saving dirty data. If you leave this parameter blank, dirty data will not be saved. | +--------------------------------------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------+ #. Click **Save**.