fix docs typo (#4725)
This commit is contained in:
@ -34,7 +34,7 @@ SET PASSWORD [FOR user_identity] =
|
||||
|
||||
The SET PASSWORD command can be used to modify a user's login password. If the [FOR user_identity] field does not exist, modify the password of the current user.
|
||||
|
||||
Note that the user_identity here must match exactly the user_identity specified when creating a user using CREATE USER, otherwise the user will be reported as non-existent. If user_identity is not specified, the current user is'username'@'ip', which may not match any user_identity. The current user can be viewed through SHOW GRANTS.
|
||||
Note that the user_identity here must match exactly the user_identity specified when creating a user using CREATE USER, otherwise the user will be reported as non-existent. If user_identity is not specified, the current user is 'username'@'ip', which may not match any user_identity. The current user can be viewed through SHOW GRANTS.
|
||||
|
||||
PASSWORD () input is a plaintext password, and direct use of strings, you need to pass the encrypted password.
|
||||
If you change the password of other users, you need to have administrator privileges.
|
||||
|
||||
@ -31,7 +31,7 @@ Syntax:
|
||||
|
||||
SET PROPERTY [FOR 'user'] 'key' = 'value' [, 'key' = 'value']
|
||||
|
||||
Set user attributes, including resources allocated to users, import cluster, etc. The user attributes set here are for user, not user_identity. That is to say, if two users'jack'@'%' and'jack'@'192%'are created through the CREATE USER statement, the SET PROPERTY statement can only be used for the jack user, not'jack'@'%' or'jack'@'192%'
|
||||
Set user attributes, including resources allocated to users, import cluster, etc. The user attributes set here are for user, not user_identity. That is to say, if two users 'jack'@'%' and 'jack'@'192%'are created through the CREATE USER statement, the SET PROPERTY statement can only be used for the jack user, not 'jack'@'%' or 'jack'@'192%'
|
||||
|
||||
Importing cluster is only applicable to Baidu internal users.
|
||||
|
||||
@ -49,7 +49,7 @@ Quota.low: Resource allocation at low level.
|
||||
|
||||
Load_cluster. {cluster_name}. hadoop_palo_path: The Hadoop directory used by Palo needs to store ETL programs and intermediate data generated by ETL for Palo to import. After the import is completed, the intermediate data will be automatically cleaned up, and the ETL program will be automatically reserved for next use.
|
||||
Load_cluster. {cluster_name}. hadoop_configs: configuration of hadoop, where fs. default. name, mapred. job. tracker, hadoop. job. UGI must be filled in.
|
||||
Load ucluster. {cluster name}. hadoop port: Hadoop HDFS name node http}
|
||||
Load_cluster. {cluster_name}. hadoop_port: Hadoop HDFS name node http}
|
||||
Default_load_cluster: The default import cluster.
|
||||
|
||||
## example
|
||||
@ -61,11 +61,11 @@ SET PROPERTY FOR 'jack' 'max_user_connections' = '1000';
|
||||
SET PROPERTY FOR 'jack' 'resource.cpu_share' = '1000';
|
||||
|
||||
3. Modify the weight of the normal group of Jack users
|
||||
Set property for'jack''quota. normal' = 400';
|
||||
Set property for 'jack''quota. normal' = 400';
|
||||
|
||||
4. Add import cluster for user jack
|
||||
SET PROPERTY FOR 'jack'
|
||||
'load 'cluster.{cluster name}.hadoop'u palo path' ='/user /palo /palo path',
|
||||
'load 'cluster.{cluster name}.hadoop' palo path' ='/user /palo /palo path',
|
||||
'load_cluster.{cluster_name}.hadoop_configs' = 'fs.default.name=hdfs://dpp.cluster.com:port;mapred.job.tracker=dpp.cluster.com:port;hadoop.job.ugi=user,password;mapred.job.queue.name=job_queue_name_in_hadoop;mapred.job.priority=HIGH;';
|
||||
|
||||
5. Delete the import cluster under user jack.
|
||||
|
||||
@ -26,7 +26,7 @@ under the License.
|
||||
|
||||
# RESTORE
|
||||
## Description
|
||||
1. RESTOR
|
||||
1. RESTORE
|
||||
This statement is used to restore the data previously backed up by the BACKUP command to the specified database. This command is an asynchronous operation. After successful submission, you need to check progress through the SHOW RESTORE command. Restoring tables of OLAP type is supported only.
|
||||
Grammar:
|
||||
SNAPSHOT RESTORE [dbu name].{snapshot name}
|
||||
@ -47,7 +47,7 @@ Explain:
|
||||
"Backup_timestamp" = "2018-05-04-16-45-08": specifies which version of the time to restore the corresponding backup must be filled in. This information can be obtained through the `SHOW SNAPSHOT ON repo;'statement.
|
||||
"Replication_num" = "3": Specifies the number of replicas of the restored table or partition. The default is 3. If an existing table or partition is restored, the number of copies must be the same as the number of copies of an existing table or partition. At the same time, there must be enough hosts to accommodate multiple copies.
|
||||
"Timeout" = "3600": Task timeout, default to one day. Unit seconds.
|
||||
"Meta_version" = 40: Use the specified meta_version to read the previously backed up metadata. Note that as a temporary solution, this parameter is only used to restore the data backed up by the older version of Doris. The latest version of the backup data already contains metaversion, no need to specify.
|
||||
"Meta_version" = 40: Use the specified meta_version to read the previously backed up metadata. Note that as a temporary solution, this parameter is only used to restore the data backed up by the older version of Doris. The latest version of the backup data already contains meta version, no need to specify.
|
||||
|
||||
## example
|
||||
1. Restore backup table backup_tbl in snapshot_1 from example_repo to database example_db1 with the time version of "2018-05-04-16-45-08". Restore to one copy:
|
||||
|
||||
@ -41,7 +41,7 @@ SHOW [FULL] [BUILTIN] FUNCTIONS [IN|FROM db] [LIKE 'function_pattern']
|
||||
|
||||
Look at all the custom(builtin) functions under the database. If the user specifies the database, then look at the corresponding database, otherwise directly query the database where the current session is located.
|
||||
|
||||
You need `SHOW'privileges for this database
|
||||
You need `SHOW` privileges for this database
|
||||
|
||||
## example
|
||||
|
||||
|
||||
@ -33,7 +33,7 @@ Subsequent imports of new features will only be supported in STEAM LOAD, MINI LO
|
||||
MINI LOAD is imported through HTTP protocol. Users can import without relying on Hadoop or Mysql client.
|
||||
The user describes the import through HTTP protocol, and the data is streamed into Doris in the process of receiving http requests. After the ** import job is completed, the ** returns to the user the imported results.
|
||||
|
||||
* Note: In order to be compatible with the old version of mini load usage habits, users can still view the import results through the'SHOW LOAD'command.
|
||||
* Note: In order to be compatible with the old version of mini load usage habits, users can still view the import results through the 'SHOW LOAD' command.
|
||||
|
||||
Grammar:
|
||||
Import:
|
||||
@ -49,13 +49,13 @@ HTTP Protocol Specification
|
||||
Privilege Authentication Currently Doris uses the Basic mode of HTTP for privilege authentication. So you need to specify a username and password when importing
|
||||
This way is to pass the password in plaintext, and does not support encrypted transmission for the time being.
|
||||
|
||||
Expect Doris needs to send an HTTP request with the'Expect'header information,'100-continue'.
|
||||
Expect Doris needs to send an HTTP request with the 'Expect' header information,'100-continue'.
|
||||
Why? Because we need to redirect the request, we have to transfer the data content before.
|
||||
This can avoid causing multiple data transmission, thereby improving efficiency.
|
||||
|
||||
Content-Length Doris needs to send a request with the header'Content-Length'. If the content ratio is sent
|
||||
'Content-Length'is less, so Doris believes that if there is a transmission problem, the submission task fails.
|
||||
NOTE: If you send more data than'Content-Length', Doris reads only'Content-Length'.
|
||||
Content-Length Doris needs to send a request with the header 'Content-Length'. If the content ratio is sent
|
||||
'Content-Length' is less, so Doris believes that if there is a transmission problem, the submission task fails.
|
||||
NOTE: If you send more data than 'Content-Length', Doris reads only 'Content-Length'.
|
||||
Length content and import
|
||||
|
||||
|
||||
@ -72,9 +72,9 @@ The specified method is comma-separated, such as columns = k1, k2, k3, K4
|
||||
|
||||
Column_separator: Used to specify the separator between columns, default is' t'
|
||||
NOTE: Url encoding is required, for example
|
||||
If you need to specify' t'as a separator, you should pass in'column_separator=% 09'
|
||||
If you need to specify'x01'as a delimiter, you should pass in'column_separator=% 01'
|
||||
If you need to specify','as a separator, you should pass in'column_separator=% 2c'
|
||||
If you need to specify '\t' as a separator, you should pass in 'column_separator=% 09'
|
||||
If you need to specify 'x01'as a delimiter, you should pass in 'column_separator=% 01'
|
||||
If you need to specify','as a separator, you should pass in 'column_separator=% 2c'
|
||||
|
||||
|
||||
Max_filter_ratio: Used to specify the maximum percentage allowed to filter irregular data, default is 0, not allowed to filter
|
||||
@ -101,22 +101,22 @@ Although the information of mini load can be found in subsequent queries, it can
|
||||
|
||||
'35;'35; example
|
||||
|
||||
1. Import the data from the local file'testData'into the table of'testTbl' in the database'testDb'(the user is in defalut_cluster)
|
||||
1. Import the data from the local file 'testData' into the table of 'testTbl' in the database 'testDb'(the user is in default_cluster)
|
||||
curl --location-trusted -u root -T testData http://host:port/api/testDb/testTbl/_load?label=123
|
||||
|
||||
2. Import the data from the local file'testData'into the table of'testTbl' in the database'testDb'(the user is in test_cluster). The timeout time is 3600 seconds.
|
||||
2. Import the data from the local file 'testData' into the table of 'testTbl' in the database'testDb'(the user is in test_cluster). The timeout time is 3600 seconds.
|
||||
curl --location-trusted -u root@test_cluster:root -T testData http://fe.host:port/api/testDb/testTbl/_load?label=123&timeout=3600
|
||||
|
||||
3. Import data from the local file'testData'into the'testTbl' table in the database'testDb', allowing a 20% error rate (the user is in defalut_cluster)
|
||||
3. Import data from the local file 'testData' into the 'testTbl' table in the database 'testDb', allowing a 20% error rate (the user is in default_cluster)
|
||||
curl --location-trusted -u root -T testData http://host:port/api/testDb/testTbl/_load?label=123\&max_filter_ratio=0.2
|
||||
|
||||
4. Import the data from the local file'testData'into the table'testTbl' in the database'testDb', allowing a 20% error rate, and specify the column name of the file (the user is in defalut_cluster)
|
||||
4. Import the data from the local file 'testData' into the table 'testTbl' in the database 'testDb', allowing a 20% error rate, and specify the column name of the file (the user is in default_cluster)
|
||||
curl --location-trusted -u root -T testData http://host:port/api/testDb/testTbl/_load?label=123\&max_filter_ratio=0.2\&columns=k1,k2,k3
|
||||
|
||||
5. Import in streaming mode (user is in defalut_cluster)
|
||||
5. Import in streaming mode (user is in default_cluster)
|
||||
seq 1 10 | awk '{OFS="\t"}{print $1, $1 * 10}' | curl --location-trusted -u root -T - http://host:port/api/testDb/testTbl/_load?label=123
|
||||
|
||||
6. Import tables containing HLL columns, which can be columns in tables or columns in data to generate HLL columns (users are in defalut_cluster)
|
||||
6. Import tables containing HLL columns, which can be columns in tables or columns in data to generate HLL columns (users are in default_cluster)
|
||||
|
||||
curl --location-trusted -u root -T testData http://host:port/api/testDb/testTbl/_load?label=123\&max_filter_ratio=0.2\&hll=hll_column1,k1:hll_column2,k2
|
||||
\&columns=k1,k2,k3
|
||||
|
||||
@ -35,23 +35,23 @@ curl --location-trusted -u user:passwd -XPOST http://host:port/api/{db}/_multi_c
|
||||
curl --location-trusted -u user:passwd -XPOST http://host:port/api/{db}/_multi_desc?label=xxx
|
||||
|
||||
'MULTI LOAD'can support users to import multiple tables at the same time on the basis of'MINI LOAD'. The specific commands are shown above.
|
||||
'/api/{db}/_multi_start'starts a multi-table import task
|
||||
'/api/{db}/{table}/_load'adds a table to be imported to an import task. The main difference from'MINI LOAD' is that the'sub_label'parameter needs to be passed in.
|
||||
'/api/{db}/_multi_commit'submits the entire multi-table import task and the background begins processing
|
||||
'/api/{db}/_multi_abort'Abandons a multi-table import task
|
||||
'/api/{db}/_multi_desc'shows the number of jobs submitted by a multi-table import task
|
||||
'/api/{db}/_multi_start' starts a multi-table import task
|
||||
'/api/{db}/{table}/_load' adds a table to be imported to an import task. The main difference from 'MINI LOAD' is that the 'sub_label' parameter needs to be passed in.
|
||||
'/api/{db}/_multi_commit' submits the entire multi-table import task and the background begins processing
|
||||
'/api/{db}/_multi_abort' Abandons a multi-table import task
|
||||
'/api/{db}/_multi_desc' shows the number of jobs submitted by a multi-table import task
|
||||
|
||||
HTTP Protocol Specification
|
||||
Privilege Authentication Currently Doris uses the Basic mode of HTTP for privilege authentication. So you need to specify a username and password when importing
|
||||
This way is to pass passwords in plaintext, since we are all in the Intranet environment at present...
|
||||
|
||||
Expect Doris needs to send an HTTP request, and needs the'Expect'header information with the content of'100-continue'.
|
||||
Expect Doris needs to send an HTTP request, and needs the 'Expect' header information with the content of'100-continue'.
|
||||
Why? Because we need to redirect the request, we have to transfer the data content before.
|
||||
This can avoid causing multiple data transmission, thereby improving efficiency.
|
||||
|
||||
Content-Length Doris needs to send a request with the header'Content-Length'. If the content ratio is sent
|
||||
Content-Length Doris needs to send a request with the header 'Content-Length'. If the content ratio is sent
|
||||
If'Content-Length'is less, Palo believes that if there is a transmission problem, the submission of the task fails.
|
||||
NOTE: If you send more data than'Content-Length', Doris reads only'Content-Length'.
|
||||
NOTE: If you send more data than 'Content-Length', Doris reads only 'Content-Length'.
|
||||
Length content and import
|
||||
|
||||
Description of parameters:
|
||||
@ -67,8 +67,8 @@ If it is not passed in, the column order in the file is considered to be the sam
|
||||
The specified method is comma-separated, such as columns = k1, k2, k3, K4
|
||||
|
||||
Column_separator: Used to specify the separator between columns, default is' t'
|
||||
NOTE: Url encoding is required, such as specifying't'as a delimiter.
|
||||
Then you should pass in'column_separator=% 09'
|
||||
NOTE: Url encoding is required, such as specifying '\t'as a delimiter.
|
||||
Then you should pass in 'column_separator=% 09'
|
||||
|
||||
Max_filter_ratio: Used to specify the maximum percentage allowed to filter irregular data, default is 0, not allowed to filter
|
||||
Custom specification should be as follows:'max_filter_ratio = 0.2', meaning that 20% error rate is allowed.
|
||||
@ -86,19 +86,19 @@ Real import behavior will occur, and the amount of data in this way can not be t
|
||||
|
||||
'35;'35; example
|
||||
|
||||
1. Import the data from the local file'testData1'into the table of'testTbl1' in the database'testDb', and
|
||||
Import the data from'testData2'into the table'testTbl2' in'testDb'(the user is in defalut_cluster)
|
||||
1. Import the data from the local file 'testData1'into the table of 'testTbl1' in the database 'testDb', and
|
||||
Import the data from 'testData2'into the table 'testTbl2' in 'testDb'(the user is in default_cluster)
|
||||
curl --location-trusted -u root -XPOST http://host:port/api/testDb/_multi_start?label=123
|
||||
curl --location-trusted -u root -T testData1 http://host:port/api/testDb/testTbl1/_load?label=123\&sub_label=1
|
||||
curl --location-trusted -u root -T testData2 http://host:port/api/testDb/testTbl2/_load?label=123\&sub_label=2
|
||||
curl --location-trusted -u root -XPOST http://host:port/api/testDb/_multi_commit?label=123
|
||||
|
||||
2. Multi-table Import Midway Abandon (User in defalut_cluster)
|
||||
2. Multi-table Import Midway Abandon (User in default_cluster)
|
||||
curl --location-trusted -u root -XPOST http://host:port/api/testDb/_multi_start?label=123
|
||||
curl --location-trusted -u root -T testData1 http://host:port/api/testDb/testTbl1/_load?label=123\&sub_label=1
|
||||
curl --location-trusted -u root -XPOST http://host:port/api/testDb/_multi_abort?label=123
|
||||
|
||||
3. Multi-table import to see how much content has been submitted (user is in defalut_cluster)
|
||||
3. Multi-table import to see how much content has been submitted (user is in default_cluster)
|
||||
curl --location-trusted -u root -XPOST http://host:port/api/testDb/_multi_start?label=123
|
||||
curl --location-trusted -u root -T testData1 http://host:port/api/testDb/testTbl1/_load?label=123\&sub_label=1
|
||||
curl --location-trusted -u root -XPOST http://host:port/api/testDb/_multi_desc?label=123
|
||||
|
||||
@ -27,7 +27,7 @@ under the License.
|
||||
# ROUTINE LOAD
|
||||
## description
|
||||
|
||||
Routine Load function allows users to submit a resident load task, and continuously load data into Doris by continuously reading data from the specified data source. Currently, only text data format (CSV) data is loaded from Kakfa by means of no authentication or SSL authentication.
|
||||
Routine Load function allows users to submit a resident load task, and continuously load data into Doris by continuously reading data from the specified data source. Currently, only text data format (CSV) data is loaded from Kafka by means of no authentication or SSL authentication.
|
||||
|
||||
Syntax:
|
||||
|
||||
@ -214,7 +214,7 @@ FROM data_source
|
||||
|
||||
`Kafka_broker_list`
|
||||
|
||||
Kafka's broker connection information. The format is ip:host. Multiple brokare separated by commas.
|
||||
Kafka's broker connection information. The format is ip:host. Multiple brokers are separated by commas.
|
||||
|
||||
Example:
|
||||
|
||||
@ -234,9 +234,9 @@ FROM data_source
|
||||
|
||||
Offset can specify a specific offset from 0 or greater, or:
|
||||
|
||||
1) OFFSET_BEGINNING: Subscribe from the location where the data is avaie.
|
||||
1) OFFSET_BEGINNING: Subscribe from the location where the data is available.
|
||||
|
||||
2) OFFSET_END: Subscribe from the end.
|
||||
2) OFFSET_END: Subscribe from the end.
|
||||
|
||||
If not specified, all partitions under topic are subscribed by default fromSET_END.
|
||||
|
||||
@ -253,7 +253,7 @@ FROM data_source
|
||||
|
||||
The function is equivalent to the "--property" parameter in the kafka shel
|
||||
|
||||
When the value of the parameter is a file, you need to add the keyword: "FILbefore the value.
|
||||
When the value of the parameter is a file, you need to add the keyword: "FILE" before the value.
|
||||
|
||||
For information on how to create a file, see "HELP CREATE FILE;"
|
||||
|
||||
@ -266,7 +266,7 @@ FROM data_source
|
||||
"property.ssl.ca.location" = "FILE:ca.pem"
|
||||
```
|
||||
|
||||
1. When connecting to Kafka using SSL, you need to specify the follg parameters:
|
||||
1. When connecting to Kafka using SSL, you need to specify the following parameters:
|
||||
|
||||
```
|
||||
"property.security.protocol" = "ssl",
|
||||
@ -278,9 +278,9 @@ FROM data_source
|
||||
|
||||
among them:
|
||||
|
||||
"property.security.protocol" and "property.ssl.ca.location" are requ to indicate the connection method is SSL and the location of the CA certate.
|
||||
"property.security.protocol" and "property.ssl.ca.location" are required to indicate the connection method is SSL and the location of the CA certificate.
|
||||
|
||||
If the client authentication is enabled on the Kafka server, you alsod to set:
|
||||
If the client authentication is enabled on the Kafka server, you also need to set:
|
||||
|
||||
```
|
||||
"property.ssl.certificate.location"
|
||||
@ -292,11 +292,11 @@ FROM data_source
|
||||
|
||||
2. Specify the default starting offset for kafka partition
|
||||
|
||||
If kafka_partitions/kafka_offsets is not specified, all partitions are umed by default, and you can specify kafka_default_offsets to specify the star offset. The default is OFFSET_END, which starts at the end of the substion.
|
||||
If kafka_partitions/kafka_offsets is not specified, all partitions are unanmed by default, and you can specify kafka_default_offsets to specify the star offset. The default is OFFSET_END, which starts at the end of the subscription.
|
||||
|
||||
Values:
|
||||
|
||||
1) OFFSET_BEGINNING: Subscribe from the location where the data is avaie.
|
||||
1) OFFSET_BEGINNING: Subscribe from the location where the data is available.
|
||||
|
||||
2) OFFSET_END: Subscribe from the end.
|
||||
|
||||
@ -309,8 +309,8 @@ FROM data_source
|
||||
Integer class (TINYINT/SMALLINT/INT/BIGINT/LARGEINT): 1, 1000, 1234
|
||||
|
||||
Floating point class (FLOAT/DOUBLE/DECIMAL): 1.1, 0.23, .356
|
||||
|
||||
Date class (DATE/DATETIME): 2017-10-03, 2017-06-13 12:34:03.
|
||||
|
||||
Date class (DATE/DATETIME): 2017-10-03, 2017-06-13 12:34:03.
|
||||
|
||||
String class (CHAR/VARCHAR) (without quotes): I am a student, a
|
||||
|
||||
@ -505,7 +505,7 @@ FROM data_source
|
||||
]
|
||||
}
|
||||
|
||||
7. Create a Kafka routine load task named test1 for the example_tbl of example_db. delete all data key colunms match v3 >100 key columns.
|
||||
7. Create a Kafka routine load task named test1 for the example_tbl of example_db. delete all data key columns match v3 >100 key columns.
|
||||
|
||||
CREATE ROUTINE LOAD example_db.test1 ON example_tbl
|
||||
WITH MERGE
|
||||
|
||||
@ -38,7 +38,7 @@ SnapshotName: The name of the backup
|
||||
DbName: Subordinate database
|
||||
State: Current phase
|
||||
PENDING: The initial state after submitting a job
|
||||
SNAPSHOTING: In the execution snapshot
|
||||
SNAPSHOTTING: In the execution snapshot
|
||||
UPLOAD_SNAPSHOT: Snapshot completed, ready for upload
|
||||
UPLOADING: Snapshot uploading
|
||||
SAVE_META: Save job meta-information as a local file
|
||||
@ -50,7 +50,7 @@ CreateTime: Task submission time
|
||||
Snapshot Finished Time: Snapshot completion time
|
||||
Upload Finished Time: Snapshot Upload Completion Time
|
||||
FinishedTime: Job End Time
|
||||
Unfinished Tasks: The unfinished sub-task ID is displayed in the SNAP HOTING and UPLOADING phases
|
||||
Unfinished Tasks: The unfinished sub-task ID is displayed in the SNAPSHOTTING and UPLOADING phases
|
||||
Status: Display failure information if the job fails
|
||||
Timeout: Job timeout, per second
|
||||
|
||||
|
||||
@ -39,11 +39,11 @@ Timestamp: Time version of backup to be restored
|
||||
DbName: Subordinate database
|
||||
State: Current phase
|
||||
PENDING: The initial state after submitting a job
|
||||
SNAPSHOTING: In the execution snapshot
|
||||
SNAPSHOTTING: In the execution snapshot
|
||||
DOWNLOAD: The snapshot is complete, ready to download the snapshot in the warehouse
|
||||
DOWNLOADING: Snapshot Download
|
||||
COMMIT: Snapshot download completed, ready to take effect
|
||||
COMMITING: In force
|
||||
COMMITTING: In force
|
||||
FINISHED: Operation Successful
|
||||
CANCELLED: Job Failure
|
||||
AllowLoad: Is import allowed on recovery (currently not supported)
|
||||
@ -54,7 +54,7 @@ MetaPreparedTime: Metadata Readiness Completion Time
|
||||
Snapshot Finished Time: Snapshot completion time
|
||||
Download Finished Time: Snapshot download completion time
|
||||
FinishedTime: Job End Time
|
||||
Unfinished Tasks: The unfinished sub-task ID is displayed in the SNAP HOTING, DOWNLOADING, and COMMITING phases
|
||||
Unfinished Tasks: The unfinished sub-task ID is displayed in the SNAPSHOTTING, DOWNLOADING, and COMMITTING phases
|
||||
Status: Display failure information if the job fails
|
||||
Timeout: Job timeout, per second
|
||||
|
||||
|
||||
@ -175,11 +175,11 @@ Where url is the url given by ErrorURL.
|
||||
|
||||
```Curl --location-trusted -u root -H "label:123" -H "where: k1=20180601" -T testData http://host:port/api/testDb/testTbl/_stream_load```
|
||||
|
||||
3. load the data from the local file 'testData' into the 'testTbl' table in the database 'testDb', allowing a 20% error rate (user is in defalut_cluster)
|
||||
3. load the data from the local file 'testData' into the 'testTbl' table in the database 'testDb', allowing a 20% error rate (user is in default_cluster)
|
||||
|
||||
```Curl --location-trusted -u root -H "label:123" -H "max_filter_ratio:0.2" -T testData http://host:port/api/testDb/testTbl/_stream_load```
|
||||
|
||||
4. load the data from the local file 'testData' into the 'testTbl' table in the database 'testDb', allow a 20% error rate, and specify the column name of the file (user is in defalut_cluster)
|
||||
4. load the data from the local file 'testData' into the 'testTbl' table in the database 'testDb', allow a 20% error rate, and specify the column name of the file (user is in default_cluster)
|
||||
|
||||
```Curl --location-trusted -u root -H "label:123" -H "max_filter_ratio:0.2" -H "columns: k2, k1, v1" -T testData http://host:port/api/testDb/testTbl/_stream_load```
|
||||
|
||||
@ -187,7 +187,7 @@ Where url is the url given by ErrorURL.
|
||||
|
||||
```Curl --location-trusted -u root -H "label:123" -H "max_filter_ratio:0.2" -H "partitions: p1, p2" -T testData http://host:port/api/testDb/testTbl/stream_load```
|
||||
|
||||
6. load using streaming mode (user is in defalut_cluster)
|
||||
6. load using streaming mode (user is in default_cluster)
|
||||
|
||||
```Seq 1 10 | awk '{OFS="\t"}{print $1, $1 * 10}' | curl --location-trusted -u root -T - http://host:port/api/testDb/testTbl/_stream_load```
|
||||
|
||||
|
||||
Reference in New Issue
Block a user