

Configuration Item | Description | |
Task Name | Name of the data processing task, for example: my_transform. | |
Enabling Status | Task start/stop, default start. | |
Preprocessing Data | Turn on the switch. The feature entry for preprocessing: Entry 1: Toggle on the Preprocessing Data switch when creating a data processing task. Entry 2: You can also click Data Processing at the bottom of the Collection Configuration page to enter the preprocessing data editing page. | |
Log Topic | Specify the log topic to write pre-processing results to. | |
Associate external data | Add an external data source, which can be used for dimension table join scenarios. Currently, only Tencent Cloud MySQL is supported. See the res_rds_mysql function. Region: The region where the cloud MySQL instance is located TencentDB for MySQL instance: Please select in the pull-down menu Username/Password: Enter your database username/password. Data processing only requires query permissions without requiring edit or delete permissions. For configuration methods, see Modifying Account Permissions for TencentDB for MySQL. Securely store your account credentials and avoid disclosure. Alias: The alias for your MySQL instance, which you will use as a parameter in res_rds_mysql. | |
Data processing service log | The operation logs of data processing tasks are stored in the cls_service_log service log topic (free). The alarm feature in the monitoring dashboard for data processing tasks depends on this log topic and is enabled by default. | |
Upload Processing Failure Logs | When enabled, logs that failed to be processed will write into the target topic. When turned off, this option will drop processing-failed logs. | |
Field Name in Processing Failure Logs | If you choose to write processing-failed logs to the target log topic, the failure logs will be stored in this field, with the field name defaulting to ETLParseFailure. | |
Advanced Settings | Add environment variable: Add environment variables for the data processing task runtime. For example, add a pair of variables with name ENV_MYSQL_INTERVAL and value 300. Then you can use refresh_interval=ENV_MYSQL_INTERVAL in the res_rds_mysql function, and the task will resolve to refresh_interval=300. | |


Configuration Item | Description |
Task Name | Name of the data processing task, for example: my_transform. |
Enabling Status | Task start/stop, default start. |
Preprocessing Data | Turn off the switch. |
Source Log Topic | Data source for the data processing task. |
Associate external data | Add an external data source for dimension table join scenarios. Currently, only Tencent Cloud MySQL is supported. See the res_rds_mysql function. Region: The region where the cloud MySQL instance is located. TencentDB for MySQL Instance: Please select in the pull-down menu. Username/Password: Enter your database username/password. Data processing only requires query permissions without requiring edit or delete permissions. For configuration methods, see Modifying Account Permissions for TencentDB for MySQL. Securely store your account credentials and avoid disclosure. Alias: The alias for your MySQL instance, which you will use as a parameter in res_rds_mysql. |
Processing Time Range | Specify the log scope for data processing. Note: Only process data within the log topic's lifecycle. |
Target Log Topic | Select fixed log topic: Log topic: destination bucket for inventory output of data processing, configured as one or multiple. Target topic ownership: You can select log topics under the current account/other primary accounts. Processed results are written to the log topic of the current account. Processed results are written to the log topic of another primary account . For example, when source log topics from Account A are processed and written to Account B's log topic, Account B must configure the access role in CAM (Cloud Access Management). After configuration, Account A enters the role ARN and external ID in the CLS console to enable writing processed results to Account B's log topic. The steps to configure the role are as follows:1. Create a role. Account B logs in to CAM and goes to the Role Management page. 1.1 Create a new cross-account access policy. Policy name, for example: cross_account. The policy syntax is as follows: Note: In the example, authorization follows the principle of least privilege. The resource is configured to only write processed results to Account B's (100012345678) log topic in the Guangzhou region (topic ID ab3456-123a-56bc-d789-abc654321). Please authorize according to the actual situation.
1.2 Create a role, set the role carrier to Tencent Cloud Account, set the cloud account type to Other Root Accounts, enter the ID of account A, such as 100012345678, select Enable Validation, and configure an external ID, such as Hello123. 1.3 Configure the access policy for the role by selecting the pre-configured policy cross_account (example). 1.4 Save the role, for example: A_ds-cross-account_B. 2. Configure the carrier for the role. In the CAM role list, find A_ds-cross-account_B (example), click on the role, select the following: role carrier > manage carrier > add product service > select CLS, and then click update. It can be seen that the current role has two trusted entities: one is Account A, and the other is cls.cloud.tencent.com (CLS Log Service). 3. Account A logs in to CLS and enters the role ARN and external ID. The two pieces of information need to be provided by account B: Account B finds the role A_ds-cross-account_B (example) in the CAM role list. Click to view the role's RoleArn, for example: qcs::cam::uin/10001234567:roleName/A_ds-cross-account_B. The external ID, such as Hello123, can be seen in the role carrier. Note: When entering the Role ARN and External ID, please ensure no extra spaces are included, as this may result in permission validation failure. Cross-account writes to the target log topic will incur log write traffic fees under Account B. Data processing fees will be billed to Account A. Target name: For example, in the source log topic, output loglevel=warning logs to Log Topic A, loglevel=error logs to Log Topic B, and loglevel=info logs to Log Topic C. You can configure the target names of Log Topic A, B, and C as warning, error, and info. |
Data processing service log | The operation logs of data processing tasks are saved in the cls_service_log service log topic (free of charge). The alarm feature in the data processing task monitoring charts relies on this log topic, which is enabled by default. |
Upload Processing Failure Logs | When enabled, logs that failed to be processed will write into the target topic. When turned off, this option will drop processing-failed logs. |
Field Name in Processing Failure Logs | If you choose to write processing-failed logs to the target log topic, the failure logs will be stored in this field, with the field name defaulting to ETLParseFailure. |
Advanced Settings | Add environment variable: Add environment variables for the data processing task runtime. For example, add a pair of variables with name ENV_MYSQL_INTERVAL and value 300. Then you can use refresh_interval=ENV_MYSQL_INTERVAL in the res_rds_mysql function, and the task will resolve to refresh_interval=300. |
Configuration Item | Description |
Task Name | Name of the data processing task, for example: my_transform. |
Enabling Status | Task start/stop, default start. |
Preprocessing Data | Turn off the switch. |
Source Log Topic | Data source for the data processing task. |
Associate external data | Add an external data source, which can be used for dimension table join scenarios. Currently, only Tencent Cloud MySQL is supported. See the res_rds_mysql function. Region: The region where the cloud MySQL instance is located. TencentDB for MySQL Instance: Please select in the pull-down menu. Username: Enter your database username. Password: Enter your database password. Alias: The alias for your MySQL instance, which you will use as a parameter in res_rds_mysql. |
Processing Time Range | Specify the log scope for data processing. Note: Only process data within the log topic's lifecycle. |
Target Log Topic | Select Dynamic Log Topic. No configuration required for target log topic, it will be automatically generated according to the specified field value. |
Overrun handling | When the topic count generated by your data processing task exceeds the product spec, you can choose: Create a fallback logset and log topic, and write logs to the fallback topic (created when creating a task). Fallback logset: auto_undertake_logset, single-region single account next. Fallback topic: auto_undertake_topic_$(data processing task name). For example, if a user creates two data processing tasks etl_A and etl_B, two fallback topics will occur: auto_undertake_topic_etl_A, auto_undertake_topic_etl_B. Discard log data: Discard logs directly without creating a fallback topic. |
Data processing service log | The operation logs of data processing tasks are stored in the cls_service_log service log topic (free). The alarm feature in the monitoring dashboard for data processing tasks depends on this log topic and is enabled by default. |
Upload Processing Failure Logs | When enabled, logs that failed to be processed will write into the target topic. When turned off, this option will drop processing-failed logs. |
Field Name in Processing Failure Logs | If you choose to write processing-failed logs to the target log topic, the failure logs will be stored in this field, with the field name defaulting to ETLParseFailure. |
Advanced Settings | Add environment variable: Add environment variables for the data processing task runtime. For example, add a pair of variables with name ENV_MYSQL_INTERVAL and value 300. Then you can use refresh_interval=ENV_MYSQL_INTERVAL in the res_rds_mysql function, and the task will resolve to refresh_interval=300. |
{"content": "[2021-11-24 11:11:08,232][328495eb-b562-478f-9d5d-3bf7e][INFO] curl -H 'Host: ' http://abc.com:8080/pc/api -d {\\"version\\": \\"1.0\\",\\"user\\": \\"CGW\\",\\"password\\": \\"123\\"}"}
Dialogue Turn | User Question | AI Assistant Reply |
First-round dialogue | Structure this log |
|
Second-round dialogue | The content is not standard JSON, and using ext_json throws an error. Please first extract the JSON part from the content, then retrieve the nodes from the JSON. |
|
{"level":"INFO","password":"123","requestid":"328495eb-b562-478f-9d5d-3bf7e","time":"2021-11-24 11:11:08,232","user":"CGW","version":"1.0"}



Function Category | Visualization Function Name | Application Scenario |
Extract Key Value | JSON: Extract fields and field values from JSON nodes Separator: Extract field values based on the separator. Users are advised to enter the field name. Regular Expression: Extract field values using regular expressions. User input is required for the field name. | Log Structuring |
Log Processing | Filter Logs: Configure conditions for filtering logs (multiple conditions are in an OR relationship). For example, if field A exists or field B does not exist, filter out the log. Distribute Logs: Configure conditions for distributing logs. If status="error" and message contains "404", distribute to topic A If status="running" and message contains "200", distribute to topic B Retain Logs: Configure conditions for preserving logs. | Delete/Retain Logs |
Field Processing | Delete fields Rename Field | Delete/Rename Field |
Was this page helpful?
You can also Contact sales or Submit a Ticket for help.
Help us improve! Rate your documentation experience in 5 mins.
Feedback