On Jobs
- Only the latest 100 jobs are displayed on DLI's SparkUI.
- A maximum of 1,000 job results can be displayed on the console. To view more or all jobs, export the job data to OBS.
- To export job run logs, you must have the permission to access OBS buckets. You need to configure a DLI job bucket on the Global Configuration > Project page in advance.
- The View Log button is not available for synchronization jobs and jobs running on the default queue.
- Only Spark jobs support custom images.
- An elastic resource pool supports a maximum of 32,000 CUs.
For details about job constraints, see Job Management.
On Queues
For more constraints on using a DLI queue, see Queue Overview.
On Elastic Resource Pools
- The region of an elastic resource pool cannot be changed.
- Jobs of Flink 1.10 or later can run in elastic resource pools.
- The network segment of an elastic resource pool cannot be changed after being set.
- You can only view the scaling history of resource pools in the last 30 days.
- Elastic resource pools cannot access the Internet.
For more constraints on elastic resource pools, see Elastic Resource Pool Overview.
On Resources
- Database
- default is the database built in DLI. You cannot create a database named default.
- DLI supports a maximum of 50 databases.
- Table
- DLI supports a maximum of 5,000 tables.
- DLI supports the following table types:
- MANAGED: Data is stored in a DLI table.
- EXTERNAL: Data is stored in an OBS table.
- View: A view can only be created using SQL statements.
- Datasource table: The table type is also EXTERNAL.
- You cannot specify a storage path when creating a DLI table.
- Data import
- Only OBS data can be imported to DLI or OBS.
- You can import data in CSV, Parquet, ORC, JSON, or Avro format from OBS to tables created on DLI.
- To import data in CSV format to a partitioned table, place the partition column in the last column of the data source.
- The encoding format of imported data can only be UTF-8.
- Data export
- Data in DLI tables (whose table type is MANAGED) can only be exported to OBS buckets, and the export path must contain a folder.
- The exported file is in JSON format, and the text format can only be UTF-8.
- Data can be exported across accounts. That is, after account B authorizes account A, account A has the permission to read the metadata and permission information of account B's OBS bucket as well as the read and write permissions on the path. Account A can export data to the OBS path of account B.
- Package
- A package can be deleted, but a package group cannot be deleted.
- The following types of packages can be uploaded:
- JAR: JAR file
- PyFile: User Python file
- File: User file
- ModelFile: User AI model file
For details about constraints on resources, see Data Management.
On Enhanced Datasource Connections
- Datasource connections cannot be created for the default queue.
- Flink jobs can directly access DIS, OBS, and SMN data sources without using datasource connections.
- VPC Administrator permissions are required for enhanced connections to use VPCs, subnets, routes, VPC peering connections.
- If you use an enhanced datasource connection, the CIDR block of the elastic resource pool or queue cannot overlap with that of the data source.
- Only queues bound with datasource connections can access datasource tables.
- Datasource tables do not support the preview function.
- When checking the connectivity of datasource connections, the constraints on IP addresses are as follows:
- When checking the connectivity of datasource connections, the constraints on domain names are as follows:
For more constraints on enhanced datasource connections, see Enhanced Datasource Connection Overview.
On Datasource Authentication
- Only Spark SQL and Flink OpenSource SQL 1.12 jobs support datasource authentication.
- DLI supports four types of datasource authentication. Select an authentication type specific to each data source.
- CSS: applies to 6.5.4 or later CSS clusters with the security mode enabled.
- Kerberos: applies to MRS security clusters with Kerberos authentication enabled.
- Kafka_SSL: applies to Kafka with SSL enabled.
- Password: applies to GaussDB(DWS), RDS, DDS, and DCS.
For more constraints on datasource authentication, see Datasource Authentication Introduction.
On SQL Syntax
- Constraints on the SQL syntax:
- You are not allowed to specify a storage path when creating a DLI table using SQL statements.
- Constraints on the size of SQL statements:
- Each SQL statement should contain less than 500,000 characters.
- The size of each SQL statement must be less than 1 MB.
Other
- For details about quota constraints, see Quotas.
- Recommended browsers for logging in to DLI:
- Google Chrome 43.0 or later
- Mozilla Firefox 38.0 or later
- Internet Explorer 9.0 or later