zhongjiajie commented on code in PR #10715:
URL: https://github.com/apache/dolphinscheduler/pull/10715#discussion_r911553180
##########
docs/docs/en/guide/security.md:
##########
@@ -139,37 +135,21 @@ worker.groups=default,test
* When executing a task, the task can be assigned to the specified worker
group, and select the corresponding environment according to the worker group.
Finally, the worker node executes the environment first and then executes the
task.
> Add or update environment
-
- The environment configuration is equivalent to the configuration in the
`dolphinscheduler_env.sh` file.
-
+
Review Comment:
we have to use the related path of img to make it work in our document
##########
docs/docs/en/guide/resource/task-group.md:
##########
@@ -20,37 +20,37 @@ You need to enter the information inside the picture:
- Resource pool size: The maximum number of concurrent task instances allowed.
-#### View Task Group Queue
+### View Task Group Queue
-
+
Click the button to view task group usage information:
-
+
-#### Use of Task Groups
+### Use of Task Groups
**Note**: The usage of task groups is applicable to tasks executed by workers,
such as [switch] nodes, [condition] nodes, [sub_process] and other node types
executed by the master are not controlled by the task group. Let's take the
shell node as an example:
-
+
Regarding the configuration of the task group, all you need to do is to
configure these parts in the red box:
- Task group name: The task group name is displayed on the task group
configuration page. Here you can only see the task group that the project has
permission to access (the project is selected when creating a task group) or
the task group that scope globally (no project is selected when creating a task
group).
- Priority: When there is a waiting resource, the task with high priority will
be distributed to the worker by the master first. The larger the value of this
part, the higher the priority.
-### Implementation Logic of Task Group
+## Implementation Logic of Task Group
-#### Get Task Group Resources:
+### Get Task Group Resources:
Review Comment:
```suggestion
### Get Task Group Resources
```
##########
docs/configs/docsdev.js:
##########
@@ -257,6 +257,10 @@ export default {
{
title: 'Resource',
children: [
+ {
+ title: 'Resources',
+ link:
'/en-us/docs/dev/user_doc/guide/resource/resources_introduction.html'
Review Comment:
I think we should better rename the `resources_introduction.md` to
`intro.md`. It is simply. and make the title to `Introduction`
##########
docs/docs/en/guide/resource/configuration.md:
##########
@@ -92,45 +60,33 @@ yarn.resourcemanager.ha.rm.ids=192.168.xx.xx,192.168.xx.xx
yarn.application.status.address=http://localhost:%s/ds/v1/cluster/apps/%s
# job history status url when application number threshold is reached(default
10000, maybe it was set to 1000)
yarn.job.history.status.address=http://localhost:19888/ds/v1/history/mapreduce/jobs/%s
-
# datasource encryption enable
datasource.encryption.enable=false
-
# datasource encryption salt
datasource.encryption.salt=!@#$%^&*
-
# data quality option
data-quality.jar.name=dolphinscheduler-data-quality-dev-SNAPSHOT.jar
-
#data-quality.error.output.path=/tmp/data-quality-error-data
-
# Network IP gets priority, default inner outer
-
# Whether hive SQL is executed in the same session
support.hive.oneSession=false
-
# use sudo or not, if set true, executing user is tenant user and deploy user
needs sudo permissions;
# if set false, executing user is the deploy user and doesn't need sudo
permissions
sudo.enable=true
-
# network interface preferred like eth0, default: empty
#dolphin.scheduler.network.interface.preferred=
-
# network IP gets priority, default: inner outer
#dolphin.scheduler.network.priority.strategy=default
-
# system env path
#dolphinscheduler.env.path=env/dolphinscheduler_env.sh
-
# development state
development.state=false
-
# rpc port
alert.rpc.port=50052
```
-> **_Note:_**
->
+> **Note:**
+>
> * If only the `api-server/conf/common.properties` file is configured, then
> resource uploading is enabled, but you can not use resources in task. If you
> want to use or execute the files in the workflow you need to configure
> `worker-server/conf/common.properties` too.
> * If you want to use the resource upload function, the deployment user in
> [installation and deployment](../installation/standalone.md) must have
> relevant operation authority.
-> * If you using a Hadoop cluster with HA, you need to enable HDFS resource
upload, and you need to copy the `core-site.xml` and `hdfs-site.xml` under the
Hadoop cluster to `worker-server/conf` and `api-server/conf`, otherwise skip
this copy step.
\ No newline at end of file
+> * If you using a Hadoop cluster with HA, you need to enable HDFS resource
upload, and you need to copy the `core-site.xml` and `hdfs-site.xml` under the
Hadoop cluster to `/opt/dolphinscheduler/conf`, otherwise skip this copy step.
Review Comment:
```suggestion
> * If you using a Hadoop cluster with HA, you need to enable HDFS resource
upload, and you need to copy the `core-site.xml` and `hdfs-site.xml` under the
Hadoop cluster to `worker-server/conf` and `api-server/conf`, otherwise skip
this copy step.
```
##########
docs/docs/en/guide/resource/file-manage.md:
##########
@@ -2,52 +2,39 @@
When third party jars are used in the scheduling process or user defined
scripts are required, these can be created from this page. The types of files
that can be created include: txt, log, sh, conf, py, java and so on. Files can
be edited, renamed, downloaded and deleted.
-
+
-> **_Note:_**
+> **Note:**
>
-> * When you manage files as `admin`, remember to set up `tenant` for `admin`
first.
+> * When you manage files as `admin`, remember to set up `tenant` for `admin`
first.
-## Basic Operations
+- Create a file
+ > The file format supports the following types: txt, log, sh, conf, cfg, py,
java, sql, xml, hql properties.
+
-### Create a File
+- Upload files
-The file format supports the following types: txt, log, sh, conf, cfg, py,
java, sql, xml, hql, properties.
+> Upload file: Click the "`Upload File`" button to upload, drag the file to
the upload area, the file name will be automatically completed with the
uploaded file name.
+
-
+- File view
-### Upload Files
+> For the files that can be viewed, click the file name to view the file
details.
+
-Click the "Upload File" button to upload, drag the file to the upload area,
the file name will be automatically completed with the uploaded file name.
+- Download file
-
+> Click the "`Download`" button in the file list to download the file or click
the "`Download`" button in the upper right corner of the file details to
download the file.
+- File rename
-### View File Content
+
- For the files that can be viewed, click the file name to view the file
details.
+- Delete
+ > File list -> Click the "`Delete`" button to delete the specified file.
+- Re-upload file
-
-
-### Download file
-
-> Click the "Download" button in the file list to download the file or click
the "Download" button in the upper right corner of the file details to download
the file.
-
-### Rename File
-
-
-
-### Delete File
-
-File list -> Click the "Delete" button to delete the specified file.
-
-### Re-upload file
-
-Click the "Re-upload File" button to upload a new file to replace the old
file, drag the file to the re-upload area, the file name will be automatically
completed with the new uploaded file name.
-
-
Review Comment:
It is seem we should not remvoe those section of content
##########
docs/docs/en/guide/resource/resources_introduction.md:
##########
@@ -0,0 +1,2 @@
+# Resources
+The Resource Center is typically used for uploading files, UDF functions, and
task group management. For a stand-alone environment, you can select the local
file directory as the upload folder (this operation does not require Hadoop
deployment). Of course, you can also choose to upload to Hadoop or MinIO
cluster. In this case, you need to have Hadoop (2.6+) or MinIOn and other
related environments.
Review Comment:
good adding, BTW
##########
docs/docs/en/guide/resource/file-manage.md:
##########
@@ -66,10 +53,14 @@ In the workflow definition module of project Manage, create
a new workflow using
- Script: 'sh hello.sh'
- Resource: Select 'hello.sh'
-
+
### View the results
You can view the log results of running the node in the workflow example. The
diagram below:
-
+
+
+
+
+
Review Comment:
```suggestion
```
##########
docs/docs/en/guide/security.md:
##########
@@ -139,37 +135,21 @@ worker.groups=default,test
* When executing a task, the task can be assigned to the specified worker
group, and select the corresponding environment according to the worker group.
Finally, the worker node executes the environment first and then executes the
task.
> Add or update environment
-
- The environment configuration is equivalent to the configuration in the
`dolphinscheduler_env.sh` file.
-
+
> Usage environment
-
- Create a task node in the workflow definition, select the worker group and
the environment corresponding to the worker group. When executing the task, the
Worker will execute the environment first before executing the task.
-
-
-## Cluster Management
-
-> Add or update cluster
-
-- Each process can be related to zero or several clusters to support multiple
environment, now just support k8s.
-
-> Usage cluster
-
-- After creation and authorization, k8s namespaces and processes will
associate clusters. Each cluster will have separate workflows and task
instances running independently.
-
-
Review Comment:
Add why we remove the section of the docs?
##########
docs/docs/en/guide/resource/configuration.md:
##########
@@ -92,45 +60,33 @@ yarn.resourcemanager.ha.rm.ids=192.168.xx.xx,192.168.xx.xx
yarn.application.status.address=http://localhost:%s/ds/v1/cluster/apps/%s
# job history status url when application number threshold is reached(default
10000, maybe it was set to 1000)
yarn.job.history.status.address=http://localhost:19888/ds/v1/history/mapreduce/jobs/%s
-
# datasource encryption enable
datasource.encryption.enable=false
-
# datasource encryption salt
datasource.encryption.salt=!@#$%^&*
-
# data quality option
data-quality.jar.name=dolphinscheduler-data-quality-dev-SNAPSHOT.jar
-
#data-quality.error.output.path=/tmp/data-quality-error-data
-
# Network IP gets priority, default inner outer
-
# Whether hive SQL is executed in the same session
support.hive.oneSession=false
-
# use sudo or not, if set true, executing user is tenant user and deploy user
needs sudo permissions;
# if set false, executing user is the deploy user and doesn't need sudo
permissions
sudo.enable=true
-
# network interface preferred like eth0, default: empty
#dolphin.scheduler.network.interface.preferred=
-
# network IP gets priority, default: inner outer
#dolphin.scheduler.network.priority.strategy=default
-
# system env path
#dolphinscheduler.env.path=env/dolphinscheduler_env.sh
-
# development state
development.state=false
-
# rpc port
alert.rpc.port=50052
```
-> **_Note:_**
->
+> **Note:**
+>
> * If only the `api-server/conf/common.properties` file is configured, then
> resource uploading is enabled, but you can not use resources in task. If you
> want to use or execute the files in the workflow you need to configure
> `worker-server/conf/common.properties` too.
> * If you want to use the resource upload function, the deployment user in
> [installation and deployment](../installation/standalone.md) must have
> relevant operation authority.
-> * If you using a Hadoop cluster with HA, you need to enable HDFS resource
upload, and you need to copy the `core-site.xml` and `hdfs-site.xml` under the
Hadoop cluster to `worker-server/conf` and `api-server/conf`, otherwise skip
this copy step.
\ No newline at end of file
+> * If you using a Hadoop cluster with HA, you need to enable HDFS resource
upload, and you need to copy the `core-site.xml` and `hdfs-site.xml` under the
Hadoop cluster to `/opt/dolphinscheduler/conf`, otherwise skip this copy step.
Review Comment:
should not change this content
##########
docs/docs/en/guide/resource/task-group.md:
##########
@@ -20,37 +20,37 @@ You need to enter the information inside the picture:
- Resource pool size: The maximum number of concurrent task instances allowed.
-#### View Task Group Queue
+### View Task Group Queue
-
+
Click the button to view task group usage information:
-
+
-#### Use of Task Groups
+### Use of Task Groups
**Note**: The usage of task groups is applicable to tasks executed by workers,
such as [switch] nodes, [condition] nodes, [sub_process] and other node types
executed by the master are not controlled by the task group. Let's take the
shell node as an example:
-
+
Regarding the configuration of the task group, all you need to do is to
configure these parts in the red box:
- Task group name: The task group name is displayed on the task group
configuration page. Here you can only see the task group that the project has
permission to access (the project is selected when creating a task group) or
the task group that scope globally (no project is selected when creating a task
group).
- Priority: When there is a waiting resource, the task with high priority will
be distributed to the worker by the master first. The larger the value of this
part, the higher the priority.
-### Implementation Logic of Task Group
+## Implementation Logic of Task Group
-#### Get Task Group Resources:
+### Get Task Group Resources:
The master judges whether the task is configured with a task group when
distributing the task. If the task is not configured, it is normally thrown to
the worker to run; if a task group is configured, it checks whether the
remaining size of the task group resource pool meets the current task operation
before throwing it to the worker for execution. , if the resource pool -1 is
satisfied, continue to run; if not, exit the task distribution and wait for
other tasks to wake up.
-#### Release and Wake Up:
+### Release and Wake Up:
Review Comment:
```suggestion
### Release and Wake Up
```
##########
docs/docs/en/guide/resource/resources_introduction.md:
##########
@@ -0,0 +1,2 @@
+# Resources
+The Resource Center is typically used for uploading files, UDF functions, and
task group management. For a stand-alone environment, you can select the local
file directory as the upload folder (this operation does not require Hadoop
deployment). Of course, you can also choose to upload to Hadoop or MinIO
cluster. In this case, you need to have Hadoop (2.6+) or MinIOn and other
related environments.
Review Comment:
```suggestion
The Resource Center is typically used for uploading files, UDF functions,
and task group management. For a stand-alone environment, you can select the
local file directory as the upload folder (this operation does not require
Hadoop deployment). Of course, you can also choose to upload to Hadoop or MinIO
cluster. In this case, you need to have Hadoop (2.6+) or MinIOn and other
related environments.
```
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
To unsubscribe, e-mail: [email protected]
For queries about this service, please contact Infrastructure at:
[email protected]