This is an automated email from the ASF dual-hosted git repository.
benjobs pushed a commit to branch dev
in repository
https://gitbox.apache.org/repos/asf/incubator-streampark-website.git
The following commit(s) were added to refs/heads/dev by this push:
new cc032d8 [Improve] localization platform install (#307)
cc032d8 is described below
commit cc032d89251b7a65cded21ec3254622277f1f8bf
Author: VampireAchao <[email protected]>
AuthorDate: Mon Jan 1 13:39:36 2024 +0800
[Improve] localization platform install (#307)
* [Improve] localization platform install
* [revert] revert yarn.lock
---------
Co-authored-by: VampireAchao <[email protected]>
---
docs/user-guide/11-platformInstall.md | 166 +++++++++++++++++-----------------
1 file changed, 84 insertions(+), 82 deletions(-)
diff --git a/docs/user-guide/11-platformInstall.md
b/docs/user-guide/11-platformInstall.md
index 0bbd936..f0f5d63 100644
--- a/docs/user-guide/11-platformInstall.md
+++ b/docs/user-guide/11-platformInstall.md
@@ -1,111 +1,113 @@
-# 引言
-## 目的和范围
-> 本文档旨在帮助用户完成 Apache StreamPark 的安装和初步配置
+# Introduction
+## Purpose and Scope
+> This document aims to assist users in installing and initially configuring
Apache StreamPark.
-## 目标受众
-> 面向需要在其系统中部署 Apache StreamPark 的系统开发和运维人员
+## Target Audience
+> Intended for system developers and operators who need to deploy Apache
StreamPark in their systems.
-# 系统要求
->
参考:[https://streampark.apache.org/docs/user-guide/deployment#environmental-requirements](https://streampark.apache.org/docs/user-guide/deployment#environmental-requirements)
+# System Requirements
+> Reference:
[https://streampark.apache.org/docs/user-guide/deployment#environmental-requirements](https://streampark.apache.org/docs/user-guide/deployment#environmental-requirements)
-## 硬件要求
-> - 本文档使用Linux:3.10.0-957.el7.x86_6
+## Hardware Requirements
+> - This document uses Linux: 3.10.0-957.el7.x86_6

-## 软件要求
+## Software Requirements
Notes:
-1. **单纯安装StreamPark,可忽略hadoop**
-2. 若采用 yarn application 模式 执行flink作业,需要hadoop
+1. **For installing StreamPark alone, Hadoop can be ignored.**
+2. If using yarn application mode for executing Flink jobs, Hadoop is required.
> - JDK : 1.8+
> - MySQL : 5.6+
> - Flink : 1.12.0+
> - Hadoop : 2.7.0+
> - StreamPark : 2.0.0+
-本文档采用的软件版本信息
-> - **JDK:1.8.0_181**
+Software versions used in this document:
+> - **JDK: 1.8.0_181**
> - **MySQL: 5.7.26**
> - **Flink : 1.14.3-scala_2.12**
> - **Hadoop : 3.2.1**
-主要组件依赖关系:
+Main component dependencies:

-# 安装前准备
-> JDK、MYSQL、HADOOP需用户自行查阅资料安装。
+# Pre-installation Preparation
+> JDK, MYSQL, HADOOP need to be installed by users themselves.
-## 下载&&配置flink
-> 下载flink
+## Download & Configure Flink
+> Download Flink
```bash
cd /usr/local
wget
https://archive.apache.org/dist/flink/flink-1.14.3/flink-1.14.3-bin-scala_2.12.tgz
```
-> 解压
+> Unzip
```bash
tar -zxvf flink-1.14.3-bin-scala_2.12.tgz
```
-> 重命名
+> Rename
```bash
mv flink-1.14.3 flink
```
-> 配置flink环境变量
+> Configure Flink environment variables
```bash
-# 配置环境变量(vim ~/.bashrc),加入以下内容
+# Set environment variables (vim ~/.bashrc), add the following content
export FLINK_HOME=/usr/local/flink
export PATH=$FLINK_HOME/bin:$PATH
-# 生效环境变量配置
+# Apply environment variable configuration
source ~/.bashrc
-# 测试(出现:'Version: 1.14.3, Commit ID: 98997ea',则说明配置成功)
+# Test (If it shows: 'Version: 1.14.3, Commit ID: 98997ea', it means
configuration is successful)
flink -v
```

-## 引入MySQL依赖包
-> 原因:**由于Apache 2.0许可与Mysql Jdbc驱动许可的不兼容,用户需要自行下载驱动jar包并放在
$STREAMPARK_HOME/lib 中,推荐使用8.x版本。**
-> 驱动包版本:mysql-connector-java-8.0.28.jar
+## Introduce MySQL Dependency Package
+> Reason: **Due to incompatibility between Apache 2.0 license and Mysql Jdbc
driver license, users need to download the driver jar package themselves and
place it in $STREAMPARK_HOME/lib, 8.x version recommended.**
+> Driver package version: mysql-connector-java-8.0.28.jar
```bash
cp mysql-connector-java-8.0.28.jar /usr/local/streampark/lib
```

-## 下载StreamPark
->
下载URL:[https://dlcdn.apache.org/incubator/streampark/2.0.0/apache-streampark_2.12-2.0.0-incubating-bin.tar.gz](https://dlcdn.apache.org/incubator/streampark/2.0.0/apache-streampark_2.12-2.0.0-incubating-bin.tar.gz)
+## Download StreamPark
+> Download URL:
[https://dlcdn.apache.org/incubator/streampark/2.0.0/apache-streampark_2.12-2.0.0-incubating-bin.tar.gz](https://dlcdn.apache.org/incubator/streampark/2.0.0/apache-streampark_2.12-2.0.0-incubating-bin.tar.gz)
-> 上传
[apache-streampark_2.12-2.0.0-incubating-bin.tar.gz](https://dlcdn.apache.org/incubator/streampark/2.0.0/apache-streampark_2.12-2.0.0-incubating-bin.tar.gz)
至 服务器 /usr/local 路径
+> Upload
[apache-streampark_2.12-2.0.0-incubating-bin.tar.gz](https://dlcdn.apache.org/incubator/streampark/2.0.0/apache-streampark_2.12-2.0.0-incubating-bin.tar.gz)
to the server /usr/local path

-> 解压
+> Unzip
```bash
tar -zxvf apache-streampark_2.12-2.0.0-incubating-bin.tar.gz
```

-# 安装
-## 初始化系统数据
-> **目的:创建StreamPark组件部署依赖的数据库(表),同时将其运行需要的数据提前初始化(比如:web页面的菜单、用户等信息),便于后续操作。**
+# Installation
+## Initialize System Data
+> **Purpose: Create databases (tables) dependent on StreamPark component
deployment, and pre-initialize the data required for its operation (e.g., web
page menus, user information), to facilitate subsequent operations.**
-### 查看执行SteamPark元数据SQL文件
-> 说明:
-> - StreamPark支持MySQL、PostgreSQL、H2
-> - 本次以MySQL为例,PostgreSQL流程基本一致
+### View Execution of SteamPark Metadata SQL File
+> Explanation:
+> - StreamPark supports MySQL, PostgreSQL, H2
+> - This document uses MySQL as an example; the PostgreSQL process is
basically the same
-> 数据库创建脚本:
/usr/local/apache-streampark_2.12-2.0.0-incubating-bin/script/schema/mysql-schema.sql
+> Database creation script: /usr/local/apache-st
+
+reampark_2.12-2.0.0-incubating-bin/script/schema/mysql-schema.sql

-> 数据库创建脚本:
/usr/local/apache-streampark_2.12-2.0.0-incubating-bin/script/data/mysql-data.sql
+> Database creation script:
/usr/local/apache-streampark_2.12-2.0.0-incubating-bin/script/data/mysql-data.sql

-### 连接MySQL数据库 && 执行初始化脚本
+### Connect to MySQL Database & Execute Initialization Script
```bash
source
/usr/local/apache-streampark_2.12-2.0.0-incubating-bin/script/schema/mysql-schema.sql
```
@@ -115,7 +117,7 @@ source source
/usr/local/apache-streampark_2.12-2.0.0-incubating-bin/script/data
```

-### 查看执行结果
+### View Execution Results
```bash
show databases;
```
@@ -129,98 +131,98 @@ show tables;
```

-## StreamPark配置
-> 目的:配置启动需要的数据源。
-> 配置文件所在路径:/usr/local/streampark/conf
+## StreamPark Configuration
+> Purpose: Configure the data sources needed for startup.
+> Configuration file location: /usr/local/streampark/conf

-### 配置mysql数据源
+### Configure MySQL Data Source
```bash
vim application-mysql.yml
```
-> **username、password、url中的数据库IP/端口号 需要改成用户自己环境的信息**
+> **The database IP/port in username, password, url need to be changed to the
user's own environment information**
```bash
spring:
datasource:
- username: 数据库用户名
- password: 数据库用户密码
+ username: Database username
+ password: Database user password
driver-class-name: com.mysql.cj.jdbc.Driver
- url:
jdbc:mysql://数据库IP地址:数据库端口号/streampark?useSSL=false&useUnicode=true&characterEncoding=UTF-8&allowPublicKeyRetrieval=false&useJDBCCompliantTimezoneShift=true&useLegacyDatetimeCode=false&serverTimezone=GMT%2B8
+ url: jdbc:mysql://Database IP address:Database port
number/streampark?useSSL=false&useUnicode=true&characterEncoding=UTF-8&allowPublicKeyRetrieval=false&useJDBCCompliantTimezoneShift=true&useLegacyDatetimeCode=false&serverTimezone=GMT%2B8
```
-### 配置应用端口、hdfs存储、应用访问密码等
+### Configure Application Port, HDFS Storage, Application Access Password, etc.
```bash
vim application.yml
```
-> **主要的配置项:**
-> 1. **server.port # 【重要】默认的web访问端口号10000,如果有冲突(如: hive服务)可以更改**
-> 2. knife4j.basic.enable # true表示允许访问页面 Swagger API
-> 3. knife4j.basic.password # 访问页面 Swagger API 时需要密码,这样可以提高接口安全性
-> 4. **spring.profiles.activemysql # 【重要】表示系统采用何种数据源,本文档采用mysql **
-> 5. **workspace.remote # 配置工作空间信息**
-> 6. **hadoop-user-name #
如果使用hadoop,该处用户需要保证有操作hdfs的权限,否则会报“org.apache.hadoop.security.AccessControlException:
Permission denied”异常**
-> 7. ldap.password # 系统登陆页面提供了2种登录模式: 用户密码 和 ldap。这里可以配置ldap密码
+> **Key configuration items:**
+> 1. **server.port # 【Important】Default web access port 10000, can be
changed if there is a conflict (e.g., hive service)**
+> 2. knife4j.basic.enable # true means allowing access to Swagger API page
+> 3. knife4j.basic.password # Password required for accessing Swagger API
page, enhancing interface security
+> 4. **spring.profiles.activemysql # 【Important】Indicates which data source
the system uses, this document uses mysql **
+> 5. **workspace.remote # Configure workspace information**
+> 6. **hadoop-user-name # If using hadoop, this user needs to have permission
to operate hdfs, otherwise an
“org.apache.hadoop.security.AccessControlException: Permission denied”
exception will be reported**
+> 7. ldap.password # The system login page offers two login modes: User
password and ldap. Here you can configure ldap password
-> 主要配置示例:
+> Main configuration example:

-> 上传flink job jar
如果过大,可能会导致上传失败,因此可以考虑修改(max-file-size和max-request-size)这两个参数;当然,实际环境还要考虑其他因素:nginx限制等。
+> If the flink job jar is too large, it may fail to upload, so consider
modifying (max-file-size and max-request-size); of course, other factors in the
actual environment should be considered: nginx restrictions, etc.

-> 支持knox配置,有些用户环境私有部署hadoop,可通过knox访问yarn web
-> workspace: 配置工作空间信息(比如:savepoint和checkpoint存储路径等)
+> Supports Knox configuration, some users have privately deployed Hadoop
environments, accessible through Knox
+> workspace: Configure workspace information (e.g., savepoint and checkpoint
storage paths)

> ldap

-### 【可选】配置kerberos
->
背景:企业级hadoop集群环境都有设置安全访问机制,比如kerberos。StreamPark也可配置kerberos,使得flink可通过kerberos认证,向hadoop集群提交作业。
+### 【Optional】Configuring Kerberos
+> Background: Enterprise-level Hadoop cluster environments have set security
access mechanisms, such as Kerberos. StreamPark can also be configured with
Kerberos, allowing Flink to authenticate through Kerberos and submit jobs to
the Hadoop cluster.
-> **修改项如下:**
+> **Modifications are as follows:**
> 1. **security.kerberos.login.enable=true**
-> 2. **security.kerberos.login.principal=实际的principal**
+> 2. **security.kerberos.login.principal=Actual principal**
> 3. **security.kerberos.login.krb5=/etc/krb5.conf**
-> 4. **security.kerberos.login.keytab=实际的keytab文件**
+> 4. **security.kerberos.login.keytab=Actual keytab file**
> 5. **java.security.krb5.conf=/etc/krb5.conf**

-## 启动StreamPark
-## 进入服务器StreamPark安装路径
+## Starting StreamPark
+## Enter the StreamPark Installation Path on the Server
```bash
cd /usr/local/streampark/
```

-## 启动StreamPark服务
+## Start the StreamPark Service
```bash
./bin/startup.sh
```

-> 查看启动日志
-> 目的:确认无报错信息
+> Check the startup logs
+> Purpose: To confirm there are no error messages
```bash
tail -100f log/streampark.out
```

-# 验证安装
+# Verifying the Installation
```bash
-# 页面正常打开,则说明部署成功。
-http://部署streampark服务IP或域名:10000/
+# If the page opens normally, it indicates a successful deployment.
+http://Deployed streampark service IP or domain:10000/
admin/streampark
```
-## 页面正常访问
+## Normal Access to the Page

-## 系统正常登录
+## System Logs in Normally

##
-# 常见问题
+# Common Issues
## Cannot load driver class: com.mysql.cj.jdbc.Driver
-> 原因:缺少mysql驱动包,参见 “3.2. 引入MySQL依赖包”
+> Reason: Missing MySQL driver package, refer to “3.2. Introducing MySQL
Dependency Package”

-# 参考资源
+# Reference Resources
> [https://streampark.apache.org/docs/user-guide/deployment/](https://streampark.apache.org/docs/user-guide/deployment/)