回复:退订

2021-08-20 文章 JasonLee
Hi


退订应该发送到 user-zh-unsubscr...@flink.apache.org 


Best
JasonLee


在2021年08月21日 09:43,牛成 写道:
退订

退订

2021-08-20 文章 牛成
退订

退订

2021-08-20 文章 牛成

退订

退订

2021-08-20 文章 牛成




Re:Re:Re:flink checkpoint 到hdfs 报错

2021-08-20 文章 东东
集群里的机器互相访问配的是内网地址呗,你这得开内网访问...


在 2021-08-20 18:56:58,"杨帅统"  写道:
>
>
>
>
>
>
>test.gl.cdh.node1 对应的是远程服务器外网地址 139.9.132.*
>192.168.0.32:9866是139.9.132.*机器的同一内网下的另一台内网地址   为啥会返回内网地址啊。。。
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>在 2021-08-20 18:28:34,"东东"  写道:
>>这不很清楚么,连 192.168.0.32:9866 超时啊
>>
>>
>>
>>
>>在 2021-08-20 18:13:10,"杨帅统"  写道:
>>>// 开启checkpoint
>>>env.enableCheckpointing(5000L, CheckpointingMode.EXACTLY_ONCE);
>>>
>>>env.getCheckpointConfig().setCheckpointStorage("hdfs://test.gl.cdh.node1:8020/flink/flink-cdc-demo");
>>>System.setProperty("HADOOP_USER_NAME", "root");
>>>
>>>
>>>报错信息如下:
>>>org.apache.hadoop.net.ConnectTimeoutException: 6 millis timeout while 
>>>waiting for channel to be ready for connect. ch : 
>>>java.nio.channels.SocketChannel[connection-pending remote=/192.168.0.32:9866]
>>>at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:534)
>>>at 
>>>org.apache.hadoop.hdfs.DataStreamer.createSocketForPipeline(DataStreamer.java:253)
>>>at 
>>>org.apache.hadoop.hdfs.DataStreamer.createBlockOutputStream(DataStreamer.java:1725)
>>>at 
>>>org.apache.hadoop.hdfs.DataStreamer.nextBlockOutputStream(DataStreamer.java:1679)
>>>at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:716)
>>>2021-08-20 18:11:49 WARN  DataStreamer:1683 - Abandoning 
>>>BP-1233461189-192.168.0.27-1615278916172:blk_1074013541_272740
>>>2021-08-20 18:11:50 WARN  DataStreamer:1688 - Excluding datanode 
>>>DatanodeInfoWithStorage[192.168.0.32:9866,DS-f45126b3-f020-473f-b25f-1b37f8540eb7,DISK]
>>>2021-08-20 18:11:50 WARN  DataStreamer:826 - DataStreamer Exception
>>>org.apache.hadoop.ipc.RemoteException(java.io.IOException): File 
>>>/flink/flink-cdc-demo/13714486ceb74d650ba104df7b202920/chk-1/_metadata could 
>>>only be written to 0 of the 1 minReplication nodes. There are 3 datanode(s) 
>>>running and 3 node(s) are excluded in this operation.
>>>at 
>>>org.apache.hadoop.hdfs.server.blockmanagement.BlockManager.chooseTarget4NewBlock(BlockManager.java:2102)
>>>at 
>>>org.apache.hadoop.hdfs.server.namenode.FSDirWriteFileOp.chooseTargetForNewBlock(FSDirWriteFileOp.java:294)
>>>at 
>>>org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:2673)


Re:Re:flink checkpoint 到hdfs 报错

2021-08-20 文章 杨帅统






test.gl.cdh.node1 对应的是远程服务器外网地址 139.9.132.*
192.168.0.32:9866是139.9.132.*机器的同一内网下的另一台内网地址   为啥会返回内网地址啊。。。














在 2021-08-20 18:28:34,"东东"  写道:
>这不很清楚么,连 192.168.0.32:9866 超时啊
>
>
>
>
>在 2021-08-20 18:13:10,"杨帅统"  写道:
>>// 开启checkpoint
>>env.enableCheckpointing(5000L, CheckpointingMode.EXACTLY_ONCE);
>>
>>env.getCheckpointConfig().setCheckpointStorage("hdfs://test.gl.cdh.node1:8020/flink/flink-cdc-demo");
>>System.setProperty("HADOOP_USER_NAME", "root");
>>
>>
>>报错信息如下:
>>org.apache.hadoop.net.ConnectTimeoutException: 6 millis timeout while 
>>waiting for channel to be ready for connect. ch : 
>>java.nio.channels.SocketChannel[connection-pending remote=/192.168.0.32:9866]
>>at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:534)
>>at 
>>org.apache.hadoop.hdfs.DataStreamer.createSocketForPipeline(DataStreamer.java:253)
>>at 
>>org.apache.hadoop.hdfs.DataStreamer.createBlockOutputStream(DataStreamer.java:1725)
>>at 
>>org.apache.hadoop.hdfs.DataStreamer.nextBlockOutputStream(DataStreamer.java:1679)
>>at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:716)
>>2021-08-20 18:11:49 WARN  DataStreamer:1683 - Abandoning 
>>BP-1233461189-192.168.0.27-1615278916172:blk_1074013541_272740
>>2021-08-20 18:11:50 WARN  DataStreamer:1688 - Excluding datanode 
>>DatanodeInfoWithStorage[192.168.0.32:9866,DS-f45126b3-f020-473f-b25f-1b37f8540eb7,DISK]
>>2021-08-20 18:11:50 WARN  DataStreamer:826 - DataStreamer Exception
>>org.apache.hadoop.ipc.RemoteException(java.io.IOException): File 
>>/flink/flink-cdc-demo/13714486ceb74d650ba104df7b202920/chk-1/_metadata could 
>>only be written to 0 of the 1 minReplication nodes. There are 3 datanode(s) 
>>running and 3 node(s) are excluded in this operation.
>>at 
>>org.apache.hadoop.hdfs.server.blockmanagement.BlockManager.chooseTarget4NewBlock(BlockManager.java:2102)
>>at 
>>org.apache.hadoop.hdfs.server.namenode.FSDirWriteFileOp.chooseTargetForNewBlock(FSDirWriteFileOp.java:294)
>>at 
>>org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:2673)


Re:flink checkpoint 到hdfs 报错

2021-08-20 文章 东东
这不很清楚么,连 192.168.0.32:9866 超时啊




在 2021-08-20 18:13:10,"杨帅统"  写道:
>// 开启checkpoint
>env.enableCheckpointing(5000L, CheckpointingMode.EXACTLY_ONCE);
>
>env.getCheckpointConfig().setCheckpointStorage("hdfs://test.gl.cdh.node1:8020/flink/flink-cdc-demo");
>System.setProperty("HADOOP_USER_NAME", "root");
>
>
>报错信息如下:
>org.apache.hadoop.net.ConnectTimeoutException: 6 millis timeout while 
>waiting for channel to be ready for connect. ch : 
>java.nio.channels.SocketChannel[connection-pending remote=/192.168.0.32:9866]
>at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:534)
>at 
>org.apache.hadoop.hdfs.DataStreamer.createSocketForPipeline(DataStreamer.java:253)
>at 
>org.apache.hadoop.hdfs.DataStreamer.createBlockOutputStream(DataStreamer.java:1725)
>at 
>org.apache.hadoop.hdfs.DataStreamer.nextBlockOutputStream(DataStreamer.java:1679)
>at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:716)
>2021-08-20 18:11:49 WARN  DataStreamer:1683 - Abandoning 
>BP-1233461189-192.168.0.27-1615278916172:blk_1074013541_272740
>2021-08-20 18:11:50 WARN  DataStreamer:1688 - Excluding datanode 
>DatanodeInfoWithStorage[192.168.0.32:9866,DS-f45126b3-f020-473f-b25f-1b37f8540eb7,DISK]
>2021-08-20 18:11:50 WARN  DataStreamer:826 - DataStreamer Exception
>org.apache.hadoop.ipc.RemoteException(java.io.IOException): File 
>/flink/flink-cdc-demo/13714486ceb74d650ba104df7b202920/chk-1/_metadata could 
>only be written to 0 of the 1 minReplication nodes. There are 3 datanode(s) 
>running and 3 node(s) are excluded in this operation.
>at 
>org.apache.hadoop.hdfs.server.blockmanagement.BlockManager.chooseTarget4NewBlock(BlockManager.java:2102)
>at 
>org.apache.hadoop.hdfs.server.namenode.FSDirWriteFileOp.chooseTargetForNewBlock(FSDirWriteFileOp.java:294)
>at 
>org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:2673)


flink checkpoint 到hdfs 报错

2021-08-20 文章 杨帅统
// 开启checkpoint
env.enableCheckpointing(5000L, CheckpointingMode.EXACTLY_ONCE);

env.getCheckpointConfig().setCheckpointStorage("hdfs://test.gl.cdh.node1:8020/flink/flink-cdc-demo");
System.setProperty("HADOOP_USER_NAME", "root");


报错信息如下:
org.apache.hadoop.net.ConnectTimeoutException: 6 millis timeout while 
waiting for channel to be ready for connect. ch : 
java.nio.channels.SocketChannel[connection-pending remote=/192.168.0.32:9866]
at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:534)
at 
org.apache.hadoop.hdfs.DataStreamer.createSocketForPipeline(DataStreamer.java:253)
at 
org.apache.hadoop.hdfs.DataStreamer.createBlockOutputStream(DataStreamer.java:1725)
at 
org.apache.hadoop.hdfs.DataStreamer.nextBlockOutputStream(DataStreamer.java:1679)
at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:716)
2021-08-20 18:11:49 WARN  DataStreamer:1683 - Abandoning 
BP-1233461189-192.168.0.27-1615278916172:blk_1074013541_272740
2021-08-20 18:11:50 WARN  DataStreamer:1688 - Excluding datanode 
DatanodeInfoWithStorage[192.168.0.32:9866,DS-f45126b3-f020-473f-b25f-1b37f8540eb7,DISK]
2021-08-20 18:11:50 WARN  DataStreamer:826 - DataStreamer Exception
org.apache.hadoop.ipc.RemoteException(java.io.IOException): File 
/flink/flink-cdc-demo/13714486ceb74d650ba104df7b202920/chk-1/_metadata could 
only be written to 0 of the 1 minReplication nodes. There are 3 datanode(s) 
running and 3 node(s) are excluded in this operation.
at 
org.apache.hadoop.hdfs.server.blockmanagement.BlockManager.chooseTarget4NewBlock(BlockManager.java:2102)
at 
org.apache.hadoop.hdfs.server.namenode.FSDirWriteFileOp.chooseTargetForNewBlock(FSDirWriteFileOp.java:294)
at 
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:2673)
at 
org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.addBlock(NameNodeRpcServer.java:872)
at 
org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.addBlock(ClientNamenodeProtocolServerSideTranslatorPB.java:550)
at 
org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java)
at 
org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:523)
at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:991)
at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:869)
at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:815)
at java.security.AccessController.doPrivileged(Native Method)
at javax.security.auth.Subject.doAs(Subject.java:422)
at 
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1875)
at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2675)


at org.apache.hadoop.ipc.Client.getRpcResponse(Client.java:1499)
at org.apache.hadoop.ipc.Client.call(Client.java:1445)
at org.apache.hadoop.ipc.Client.call(Client.java:1355)
at 
org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:228)
at 
org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:116)
at com.sun.proxy.$Proxy184.addBlock(Unknown Source)
at 
org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.addBlock(ClientNamenodeProtocolTranslatorPB.java:497)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at 
org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:422)
at 
org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invokeMethod(RetryInvocationHandler.java:165)
at 
org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invoke(RetryInvocationHandler.java:157)
at 
org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invokeOnce(RetryInvocationHandler.java:95)
at 
org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:359)
at com.sun.proxy.$Proxy185.addBlock(Unknown Source)
at org.apache.hadoop.hdfs.DFSOutputStream.addBlock(DFSOutputStream.java:1085)
at 
org.apache.hadoop.hdfs.DataStreamer.locateFollowingBlock(DataStreamer.java:1865)
at 
org.apache.hadoop.hdfs.DataStreamer.nextBlockOutputStream(DataStreamer.java:1668)
at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:716)
2021-08-20 18:11:50 WARN  JobMaster:121 - Error while processing 
AcknowledgeCheckpoint message
org.apache.flink.runtime.checkpoint.CheckpointException: Could not finalize the 
pending checkpoint 1. Failure reason: Failure to finalize checkpoint.
at 
org.apache.flink.runtime.checkpoint.CheckpointCoordinator.completePendingCheckpoint(CheckpointCoordinator.java:1199)
at 
org.apache.flink.runtime.checkpoint.CheckpointCoordinator.receiveAcknowledgeMessage(CheckpointCoordinator.java:1072)
at 

Re:Re: cumulate累加函数输出问题

2021-08-20 文章 李航飞



你好:
我使用的场景是要实时统计一天的数据,在小窗口进行即时输出,延迟不能太高,cumulate window符合要求,tumble window 延迟太高了。
在 2021-08-20 16:01:57,"Caizhi Weng"  写道:
>Hi!
>
>你可能想要的是 tumble window 而不是 cumulate window。
>
>李航飞  于2021年8月20日周五 下午3:26写道:
>
>> 能不能通过设置,每分钟只输出 当前小窗口中计算数据的结果,之前计算的结果不再输出
>>
>>
>> 目前测试发现,输入的数据越多,到下次输出的数据也会越来越多,
>> 不同窗口的计算结果,都会再下次窗口中输出,
>>
>>


Re: cumulate累加函数输出问题

2021-08-20 文章 Caizhi Weng
Hi!

你可能想要的是 tumble window 而不是 cumulate window。

李航飞  于2021年8月20日周五 下午3:26写道:

> 能不能通过设置,每分钟只输出 当前小窗口中计算数据的结果,之前计算的结果不再输出
>
>
> 目前测试发现,输入的数据越多,到下次输出的数据也会越来越多,
> 不同窗口的计算结果,都会再下次窗口中输出,
>
>


Re: 怎么才能取消订阅呢

2021-08-20 文章 changfeng
To remove your address from the list, send a message to:
  mailto:user-zh-unsubscr...@flink.apache.org>>

> 2021年8月20日 下午3:36,18221112048 <18221112...@163.com> 写道:
> 
> 



怎么才能取消订阅呢

2021-08-20 文章 18221112048



spring????flinkclient????jar??????????

2021-08-20 文章 ????
??  
spring??jarjarflink??flinkclientjar??spring??jar??webjar??flinkjar??web??


java.util.zip.ZipException: invalid distance too far back
at java.util.zip.InflaterInputStream.read(InflaterInputStream.java:164)
at 
org.springframework.boot.loader.jar.ZipInflaterInputStream.read(ZipInflaterInputStream.java:52)
at java.io.FilterInputStream.read(FilterInputStream.java:107)
at java.util.Properties$LineReader.readLine(Properties.java:435)
at java.util.Properties.load0(Properties.java:353)
at java.util.Properties.load(Properties.java:341)
at 
org.apache.flink.api.java.utils.ParameterTool.fromPropertiesFile(ParameterTool.java:137)
at com.chinaunicom.streamjob.job.HawkeyeTask.main(HawkeyeTask.java:52)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at 
org.apache.flink.client.program.PackagedProgram.callMainMethod(PackagedProgram.java:349)
at 
org.apache.flink.client.program.PackagedProgram.invokeInteractiveModeForExecution(PackagedProgram.java:219)
at 
org.apache.flink.client.program.PackagedProgramUtils.getPipelineFromProgram(PackagedProgramUtils.java:158)
at 
org.apache.flink.client.program.PackagedProgramUtils.createJobGraph(PackagedProgramUtils.java:82)
at 
org.apache.flink.client.program.PackagedProgramUtils.createJobGraph(PackagedProgramUtils.java:117)
at 
com.chinaunicom.dataengine.graph.JobManager.submitJarJob(JobManager.java:422)
at 
org.springblade.modules.dstream.service.impl.JobServiceImpl.submit(JobServiceImpl.java:295)





java.util.zip.ZipException: invalid block type
at java.util.zip.InflaterInputStream.read(InflaterInputStream.java:164)
at 
org.springframework.boot.loader.jar.ZipInflaterInputStream.read(ZipInflaterInputStream.java:52)
at java.io.FilterInputStream.read(FilterInputStream.java:107)
at java.util.Properties$LineReader.readLine(Properties.java:435)
at java.util.Properties.load0(Properties.java:353)
at java.util.Properties.load(Properties.java:341)
at 
org.apache.flink.api.java.utils.ParameterTool.fromPropertiesFile(ParameterTool.java:137)
at 
com.chinaunicom.streamjob.job.CbssCdrCheatTask.main(CbssCdrCheatTask.java:54)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at 
org.apache.flink.client.program.PackagedProgram.callMainMethod(PackagedProgram.java:349)
at 
org.apache.flink.client.program.PackagedProgram.invokeInteractiveModeForExecution(PackagedProgram.java:219)
org.apache.flink.client.program.ProgramInvocationException: The program 
plan could not be fetched - the program aborted pre-maturely.
System.out: (none)

cumulate累加函数输出问题

2021-08-20 文章 李航飞
能不能通过设置,每分钟只输出 当前小窗口中计算数据的结果,之前计算的结果不再输出


目前测试发现,输入的数据越多,到下次输出的数据也会越来越多,
不同窗口的计算结果,都会再下次窗口中输出,



Re: 请问同一个flink history server能够支持多个flink application cluster吗?

2021-08-20 文章 Tony Wei
Hi Chenyu,

這確實是目前尚未解決的一個問題,相關的 jira issue 可以看這 [1]。
jira issue 底下的討論串有提到一個替代方案是:使用 -D\$internal.pipeline.job-id=$(cat
/proc/sys/kernel/random/uuid|tr -d "-") 主動為 application 模式的任務產生隨機的 jobid。
但由於此配置參數屬於 flink 內部參數,可能不保證未來任何改動後的向後兼容性,請謹慎考慮後再使用。

[1] https://issues.apache.org/jira/browse/FLINK-19358


Chenyu Zheng  於 2021年8月20日 週五 下午12:16寫道:

> History Server的API也是使用jobid作为区分
>
>   *   /config
>   *   /jobs/overview
>   *   /jobs/
>   *   /jobs//vertices
>   *   /jobs//config
>   *   /jobs//exceptions
>   *   /jobs//accumulators
>   *   /jobs//vertices/
>   *   /jobs//vertices//subtasktimes
>   *   /jobs//vertices//taskmanagers
>   *   /jobs//vertices//accumulators
>   *   /jobs//vertices//subtasks/accumulators
>   *   /jobs//vertices//subtasks/
>   *
>  /jobs//vertices//subtasks//attempts/
>   *
>  
> /jobs//vertices//subtasks//attempts//accumulators
>   *   /jobs//plan
>
>
> From: Chenyu Zheng 
> Reply-To: "user-zh@flink.apache.org" 
> Date: Friday, August 20, 2021 at 11:43 AM
> To: "user-zh@flink.apache.org" 
> Subject: 请问同一个flink history server能够支持多个flink application cluster吗?
>
> 您好,
>
> 我们目前在k8s上以flink application模式运行作业,现在希望部署一个history
> server方便debug。但是根据文档,flink
> historyserver貌似只支持单个cluster下不同job的使用方法,如果存在多个cluster,相同的jobID将会出现错误。
>
> 请问对于多个application cluster,history使用的最佳姿势是什么样的?
>
> 谢谢[cid:image001.png@01D795B8.6430A670]
>