[ 
https://issues.apache.org/jira/browse/BAHIR-104?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15959299#comment-15959299
 ] 

Francesco Beneventi commented on BAHIR-104:
-------------------------------------------

[~anntinutj], I fixed the wrong method call before running my tests. You can 
check it in my test repo 
[Github|https://github.com/fbeneventi/bahir/commit/3755ecc56e19ac58fe20089a475253cc0bd19486#diff-f63d51c23efe7ce897bcc35fc52039f7]

Even with the correct API (helper.createPairedStream) I have the same issue.

I'm not really a java/scala programmer but can you help me figure out how to 
address this problem?
After solving this problem I can provide a simple test case to use this API in 
python.

> MQTT Dstream returned by the new multi topic support API is not a pairRDD
> -------------------------------------------------------------------------
>
>                 Key: BAHIR-104
>                 URL: https://issues.apache.org/jira/browse/BAHIR-104
>             Project: Bahir
>          Issue Type: Bug
>          Components: Spark Streaming Connectors
>    Affects Versions: Spark-2.1.0
>            Reporter: Francesco Beneventi
>              Labels: MQTT, SPARK
>
> The new multi topic support API added with [BAHIR-89], when used in pyspark, 
> does not return a Dstream of <topic,message> tuples. 
> Example: 
> In pyspark, when creating a Dstream using the new API ( mqttstream = 
> MQTTUtils.createPairedStream(ssc, brokerUrl, topics) ) the expected contents 
> of mqttstream should be a collections of tuples:
> (topic,message) , (topic,message) , (topic,message) , ...
> Instead, the current content is a flattened list:
> topic, message, topic, message, topic, message, ...
> that is hard to use.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

Reply via email to