Do we have option to say to spark to preserve time stamp while creating
struct.
Regards,
Sudhir
Hello,
I am using createDataframe and passing java row rdd and schema . But it is
changing the time value when I write that data frame to a parquet file.
Can any one help .
Thank you,
Sudhir
> applicable.
>
>
> Thanks.
>
>
> 2017-12-30 2:19 GMT+09:00 sk skk :
>
>> Hi,
>>
>> Do we have an option to write a csv or text file with a custom
>> record/line separator through spark ?
>>
>> I could not find any ref on the api. I have a is
new line character .
Thank you ,
Sk
I have registered a udf with sqlcontext , I am trying to read another
parquet using sqlcontext under same udf it’s throwing null pointer
exception .
Any help how to access sqlcontext inside a udf ?
Regards,
Sk
literal as if I register it as a temp table and fetch that column value and
assigning it to a string it is return a row to string schema and not
getting a literal .
Is there a better way to handle this or how to get a literal value from
temporary table .
Thank you ,
Sk
Can we create a dataframe from a Java pair rdd of String . I don’t have a
schema as it will be a dynamic Json. I gave encoders.string class.
Any help is appreciated !!
Thanks,
SK
this data array which
will be performance impact, do we have any options or better way to handle..
Thanks in advance.
sk
Hi,
I need to parse a json input file where the nested objects take on a
different structure based on the typeId field, as follows:
{ "d":
{ "uid" : "12345"
"contents": [{"info": {"eventId": "event1"}, "typeId": 19}]
}
}
{ "d":
{ "uid" : "56780"
Hi,
I need to access data on S3 from another account and I have been given the
IAM role information to access that S3 bucket. From what I understand, AWS
allows us to attach a role to a resource at the time it is created. However,
I don't see an option for specifying the role using the spark_ec2.p
Hi,
I tried to develop a RandomForest model for my data in PySpark as follows:
rf_model = RandomForest.trainClassifier(train_idf, 2, {},
numTrees=15, seed=144)
print "RF: Num trees = %d, Num nodes = %d\n" %(rf_model.numTrees(),
rf_model.totalNumNodes())
pred_label = test_id
The following statement appears in the Scala API example at
https://spark.apache.org/docs/latest/api/scala/index.html#org.apache.spark.sql.DataFrame
people.filter("age" > 30).
I tried this example and it gave a compilation error. I think this needs to
be changed to people.filter(people("age") >
Hi,
I am trying to port some code that was working in Spark 1.2.0 on the latest
version, Spark 1.3.0. This code involves a left outer join between two
SchemaRDDs which I am now trying to change to a left outer join between 2
DataFrames. I followed the example for left outer join of DataFrame at
h
Hi,
I am using the filter() method to separate the rdds based on a predicate,as
follows:
val rdd1 = data.filter (t => { t._2 >0.0 && t._2 <= 1.0}) // t._2 is a
Double
val rdd2 = data.filter (t => { t._2 >1.0 && t._2 <= 4.0})
val rdd3 = data.filter (t => { t._2 >0.0 && t._2 <= 4.0}) // this sho
Hi,
I have a dataset in csv format and I am trying to standardize the features
before using k-means clustering. The data does not have any labels but has
the following format:
s1, f12,f13,...
s2, f21,f22,...
where s is a string id, and f is a floating point feature value.
To perform feature stan
Hi,
I am trying to get the final cluster centers after running the KMeans
algorithm in MLlib in order to characterize the clusters. But the
KMeansModel does not have any public method to retrieve this info. There
appears to be only a private method called clusterCentersWithNorm. I guess
I could c
Hi,
histogram() returns an object that is a pair of Arrays. There appears to be
no saveAsTextFile() for this paired object.
Currently I am using the following to save the output to a file:
val hist = a.histogram(10)
val arr1 = sc.parallelize(hist._1).saveAsTextFile("file1")
val arr2 = sc.parall
Hi,
I use the following code to read in data and extract the unique users using
Spark SQL. The data is 1.2 TB and I am running this on a cluster with 3 TB
memory. It appears that there is enough memory, but the program just freezes
after sometime where it maps the rdd to the case class Play. (If
Hi,
Is there any document that provides some guidelines with some examples that
illustrate when different performance optimizations would be useful? I am
interested in knowing the guidelines for using optimizations like cache(),
persist(), repartition(), coalesce(), and broadcast variables. I stu
Hi,
I am using sc.textFile("shared_dir/*") to load all the files in a directory
on a shared partition. The total size of the files in this directory is 1.2
TB. We have a 16 node cluster with 3 TB memory (1 node is driver, 15 nodes
are workers). But the loading fails after around 1 TB of data is
Hi,
I am using the following code to generate the (score, count) for each
window:
val score_count_by_window = topic.map(r => r._2) // r._2 is the integer
score
.countByValue()
score_count_by_window.print()
E.
Hi.
1) I dont see a groupBy() method for a DStream object. Not sure why that is
not supported. Currently I am using filter () to separate out the different
groups. I would like to know if there is a way to convert a DStream object
to a regular RDD so that I can apply the RDD methods like groupBy.
Small typo in my code in the previous post. That should be:
tweets.filter(_.getLang()=="en")
--
View this message in context:
http://apache-spark-user-list.1001560.n3.nabble.com/filtering-out-non-English-tweets-using-TwitterUtils-tp18614p18622.html
Sent from the Apache Spark User List mail
Thanks for the response. I tried the following :
tweets.filter(_.getLang()="en")
I get a compilation error:
value getLang is not a member of twitter4j.Status
But getLang() is one of the methods of twitter4j.Status since version 3.0.6
according to the doc at:
http://twitter4j.org/javadoc
Hi,
Is there a way to extract only the English language tweets when using
TwitterUtils.createStream()? The "filters" argument specifies the strings
that need to be contained in the tweets, but I am not sure how this can be
used to specify the language.
thanks
--
View this message in context:
Thanks. I was looking at an older RDD documentation that did not specify the
ordering option.
--
View this message in context:
http://apache-spark-user-list.1001560.n3.nabble.com/specifying-sort-order-for-sort-by-value-tp18289p18292.html
Sent from the Apache Spark User List mailing list archiv
Hi,
I am using rdd.sortBy(_._2) to get an RDD sorted by value. The default order
is ascending order. How can I get it sorted in descending order? I could
not find an option to specify the order.
I need to get the top K elements of the list sorted in descending order. If
there is no option to ge
SchemaRDD supports some of the SQL-like functionality like groupBy(),
distinct(), select(). However, SparkSQL also supports SQL statements which
provide this functionality. In terms of future support and performance, is
it better to use SQL statements or the SchemaRDD methods that provide
equivale
Hi,
I am getting an error in the Query Plan when I use the SQL statement
exactly as you have suggested. Is that the exact SQL statement I should be
using (I am not very familiar with SQL syntax)?
I also tried using the SchemaRDD's subtract method to perform this query.
usersRDD.subtract(deviceR
Hi,
I am using Spark 1.1.0. I have the following SQL statement where I am trying
to count the number of UIDs that are in the "tusers" table but not in the
"device" table.
val users_with_no_device = sql_cxt.sql("SELECT COUNT (u_uid) FROM tusers
WHERE tusers.u_uid NOT IN (SELECT d_uid FROM device)"
Hi,
As a result of a reduction operation, the resultant value "score" is a
DStream[Int] . How can I get the simple Int value?
I tried score[0], and score._1, but neither worked and can't find a
getValue() in the DStream API.
thanks
--
View this message in context:
http://apache-spark-user
You are right. Creating the StreamingContext from the SparkContext instead of
SparkConf helped. Thanks for the help.
--
View this message in context:
http://apache-spark-user-list.1001560.n3.nabble.com/Spark-Streaming-Sentiment-Analysis-of-Twitter-streams-tp16410p16520.html
Sent from the Apache
Hi,
I am trying to implement simple sentiment analysis of Twitter streams in
Spark/Scala. I am getting an exception and it appears when I combine
SparkContext with StreamingContext in the same program. When I read the
positive and negative words using only SparkContext.textFile (without
creating
Thanks. I made the change and ran the code. But I dont get any tweets for my
handle, although I do see the tweets when I search for it on twitter. Does
Spark allow us to get the tweets from the past (say the last 100 tweets?
tweets that appeared in the last 10 minutes)?
thanks
--
View this mess
Hi,
I am using Spark 1.1.0. Is there a way to get the complete tweets
corresponding to a handle (for e.g. @Delta)? I tried using the following
example that extracts just the hashtags and replaced the "#" with "@" as
follows. I need the complete tweet and not just the tags.
// val hashTag
- We set ulimit to 50. But I still get the same "too many open files"
warning.
- I tried setting consolidateFiles to True, but that did not help either.
I am using a Mesos cluster. Does Mesos have any limit on the number of
open files?
thanks
--
View this message in context:
http:/
Thanks for the help. Yes, I did not realize that the first header line has a
different separator.
By the way, is there a way to drop the first line that contains the header?
Something along the following lines:
sc.textFile(inp_file)
.drop(1) // or tail() to drop the header line
Hi,
I am trying to extract the number of distinct users from a file using Spark
SQL, but I am getting the following error:
ERROR Executor: Exception in task 1.0 in stage 8.0 (TID 15)
java.lang.ArrayIndexOutOfBoundsException: 1
I am following the code in examples/sql/RDDRelation.scala. My co
Hi,
Currently the history server provides application details for only the
successfully completed jobs (where the APPLICATION_COMPLETE file is
generated). However, (long-running) jobs that we terminate manually or
failed jobs where the APPLICATION_COMPLETE may not be generated, dont show
up on th
Hi,
I am using spark v 1.1.0. The default value of spark.cleaner.ttl is infinite
as per the online docs. Since a lot of shuffle files are generated in
/tmp/spark-local* and the disk is running out of space, we tested with a
smaller value of ttl. However, even when job has completed and the timer
Hi,
I am trying to compute the number of unique users from a year's worth of
data. So there are about 300 files and each file is quite large (~GB). I
first tried this without a loop by reading all the files in the directory
using the glob pattern: sc.textFile("dir/*"). But the tasks were stalli
Hi,
I am using Spark 1.1.0 on a cluster. My job takes as input 30 files in a
directory (I am using sc.textfile("dir/*") ) to read in the files. I am
getting the following warning:
WARN TaskSetManager: Lost task 99.0 in stage 1.0 (TID 99,
mesos12-dev.sccps.net): java.io.FileNotFoundException:
/t
I execute it as follows:
$SPARK_HOME/bin/spark-submit --master --class
org.apache.spark.examples.streaming.HdfsWordCount
target/scala-2.10/spark_stream_examples-assembly-1.0.jar
After I start the job, I add a new test file in hdfsdir. It is a large text
file which I will not be able to c
Hi,
I tried out the HdfsWordCount program in the Streaming module on a cluster.
Based on the output, I find that it counts only a few of the words. How can
I have it count all the words in the text? I have only one text file in the
directory.
thanks
--
View this message in context:
http://a
This issue is resolved. The file needs to be created after the program has
started to execute.
thanks
--
View this message in context:
http://apache-spark-user-list.1001560.n3.nabble.com/Streaming-HdfsWordCount-does-not-print-any-output-tp14849p14852.html
Sent from the Apache Spark User List
Hi,
I tried running the HdfsWordCount program in the streaming examples in Spark
1.1.0. I provided a directory in the distributed filesystem as input. This
directory has one text file. However, the only thing that the program keeps
printing is the time - but not the word count. I have not used the
Hi,
I am using the latest release Spark 1.1.0. I am trying to build the
streaming examples (under examples/streaming) as a standalone project with
the following streaming.sbt file. When I run sbt assembly, I get an error
stating that object algebird is not a member of package com.twitter. I
tried
I have created JIRA ticket 3610 for the issue. thanks.
--
View this message in context:
http://apache-spark-user-list.1001560.n3.nabble.com/Unable-to-load-app-logs-for-MLLib-programs-in-history-server-tp14627p14706.html
Sent from the Apache Spark User List mailing list archive at Nabble.com.
-
Hi,
I have a program similar to the BinaryClassifier example that I am running
using my data (which is fairly small). I run this for 100 iterations. I
observed the following performance:
Standalone mode cluster with 10 nodes (with Spark 1.0.2): 5 minutes
Standalone mode cluster with 10 nodes (wi
Hi,
The default log files for the Mllib examples use a rather long naming
convention that includes special characters like parentheses and comma.For
e.g. one of my log files is named
"binaryclassifier-with-params(input.txt,100,1.0,svm,l2,0.1)-1410566770032".
When I click on the program on the hi
This issue is resolved. Looks like in the new spark-submit, the jar path has
to be at the end of the options. Earlier I could specify this path in any
order on the command line.
thanks
--
View this message in context:
http://apache-spark-user-list.1001560.n3.nabble.com/Spark-1-1-0-Cannot-load
Hi,
I am using the Spark 1.1.0 version that was released yesterday. I recompiled
my program to use the latest version using "sbt assembly" after modifying
.sbt to use the 1.1.0 version. The compilation goes through and the
jar is built. When I run the jar using spark-submit, I get an error: "Canno
Hi,
I am using Spark 1.0.2 on a mesos cluster. After I run my job, when I try to
look at the detailed application stats using a history server@18080, the
stats don't show up for some of the jobs even though the job completed
successfully and the event logs are written to the log folder. The log f
Hi,
I am running Spark 1.0.2 on a cluster in Mesos mode. I am not able to access
the Spark master Web UI at port 8080 but am able to access it at port 5050.
Is 5050 the standard port?
Also, in the the standalone mode, there is a link to the Application detail
UI directly from the master UI. I do
The dataset is quite small : 5.6 KB. It has 200 rows and 3 features, and 1
column of labels. From this dataset, I split 80% for training set and 20%
for test set. The features are integer counts and labels are binary (1/0).
thanks
--
View this message in context:
http://apache-spark-user-lis
NUm Iterations: For LR and SVM, I am using the default value of 100. All
the other parameters also I am using the default values. I am pretty much
reusing the code from BinaryClassification.scala. For Decision Tree, I dont
see any parameter for number of iterations inthe example code, so I did
Hi,
I evaluated the runtime performance of some of the MLlib classification
algorithms on a local machine and a cluster with 10 nodes. I used standalone
mode and Spark 1.0.1 in both cases. Here are the results for the total
runtime:
Local Cluster
Logi
Hi,
I am having the same problem reported by Michael. I am trying to open 30
files. ulimit -n shows the limit is 1024. So I am not sure why the program
is failing with "Too many open files" error. The total size of all the 30
files is 230 GB.
I am running the job on a cluster with 10 nodes, eac
Hi,
Thanks for the responses. I understand that the second values in the Memory
Used column for the executors add up to 95.5 GB and the first values add up
to 17.3 KB. If 95.5 GB is the memory used to store the RDDs, then what is
17.3 KB ? is that the memory used for shuffling operations? For non
Hi,
I am using a cluster where each node has 16GB (this is the executor memory).
After I complete an MLlib job, the executor tab shows the following:
Memory: 142.6 KB Used (95.5 GB Total)
and individual worker nodes have the Memory Used values as 17.3 KB / 8.6 GB
(this is different for differe
Hi,
How do I convert a Map object to an RDD so that I can use the
saveAsTextFile() operation to output the Map object?
thanks
--
View this message in context:
http://apache-spark-user-list.1001560.n3.nabble.com/transforming-a-Map-object-to-RDD-tp13071.html
Sent from the Apache Spark User Li
Hi,
Thanks for the response. I tried to use countByKey. But I am not able to
write the output to console or to a file. Neither collect() nor
saveAsTextFile() work for the Map object that is generated after
countByKey().
valx = sc.textFile(baseFile)).map { line =>
val field
I was able to recently solve this problem for standalone mode. For this mode,
I did not use a history server. Instead, I set spark.eventLog.dir (in
conf/spark-defaults.conf) to a directory in hdfs (basically this directory
should be in a place that is writable by the master and accessible globally
Hi,
I have the following piece of code that I am running on a cluster with 10
nodes with 2GB memory per node. The tasks seem to complete, but at the point
where it is generating output (saveAsTextFile), the program freezes after
some time and reports an out of memory error (error transcript attach
I have already tried setting the history server and accessing it on
:18080 as per the link. But the page does not list any completed
applications. As I mentioned in my previous mail, I am running Spark in
standalone mode on the cluster (as well as on my local machine). According
to the link, it ap
Hi,
I am able to access the Application details web page from the master UI page
when I run Spark in standalone mode on my local machine. However, I am not
able to access it when I run Spark on our private cluster. The Spark master
runs on one of the nodes in the cluster. I am able to access the
Without the sc.union, my program crashes with the following error:
Exception in thread "main" org.apache.spark.SparkException: Job aborted due
to stage failure: Master removed our application: FAILED at
org.apache.spark.scheduler.DAGScheduler.org$apache$spark$scheduler$DAGScheduler$$failJobAndInde
Hi,
I have a piece of code that reads all the (csv) files in a folder. For each
file, it parses each line, extracts the first 2 elements from each row of
the file, groups the tuple by the key and finally outputs the number of
unique values for each key.
val conf = new SparkConf().setAppNam
Hi,
I have a piece of code in which the result of a groupByKey operation is as
follows:
(2013-04, ArrayBuffer(s1, s2, s3, s1, s2, s4))
The first element is a String value representing a date and the ArrayBuffer
consists of (non-unique) strings. I want to extract the unique elements of
the Array
Hi,
Ok, I was specifying --master local. I changed that to --master
spark://:7077 and am now able to see the completed
applications. It provides summary stats about runtime and memory usage,
which is sufficient for me at this time.
However it doesn't seem to archive the info in the "application
More specifically, as indicated by Patrick above, in 1.0+, apps will have
persistent state so that the UI can be reloaded. Is there a way to enable
this feature in 1.0.1?
thanks
--
View this message in context:
http://apache-spark-user-list.1001560.n3.nabble.com/Spark-webUI-application-details
I set "spark.eventLog.enabled" to true in
$SPARK_HOME/conf/spark-defaults.conf and also configured the logging to a
file as well as console in log4j.properties. But I am not able to get the
log of the statistics in a file. On the console there is a lot of log
messages along with the stats - so ha
Hi,
I am using Spark 1.0.1. But I am still not able to see the stats for
completed apps on port 4040 - only for running apps. Is this feature
supported or is there a way to log this info to some file? I am interested
in stats about the total # of executors, total runtime, and total memory
used by
Hi,
Thanks for the reference to the LBFGS optimizer.
I tried to use the LBFGS optimizer, but I am not able to pass it as an
input to the LogisticRegression model for binary classification. After
studying the code in mllib/classification/LogisticRegression.scala, it
appears that the only impleme
i was using sbt package when I got this error. Then I switched to using sbt
assembly and that solved the issue. To run "sbt assembly", you need to have
a file called plugins.sbt in the "/project" directory and it
has the following line:
addSbtPlugin("com.eed3si9n" % "sbt-assembly" % "0.11.2")
T
Hi,
According to the MLLib guide, there seems to be support for different loss
functions. But I could not find a command line parameter to choose the loss
function but only found regType to choose the regularization. Does MLLib
support a parameter to choose the loss function?
thanks
--
View t
What is the definition of regParam and what is the range of values it is
allowed to take?
thanks
--
View this message in context:
http://apache-spark-user-list.1001560.n3.nabble.com/Regularization-parameters-tp11601p11737.html
Sent from the Apache Spark User List mailing list archive at Nabbl
Ok, thanks for clarifying. So looks like numFeatures is only relevant for lib
SVM format. I am using LabeledPoint, so if data is not sparse, perhaps
numFeatures is not required. I thought that the Params class defines all
the parameters passed to the ML algorithm. But it looks like it also
include
Spark 1.0.1
thanks
--
View this message in context:
http://apache-spark-user-list.1001560.n3.nabble.com/Regularization-parameters-tp11601p11631.html
Sent from the Apache Spark User List mailing list archive at Nabble.com.
-
T
Hi,
I am following the code in
examples/src/main/scala/org/apache/spark/examples/mllib/BinaryClassification.scala
For setting the parameters and parsing the command line options, I am just
reusing that code.Params is defined as follows.
case class Params(
input: String = null,
numIt
I followed the example in
examples/src/main/scala/org/apache/spark/examples/mllib/SparseNaiveBayes.scala.
IN this file Params is defined as follows:
case class Params (
input: String = null,
minPartitions: Int = 0,
numFeatures: Int = -1,
lambda: Double = 1.0)
In the main functio
Hi,
I tried different regularization parameter values with Logistic Regression
for binary classification of my dataset and would like to understand the
following results:
regType = L2, regParam = 0.0 , I am getting AUC = 0.80 and accuracy of 80%
regType = L1, regParam = 0.0 , I am getting AUC =
1) How is the minPartitions parameter in NaiveBayes example used? What is
the default value?
2) Why is the numFeatures specified as a parameter? Can this not be
obtained from the data? This parameter is not specified for the other MLlib
algorithms.
thanks
--
View this message in context:
Hi,
So I again ran "sbt clean" followed by all of the steps listed above to
rebuild the jars after cleaning. My compilation error still persists.
Specifically, I am trying to extract an element from the feature vector that
is part of a LabeledPoint as follows:
data.features(i)
This gives the fo
Hi,
I upgraded to 1.0.1 from 1.0 a couple of weeks ago and have been able to use
some of the features advertised in 1.0.1. However, I get some compilation
errors in some cases and based on user response, these errors have been
addressed in the 1.0.1 version and so I should not be getting these er
I am using 1.0.1. It does not matter to me whether it is the first or second
element. I would like to know how to extract the i-th element in the feature
vector (not the label).
data.features(i) gives the following error:
method apply in trait Vector cannot be accessed in
org.apache.spark.mllib.l
Hi,
I want to extract the individual elements of a feature vector that is part
of a LabeledPoint. I tried the following:
data.features._1
data.features(1)
data.features.map(_.1)
data is a LabeledPoint with a feature vector containing 3 features. All of
the above resulted in compilation errors
I have also used labeledPoint or libSVM format (for sparse data) for
DecisionTree. When I had categorical labels (not features), I mapped the
categories to numerical data as part of the data transformation step (i.e.
before creating the LabeledPoint).
--
View this message in context:
http://apa
I am using 1.0.1 and I am running locally (I am not providing any master
URL). But the zip() does not produce the correct count as I mentioned above.
So not sure if the issue has been fixed in 1.0.1. However, instead of using
zip, I am now using the code that Sean has mentioned and am getting the
c
Hi,
In order to evaluate the ML classification accuracy, I am zipping up the
prediction and test labels as follows and then comparing the pairs in
predictionAndLabel:
val prediction = model.predict(test.map(_.features))
val predictionAndLabel = prediction.zip(test.map(_.label))
However, I am fi
yes, the output is continuous. So I used a threshold to get binary labels.
If prediction < threshold, then class is 0 else 1. I use this binary label
to then compute the accuracy. Even with this binary transformation, the
accuracy with decision tree model is low compared to LR or SVM (for the
spec
Hi,
The mllib.clustering.kmeans implementation supports a random or parallel
initialization mode to pick the initial centers. is there a way to specify
the initial centers explictly? It would be useful to have a setCenters()
method where we can explicitly specify the initial centers. (For e.g. R
To add to my previous post, the error at runtime is teh following:
Exception in thread "main" org.apache.spark.SparkException: Job aborted due
to stage failure: Task 0.0:0 failed 1 times, most recent failure: Exception
failure in TID 0 on host localhost: org.json4s.package$MappingException:
Expect
Hi,
I have a json file where the object definition in each line includes an
array component "obj" that contains 0 or more elements as shown by the
example below.
{"name": "16287e9cdf", "obj": [{"min": 50,"max": 59 }, {"min": 20, "max":
29}]},
{"name": "17087e9cdf", "obj": [{"min": 30,"max": 3
The problem is resolved. Thanks.
--
View this message in context:
http://apache-spark-user-list.1001560.n3.nabble.com/jsonRDD-NoSuchMethodError-tp9688p9742.html
Sent from the Apache Spark User List mailing list archive at Nabble.com.
I am running this in standalone mode on a single machine. I built the spark
jar from scratch (sbt assembly) and then included that in my application
(the same process I have done for earlier versions).
thanks.
--
View this message in context:
http://apache-spark-user-list.1001560.n3.nabble.co
Hi,
I am using Spark 1.0.1. I am using the following piece of code to parse a
json file. It is based on the code snippet in the SparkSQL programming
guide. However, the compiler outputs an error stating:
java.lang.NoSuchMethodError:
org.apache.spark.sql.SQLContext.jsonRDD(Lorg/apache/spark/rdd/R
Hi,
I have a json file where the definition of each object spans multiple lines.
An example of one object definition appears below.
{
"name": "16287e9cdf",
"width": 500,
"height": 325,
"width": 1024,
"height": 665,
"obj": [
{
"x": 395.08,
"y": 82.09,
Hi,
I need to perform binary classification on an image dataset. Each image is a
data point described by a Json object. The feature set for each image is a
set of feature vectors, each feature vector corresponding to a distinct
object in the image. For example, if an image has 5 objects, its featu
I dont have a proxy server.
thanks.
--
View this message in context:
http://apache-spark-user-list.1001560.n3.nabble.com/Streaming-training-Spark-Summit-2014-tp9465p9481.html
Sent from the Apache Spark User List mailing list archive at Nabble.com.
1 - 100 of 130 matches
Mail list logo