Yep, Opened SPARK-5054 <https://issues.apache.org/jira/browse/SPARK-5054>

Thanks
Best Regards

On Tue, Dec 30, 2014 at 5:52 AM, Michael Armbrust <mich...@databricks.com>
wrote:

> Yeah, this looks like a regression in the API due to the addition of
> arbitrary decimal support.  Can you open a JIRA?
>
> On Sun, Dec 28, 2014 at 12:23 AM, Akhil Das <ak...@sigmoidanalytics.com>
> wrote:
>
>> Hi Zigen,
>>
>> Looks like they missed it.
>>
>> Thanks
>> Best Regards
>>
>> On Sat, Dec 27, 2014 at 12:43 PM, Zigen Zigen <dbviewer.zi...@gmail.com>
>> wrote:
>>
>>> Hello , I am zigen.
>>>
>>> I am using the Spark SQL 1.1.0.
>>>
>>> I want to use the Spark SQL 1.2.0.
>>>
>>>
>>> but my Spark application is a compile error.
>>>
>>> Spark 1.1.0 had a DataType.DecimalType.
>>>
>>> but Spark1.2.0 had not DataType.DecimalType.
>>>
>>> Why ?
>>>
>>>
>>> JavaDoc (Spark 1.1.0)
>>>
>>> http://people.apache.org/~pwendell/spark-1.1.0-rc1-docs/api/java/org/apache/spark/sql/api/java/DataType.html
>>>
>>>
>>> JavaDoc (Spark 1.2.0)
>>>
>>> http://people.apache.org/~pwendell/spark-1.2.0-rc1-docs/api/java/org/apache/spark/sql/api/java/DataType.html
>>>
>>>
>>> programing guild (Spark 1.2.0)
>>>
>>> https://spark.apache.org/docs/latest/sql-programming-guide.html#spark-sql-datatype-reference
>>>
>>>
>>
>

Reply via email to