Hi,
I also filed a jira yesterday:
https://issues.apache.org/jira/browse/SPARK-26538
Looks like one needs to be closed as duplicate. Sorry for the late update.
Best regards
--
Sent from: http://apache-spark-user-list.1001560.n3.nabble.com/
Hi,
I filed a jira: https://issues.apache.org/jira/browse/SPARK-26540
On Thu, Jan 3, 2019 at 10:04 PM Takeshi Yamamuro
wrote:
> Hi,
>
> I checked that v2.2/v2.3/v2.4/master had the same issue, so can you file a
> jira?
> I looked over the related code and then I think we need more logics to
>
Hi,
I checked that v2.2/v2.3/v2.4/master had the same issue, so can you file a
jira?
I looked over the related code and then I think we need more logics to
handle this issue;
Hi,
I came across strange behavior when dealing with postgres columns of type
numeric[] using Spark 2.3.2, PostgreSQL 10.4, 9.6.9.
Consider the following table definition:
create table test1
(
v numeric[],
d numeric
);
insert into test1 values('{.222,.332}', 222.4555);
When