+1

On Thu, Jul 30, 2015 at 11:18 AM, Patrick Wendell <pwend...@gmail.com>
wrote:

> Yeah this could make sense - allowing data sources to register a short
> name. What mechanism did you have in mind? To use the jar service loader?
>
> The only issue is that there could be conflicts since many of these are
> third party packages. If the same name were registered twice I'm not sure
> what the best behavior would be. Ideally in my mind if the same shortname
> were registered twice we'd force the user to use a fully qualified name and
> say the short name is ambiguous.
>
> Patrick
> On Jul 30, 2015 9:44 AM, "Joseph Batchik" <josephbatc...@gmail.com> wrote:
>
>> Hi all,
>>
>> There are now starting to be a lot of data source packages for Spark. A
>> annoyance I see is that I have to type in the full class name like:
>>
>> sqlContext.read.format("com.databricks.spark.avro").load(path).
>>
>> Spark internally has formats such as "parquet" and "jdbc" registered and
>> it would be nice to be able just to type in "avro", "redshift", etc. as
>> well. Would it be a good idea to use something like a service loader to
>> allow data sources defined in other packages to register themselves with
>> Spark? I think that this would make it easier for end users. I would be
>> interested in adding this, please let me know what you guys think.
>>
>> - Joe
>>
>>
>>

Reply via email to