[ 
https://issues.apache.org/jira/browse/HDFS-201?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Allen Wittenauer resolved HDFS-201.
-----------------------------------

    Resolution: Duplicate

> Spring and OSGi support
> -----------------------
>
>                 Key: HDFS-201
>                 URL: https://issues.apache.org/jira/browse/HDFS-201
>             Project: Hadoop HDFS
>          Issue Type: New Feature
>            Reporter: Jon Brisbin
>            Assignee: Jean-Baptiste Onofré
>         Attachments: HDFS-201.patch
>
>
> I was able to compile 0.18.2 in eclipse into a new OSGi bundle using eclipse 
> PDE. Using Spring to control the HDFS nodes, however, seems out of the 
> question for the time being because of inter-dependencies between packages 
> that should be separate OSGi bundles (for example, SecondaryNameNode includes 
> direct references to StatusHttpServer, which should be in a bundle with a 
> "web" personality that is separate from Hadoop Core). Looking through the 
> code that starts the daemons, it would seem code changes are necessary to 
> allow for components to be dependency-injected. Rather than instantiating a 
> StatusHttpServer inside the SecondaryNameNode, that reference should (at the 
> very least) be able to be dependency-injected (for example from an OSGi 
> service from another bundle). Adding setters for infoServer would allow that 
> reference to be injected by Spring. This is just an example of the changes 
> that would need to be made to get Hadoop to live happily inside an OSGi 
> container.
> As a starting point, it would be nice if Hadoop core was able to be split 
> into a client bundle that could be deployed into OSGi containers that would 
> provide client-only access to HDFS clusters.



--
This message was sent by Atlassian JIRA
(v6.2#6252)

Reply via email to