[
https://issues.apache.org/jira/browse/HDFS-6819?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14087070#comment-14087070
]
George Wong commented on HDFS-6819:
-----------------------------------
I propose to use a separate project to build AOP FI framework in HDFS.
We can put the project in "hadoop-hdfs-project" or in
"hadoop-hdfs-project/hadoop-hdfs/src/contrib/".
The HDFS-FI project depends on the hadoop-hdfs project. When building HDFS-FI,
AspectJ plugin weaves the Aspect advice into the hdfs jars like
hadoop-hdfs-2.4.0.jar etc.
There are several pros for this proposal.
# fault injection code and UT are isolated with the normal HDFS code and UT.
# Aspect advice is weaved into the hdfs jars which are built in
hadoop-hdfs-project. So, the weaved jars and normal jars are separated jars.
What's your opinions?
> make HDFS fault injection framework working with maven
> ------------------------------------------------------
>
> Key: HDFS-6819
> URL: https://issues.apache.org/jira/browse/HDFS-6819
> Project: Hadoop HDFS
> Issue Type: Task
> Reporter: George Wong
> Assignee: George Wong
>
> In current trunk code repo, the FI framework does not work. Because maven
> build process does not execute the AspectJ injection.
> Since FI is very useful for testing and bug reproduce, it is better to make
> FI framework working in the trunk code.
--
This message was sent by Atlassian JIRA
(v6.2#6252)