Forwarding my last email since I didn't get any feedback.
Thanks
---------- Forwarded message ----------
Hi Marco and others
I like to work on the Gsoc project "Fact Extraction from Wikipedia Text"
during this summer.
I went through the project description and the research papers mentioned
under the description. I have few questions to clarify.
1- As mentioned in the project idea the main objective is the
implementation of a new text extractor. Will this need to be implemented
inside the current extraction-framework? Or would it be a completely new
tool?
2- Also it mentioned the use of NLP techniques to process Wikipedia text.
Does this means extraction of Dependency relationships to get the frame
elements (FE) and lexical unit(LU)? There are several NLP libraries like
Stanford parser, RelEx, NLTK etc. Is there any decision made which NLP
library to use?
3- Also regarding the content of a Wikipedia page; do we use all the
sentences from the Wikipedia page? My idea is it's better if we can use
important sentences rather than all the sentences. If that is the better
idea we have to come up with a criteria to select important sentences.
--
Regards
Kasun Perera
--
Regards
Kasun Perera
------------------------------------------------------------------------------
Dive into the World of Parallel Programming The Go Parallel Website, sponsored
by Intel and developed in partnership with Slashdot Media, is your hub for all
things parallel software development, from weekly thought leadership blogs to
news, videos, case studies, tutorials and more. Take a look and join the
conversation now. http://goparallel.sourceforge.net/
_______________________________________________
Dbpedia-developers mailing list
[email protected]
https://lists.sourceforge.net/lists/listinfo/dbpedia-developers