Hi Cristian

On Mon, Mar 17, 2014 at 9:43 PM, Cristian Petroaca
<cristian.petro...@gmail.com> wrote:
> 1. Updated to the latest code and it's gone. Cool
>
> 2. I start the stable launcher -> create a new instance of the
> PosChunkerEngine -> add it to the default chain. At this point everything
> looks good and works ok.
> After I restart the server the default chain is gone and instead I see this
> in the enhancement chains page : all-active (default, id: 149, ranking: 0,
> impl: AllActiveEnginesChain ). all-active did not contain the 'default'
> word before the restart.
>

Please note the default chain selection rules as described at [1]. You
can also access chains chains under '/enhancer/chain/{chain-name}'

best
Rupert

[1] 
http://stanbol.staging.apache.org/docs/trunk/components/enhancer/chains/#default-chain

> It looks like the config files are exactly what I need. Thanks.
>
>
> 2014-03-17 9:26 GMT+02:00 Rupert Westenthaler <rupert.westentha...@gmail.com
>>:
>
>> On Sat, Mar 15, 2014 at 8:34 PM, Cristian Petroaca
>> <cristian.petro...@gmail.com> wrote:
>> > Thanks Rupert.
>> >
>> > A couple more questions/issues :
>> >
>> > 1. Whenever I start the stanbol server I'm seeing this in the console
>> > output :
>> >
>>
>> This should be fixed with STANBOL-1278 [1] [2]
>>
>> > 2. Whenever I restart the server the Weighted Chains get messed up. I
>> > usually use the 'default' chain and add my engine to it so there are 11
>> > engines in it. After the restart this chain now contains around 23
>> engines
>> > in total.
>>
>> I was not able to replicate this. What I tried was
>>
>> (1) start up the stable launcher
>> (2) add an additional engine to the default chain
>> (3) restart the launcher
>>
>> The default chain was not changed after (2) and (3). So I would need
>> further information for knowing why this is happening.
>>
>> Generally it is better to create you own chain instance as modifying
>> one that is provided by the default configuration. I would also
>> recommend that you keep your test configuration in text files and to
>> copy those to the 'stanbol/fileinstall' folder. Doing so prevent you
>> from manually entering the configuration after a software update. The
>> production-mode section [3] provides information on how to do that.
>>
>> best
>> Rupert
>>
>> [1] https://issues.apache.org/jira/browse/STANBOL-1278
>> [2] http://svn.apache.org/r1576623
>> [3] http://stanbol.apache.org/docs/trunk/production-mode
>>
>> > ERROR: Bundle org.apache.stanbol.enhancer.engine.topic.web [153]: Error
>> > starting
>> >
>>  
>> slinginstall:c:\Data\Projects\Stanbol\main\launchers\stable\target\stanbol\star
>> > tup\35\org.apache.stanbol.enhancer.engine.topic.web-1.0.0-SNAPSHOT.jar
>> > (org.osgi
>> > .framework.BundleException: Unresolved constraint in bundle
>> > org.apache.stanbol.e
>> > nhancer.engine.topic.web [153]: Unable to resolve 153.0: missing
>> > requirement [15
>> > 3.0] package; (&(package=javax.ws.rs
>> )(version>=0.0.0)(!(version>=2.0.0))))
>> > org.osgi.framework.BundleException: Unresolved constraint in bundle
>> > org.apache.s
>> > tanbol.enhancer.engine.topic.web [153]: Unable to resolve 153.0: missing
>> > require
>> > ment [153.0] package; (&(package=javax.ws.rs
>> > )(version>=0.0.0)(!(version>=2.0.0))
>> > )
>> >         at
>> org.apache.felix.framework.Felix.resolveBundle(Felix.java:3443)
>> >         at org.apache.felix.framework.Felix.startBundle(Felix.java:1727)
>> >         at
>> > org.apache.felix.framework.Felix.setActiveStartLevel(Felix.java:1156)
>> >
>> >         at
>> > org.apache.felix.framework.StartLevelImpl.run(StartLevelImpl.java:264
>> > )
>> >         at java.lang.Thread.run(Unknown Source)
>> >
>> > Despite of this the server starts fine and I can use the enhancer fine.
>> Do
>> > you guys see this as well?
>> >
>> >
>> > 2. Whenever I restart the server the Weighted Chains get messed up. I
>> > usually use the 'default' chain and add my engine to it so there are 11
>> > engines in it. After the restart this chain now contains around 23
>> engines
>> > in total.
>> >
>> >
>> >
>> >
>> > 2014-03-11 9:47 GMT+02:00 Rupert Westenthaler <
>> rupert.westentha...@gmail.com
>> >>:
>> >
>> >> Hi Cristian,
>> >>
>> >> NER Annotations are typically available as both
>> >> NlpAnnotations.NER_ANNOTATION and  fise:TextAnnotation [1] in the
>> >> enhancement metadata. As you are already accessing the AnayzedText I
>> >> would prefer using the  NlpAnnotations.NER_ANNOTATION.
>> >>
>> >> best
>> >> Rupert
>> >>
>> >> [1]
>> >>
>> http://stanbol.apache.org/docs/trunk/components/enhancer/enhancementstructure.html#fisetextannotation
>> >>
>> >> On Mon, Mar 10, 2014 at 10:07 PM, Cristian Petroaca
>> >> <cristian.petro...@gmail.com> wrote:
>> >> > Thanks.
>> >> > I assume I should get the Named entities using the same but with
>> >> > NlpAnnotations.NER_ANNOTATION?
>> >> >
>> >> >
>> >> >
>> >> > 2014-03-10 13:29 GMT+02:00 Rupert Westenthaler <
>> >> > rupert.westentha...@gmail.com>:
>> >> >
>> >> >> Hallo Cristian,
>> >> >>
>> >> >> NounPhrases are not added to the RDF enhancement results. You need to
>> >> >> use the AnalyzedText ContentPart [1]
>> >> >>
>> >> >> here is some demo code you can use in the computeEnhancement method
>> >> >>
>> >> >>         AnalysedText at = NlpEngineHelper.getAnalysedText(this, ci,
>> >> true);
>> >> >>         Iterator<? extends Section> sections = at.getSentences();
>> >> >>         if(!sections.hasNext()){ //process as single sentence
>> >> >>             sections = Collections.singleton(at).iterator();
>> >> >>         }
>> >> >>
>> >> >>         while(sections.hasNext()){
>> >> >>             Section section = sections.next();
>> >> >>             Iterator<Span> chunks =
>> >> >> section.getEnclosed(EnumSet.of(SpanTypeEnum.Chunk));
>> >> >>             while(chunks.hasNext()){
>> >> >>                 Span chunk = chunks.next();
>> >> >>                 Value<PhraseTag> phrase =
>> >> >> chunk.getAnnotation(NlpAnnotations.PHRASE_ANNOTATION);
>> >> >>                 if(phrase.value().getCategory() ==
>> >> LexicalCategory.Noun){
>> >> >>                     log.info(" - NounPhrase [{},{}] {}", new
>> Object[]{
>> >> >>
>> >> >> chunk.getStart(),chunk.getEnd(),chunk.getSpan()});
>> >> >>                 }
>> >> >>             }
>> >> >>         }
>> >> >>
>> >> >> hope this helps
>> >> >>
>> >> >> best
>> >> >> Rupert
>> >> >>
>> >> >> [1]
>> >> >>
>> >>
>> http://stanbol.apache.org/docs/trunk/components/enhancer/nlp/analyzedtext
>> >> >>
>> >> >> On Sun, Mar 9, 2014 at 6:07 PM, Cristian Petroaca
>> >> >> <cristian.petro...@gmail.com> wrote:
>> >> >> > I started to implement the engine and I'm having problems with
>> getting
>> >> >> > results for noun phrases. I modified the "default" weighted chain
>> to
>> >> also
>> >> >> > include the PosChunkerEngine and ran a sample text : "Angela Merkel
>> >> >> visted
>> >> >> > China. The german chancellor met with various people". I expected
>> that
>> >> >> the
>> >> >> > RDF XML output would contain some info about the noun phrases but I
>> >> >> cannot
>> >> >> > see any.
>> >> >> > Could you point me to the correct way to generate the noun phrases?
>> >> >> >
>> >> >> > Thanks,
>> >> >> > Cristian
>> >> >> >
>> >> >> >
>> >> >> > 2014-02-09 14:15 GMT+02:00 Cristian Petroaca <
>> >> >> cristian.petro...@gmail.com>:
>> >> >> >
>> >> >> >> Opened https://issues.apache.org/jira/browse/STANBOL-1279
>> >> >> >>
>> >> >> >>
>> >> >> >> 2014-02-07 10:53 GMT+02:00 Cristian Petroaca <
>> >> >> cristian.petro...@gmail.com>
>> >> >> >> :
>> >> >> >>
>> >> >> >> Hi Rupert,
>> >> >> >>>
>> >> >> >>> The "spatial" dimension is a good idea. I'll also take a look at
>> >> Yago.
>> >> >> >>>
>> >> >> >>> I will create a Jira with what we talked about here. It will
>> >> probably
>> >> >> >>> have just a draft-like description for now and will be updated
>> as I
>> >> go
>> >> >> >>> along.
>> >> >> >>>
>> >> >> >>> Thanks,
>> >> >> >>> Cristian
>> >> >> >>>
>> >> >> >>>
>> >> >> >>> 2014-02-06 15:39 GMT+02:00 Rupert Westenthaler <
>> >> >> >>> rupert.westentha...@gmail.com>:
>> >> >> >>>
>> >> >> >>> Hi Cristian,
>> >> >> >>>>
>> >> >> >>>> definitely an interesting approach. You should have a look at
>> Yago2
>> >> >> >>>> [1]. As far as I can remember the Yago taxonomy is much better
>> >> >> >>>> structured as the one used by dbpedia. Mapping suggestions of
>> >> dbpedia
>> >> >> >>>> to concepts in Yago2 is easy as both dbpedia and yago2 do
>> provide
>> >> >> >>>> mappings [2] and [3]
>> >> >> >>>>
>> >> >> >>>> > 2014-02-05 15:39 GMT+02:00 Rafa Haro <rh...@apache.org>:
>> >> >> >>>> >>
>> >> >> >>>> >> "Microsoft posted its 2013 earnings. The Redmond's company
>> made
>> >> a
>> >> >> >>>> >> huge profit".
>> >> >> >>>>
>> >> >> >>>> Thats actually a very good example. Spatial contexts are very
>> >> >> >>>> important as they tend to be often used for referencing. So I
>> would
>> >> >> >>>> suggest to specially treat the spatial context. For spatial
>> >> Entities
>> >> >> >>>> (like a City) this is easy, but even for other (like a Person,
>> >> >> >>>> Company) you could use relations to spatial entities define
>> their
>> >> >> >>>> spatial context. This context could than be used to correctly
>> link
>> >> >> >>>> "The Redmond's company" to "Microsoft".
>> >> >> >>>>
>> >> >> >>>> In addition I would suggest to use the "spatial" context of each
>> >> >> >>>> entity (basically relation to entities that are cities, regions,
>> >> >> >>>> countries) as a separate dimension, because those are very often
>> >> used
>> >> >> >>>> for coreferences.
>> >> >> >>>>
>> >> >> >>>> [1] http://www.mpi-inf.mpg.de/yago-naga/yago/
>> >> >> >>>> [2] http://downloads.dbpedia.org/3.9/links/yago_links.nt.bz2
>> >> >> >>>> [3]
>> >> >> >>>>
>> >> >>
>> >>
>> http://www.mpi-inf.mpg.de/yago-naga/yago/download/yago/yagoDBpediaInstances.ttl.7z
>> >> >> >>>>
>> >> >> >>>>
>> >> >> >>>> On Thu, Feb 6, 2014 at 10:33 AM, Cristian Petroaca
>> >> >> >>>> <cristian.petro...@gmail.com> wrote:
>> >> >> >>>> > There are several dbpedia categories for each entity, in this
>> >> case
>> >> >> for
>> >> >> >>>> > Microsoft we have :
>> >> >> >>>> >
>> >> >> >>>> > category:Companies_in_the_NASDAQ-100_Index
>> >> >> >>>> > category:Microsoft
>> >> >> >>>> > category:Software_companies_of_the_United_States
>> >> >> >>>> > category:Software_companies_based_in_Washington_(state)
>> >> >> >>>> > category:Companies_established_in_1975
>> >> >> >>>> > category:1975_establishments_in_the_United_States
>> >> >> >>>> > category:Companies_based_in_Redmond,_Washington
>> >> >> >>>> >
>> >> category:Multinational_companies_headquartered_in_the_United_States
>> >> >> >>>> > category:Cloud_computing_providers
>> >> >> >>>> > category:Companies_in_the_Dow_Jones_Industrial_Average
>> >> >> >>>> >
>> >> >> >>>> > So we also have "Companies based in Redmont,Washington" which
>> >> could
>> >> >> be
>> >> >> >>>> > matched.
>> >> >> >>>> >
>> >> >> >>>> >
>> >> >> >>>> > There is still other contextual information from dbpedia which
>> >> can
>> >> >> be
>> >> >> >>>> used.
>> >> >> >>>> > For example for an Organization we could also include :
>> >> >> >>>> > dbpprop:industry = Software
>> >> >> >>>> > dbpprop:service = Online Service Providers
>> >> >> >>>> >
>> >> >> >>>> > and for a Person (that's for Barack Obama) :
>> >> >> >>>> >
>> >> >> >>>> > dbpedia-owl:profession:
>> >> >> >>>> >                                dbpedia:Author
>> >> >> >>>> >                                dbpedia:Constitutional_law
>> >> >> >>>> >                                dbpedia:Lawyer
>> >> >> >>>> >                                dbpedia:Community_organizing
>> >> >> >>>> >
>> >> >> >>>> > I'd like to continue investigating this as I think that it may
>> >> have
>> >> >> >>>> some
>> >> >> >>>> > value in increasing the number of coreference resolutions and
>> I'd
>> >> >> like
>> >> >> >>>> to
>> >> >> >>>> > concentrate more on precision rather than recall since we
>> already
>> >> >> have
>> >> >> >>>> a
>> >> >> >>>> > set of coreferences detected by the stanford nlp tool and this
>> >> would
>> >> >> >>>> be as
>> >> >> >>>> > an addition to that (at least this is how I would like to use
>> >> it).
>> >> >> >>>> >
>> >> >> >>>> > Is it ok if I track this by opening a jira? I could update it
>> to
>> >> >> show
>> >> >> >>>> my
>> >> >> >>>> > progress and also my conclusions and if it turns out that it
>> was
>> >> a
>> >> >> bad
>> >> >> >>>> idea
>> >> >> >>>> > then that's the situation at least I'll end up with more
>> >> knowledge
>> >> >> >>>> about
>> >> >> >>>> > Stanbol in the end :).
>> >> >> >>>> >
>> >> >> >>>> >
>> >> >> >>>> > 2014-02-05 15:39 GMT+02:00 Rafa Haro <rh...@apache.org>:
>> >> >> >>>> >
>> >> >> >>>> >> Hi Cristian,
>> >> >> >>>> >>
>> >> >> >>>> >> The approach sounds nice. I don't want to be the devil's
>> >> advocate
>> >> >> but
>> >> >> >>>> I'm
>> >> >> >>>> >> just not sure about the recall using the dbpedia categories
>> >> >> feature.
>> >> >> >>>> For
>> >> >> >>>> >> example, your sentence could be also "Microsoft posted its
>> 2013
>> >> >> >>>> earnings.
>> >> >> >>>> >> The Redmond's company made a huge profit". So, maybe
>> including
>> >> more
>> >> >> >>>> >> contextual information from dbpedia could increase the recall
>> >> but
>> >> >> of
>> >> >> >>>> course
>> >> >> >>>> >> will reduce the precision.
>> >> >> >>>> >>
>> >> >> >>>> >> Cheers,
>> >> >> >>>> >> Rafa
>> >> >> >>>> >>
>> >> >> >>>> >> El 04/02/14 09:50, Cristian Petroaca escribió:
>> >> >> >>>> >>
>> >> >> >>>> >>  Back with a more detailed description of the steps for
>> making
>> >> this
>> >> >> >>>> kind of
>> >> >> >>>> >>> coreference work.
>> >> >> >>>> >>>
>> >> >> >>>> >>> I will be using references to the following text in the
>> steps
>> >> >> below
>> >> >> >>>> in
>> >> >> >>>> >>> order to make things clearer : "Microsoft posted its 2013
>> >> >> earnings.
>> >> >> >>>> The
>> >> >> >>>> >>> software company made a huge profit."
>> >> >> >>>> >>>
>> >> >> >>>> >>> 1. For every noun phrase in the text which has :
>> >> >> >>>> >>>      a. a determinate pos which implies reference to an
>> entity
>> >> >> local
>> >> >> >>>> to
>> >> >> >>>> >>> the
>> >> >> >>>> >>> text, such as "the, this, these") but not "another, every",
>> etc
>> >> >> which
>> >> >> >>>> >>> implies a reference to an entity outside of the text.
>> >> >> >>>> >>>      b. having at least another noun aside from the main
>> >> required
>> >> >> >>>> noun
>> >> >> >>>> >>> which
>> >> >> >>>> >>> further describes it. For example I will not count "The
>> >> company"
>> >> >> as
>> >> >> >>>> being
>> >> >> >>>> >>> a
>> >> >> >>>> >>> legitimate candidate since this could create a lot of false
>> >> >> >>>> positives by
>> >> >> >>>> >>> considering the double meaning of some words such as "in the
>> >> >> company
>> >> >> >>>> of
>> >> >> >>>> >>> good people".
>> >> >> >>>> >>> "The software company" is a good candidate since we also
>> have
>> >> >> >>>> "software".
>> >> >> >>>> >>>
>> >> >> >>>> >>> 2. match the nouns in the noun phrase to the contents of the
>> >> >> dbpedia
>> >> >> >>>> >>> categories of each named entity found prior to the location
>> of
>> >> the
>> >> >> >>>> noun
>> >> >> >>>> >>> phrase in the text.
>> >> >> >>>> >>> The dbpedia categories are in the following format (for
>> >> Microsoft
>> >> >> for
>> >> >> >>>> >>> example) : "Software companies of the United States".
>> >> >> >>>> >>>   So we try to match "software company" with that.
>> >> >> >>>> >>> First, as you can see, the main noun in the dbpedia category
>> >> has a
>> >> >> >>>> plural
>> >> >> >>>> >>> form and it's the same for all categories which I saw. I
>> don't
>> >> >> know
>> >> >> >>>> if
>> >> >> >>>> >>> there's an easier way to do this but I thought of applying a
>> >> >> >>>> lemmatizer on
>> >> >> >>>> >>> the category and the noun phrase in order for them to have a
>> >> >> common
>> >> >> >>>> >>> denominator.This also works if the noun phrase itself has a
>> >> plural
>> >> >> >>>> form.
>> >> >> >>>> >>>
>> >> >> >>>> >>> Second, I'll need to use for comparison only the words in
>> the
>> >> >> >>>> category
>> >> >> >>>> >>> which are themselves nouns and not prepositions or
>> determiners
>> >> >> such
>> >> >> >>>> as "of
>> >> >> >>>> >>> the".This means that I need to pos tag the categories
>> contents
>> >> as
>> >> >> >>>> well.
>> >> >> >>>> >>> I was thinking of running the pos and lemma on the dbpedia
>> >> >> >>>> categories when
>> >> >> >>>> >>> building the dbpedia backed entity hub and storing them for
>> >> later
>> >> >> >>>> use - I
>> >> >> >>>> >>> don't know how feasible this is at the moment.
>> >> >> >>>> >>>
>> >> >> >>>> >>> After this I can compare each noun in the noun phrase with
>> the
>> >> >> >>>> equivalent
>> >> >> >>>> >>> nouns in the categories and based on the number of matches I
>> >> can
>> >> >> >>>> create a
>> >> >> >>>> >>> confidence level.
>> >> >> >>>> >>>
>> >> >> >>>> >>> 3. match the noun of the noun phrase with the rdf:type from
>> >> >> dbpedia
>> >> >> >>>> of the
>> >> >> >>>> >>> named entity. If this matches increase the confidence level.
>> >> >> >>>> >>>
>> >> >> >>>> >>> 4. If there are multiple named entities which can match a
>> >> certain
>> >> >> >>>> noun
>> >> >> >>>> >>> phrase then link the noun phrase with the closest named
>> entity
>> >> >> prior
>> >> >> >>>> to it
>> >> >> >>>> >>> in the text.
>> >> >> >>>> >>>
>> >> >> >>>> >>> What do you think?
>> >> >> >>>> >>>
>> >> >> >>>> >>> Cristian
>> >> >> >>>> >>>
>> >> >> >>>> >>> 2014-01-31 Cristian Petroaca <cristian.petro...@gmail.com>:
>> >> >> >>>> >>>
>> >> >> >>>> >>>  Hi Rafa,
>> >> >> >>>> >>>>
>> >> >> >>>> >>>> I don't yet have a concrete heursitic but I'm working on
>> it.
>> >> I'll
>> >> >> >>>> provide
>> >> >> >>>> >>>> it here so that you guys can give me a feedback on it.
>> >> >> >>>> >>>>
>> >> >> >>>> >>>> What are "locality" features?
>> >> >> >>>> >>>>
>> >> >> >>>> >>>> I looked at Bart and other coref tools such as ArkRef and
>> >> >> >>>> CherryPicker
>> >> >> >>>> >>>> and
>> >> >> >>>> >>>> they don't provide such a coreference.
>> >> >> >>>> >>>>
>> >> >> >>>> >>>> Cristian
>> >> >> >>>> >>>>
>> >> >> >>>> >>>>
>> >> >> >>>> >>>> 2014-01-30 Rafa Haro <rh...@apache.org>:
>> >> >> >>>> >>>>
>> >> >> >>>> >>>> Hi Cristian,
>> >> >> >>>> >>>>
>> >> >> >>>> >>>>> Without having more details about your concrete heuristic,
>> >> in my
>> >> >> >>>> honest
>> >> >> >>>> >>>>> opinion, such approach could produce a lot of false
>> >> positives. I
>> >> >> >>>> don't
>> >> >> >>>> >>>>> know
>> >> >> >>>> >>>>> if you are planning to use some "locality" features to
>> detect
>> >> >> such
>> >> >> >>>> >>>>> coreferences but you need to take into account that it is
>> >> quite
>> >> >> >>>> usual
>> >> >> >>>> >>>>> that
>> >> >> >>>> >>>>> coreferenced mentions can occurs even in different
>> >> paragraphs.
>> >> >> >>>> Although
>> >> >> >>>> >>>>> I'm
>> >> >> >>>> >>>>> not an expert in Natural Language Understanding, I would
>> say
>> >> it
>> >> >> is
>> >> >> >>>> quite
>> >> >> >>>> >>>>> difficult to get decent precision/recall rates for
>> >> coreferencing
>> >> >> >>>> using
>> >> >> >>>> >>>>> fixed rules. Maybe you can give a try to others tools like
>> >> BART
>> >> >> (
>> >> >> >>>> >>>>> http://www.bart-coref.org/).
>> >> >> >>>> >>>>>
>> >> >> >>>> >>>>> Cheers,
>> >> >> >>>> >>>>> Rafa Haro
>> >> >> >>>> >>>>>
>> >> >> >>>> >>>>> El 30/01/14 10:33, Cristian Petroaca escribió:
>> >> >> >>>> >>>>>
>> >> >> >>>> >>>>>   Hi,
>> >> >> >>>> >>>>>
>> >> >> >>>> >>>>>> One of the necessary steps for implementing the Event
>> >> >> extraction
>> >> >> >>>> Engine
>> >> >> >>>> >>>>>> feature :
>> >> https://issues.apache.org/jira/browse/STANBOL-1121is
>> >> >> >>>> to
>> >> >> >>>> >>>>>> have
>> >> >> >>>> >>>>>> coreference resolution in the given text. This is
>> provided
>> >> now
>> >> >> >>>> via the
>> >> >> >>>> >>>>>> stanford-nlp project but as far as I saw this module is
>> >> >> performing
>> >> >> >>>> >>>>>> mostly
>> >> >> >>>> >>>>>> pronomial (He, She) or nominal (Barack Obama and Mr.
>> Obama)
>> >> >> >>>> coreference
>> >> >> >>>> >>>>>> resolution.
>> >> >> >>>> >>>>>>
>> >> >> >>>> >>>>>> In order to get more coreferences from the text I though
>> of
>> >> >> >>>> creating
>> >> >> >>>> >>>>>> some
>> >> >> >>>> >>>>>> logic that would detect this kind of coreference :
>> >> >> >>>> >>>>>> "Apple reaches new profit heights. The software company
>> just
>> >> >> >>>> announced
>> >> >> >>>> >>>>>> its
>> >> >> >>>> >>>>>> 2013 earnings."
>> >> >> >>>> >>>>>> Here "The software company" obviously refers to "Apple".
>> >> >> >>>> >>>>>> So I'd like to detect coreferences of Named Entities
>> which
>> >> are
>> >> >> of
>> >> >> >>>> the
>> >> >> >>>> >>>>>> rdf:type of the Named Entity , in this case "company" and
>> >> also
>> >> >> >>>> have
>> >> >> >>>> >>>>>> attributes which can be found in the dbpedia categories
>> of
>> >> the
>> >> >> >>>> named
>> >> >> >>>> >>>>>> entity, in this case "software".
>> >> >> >>>> >>>>>>
>> >> >> >>>> >>>>>> The detection of coreferences such as "The software
>> >> company" in
>> >> >> >>>> the
>> >> >> >>>> >>>>>> text
>> >> >> >>>> >>>>>> would also be done by either using the new Pos Tag Based
>> >> Phrase
>> >> >> >>>> >>>>>> extraction
>> >> >> >>>> >>>>>> Engine (noun phrases) or by using a dependency tree of
>> the
>> >> >> >>>> sentence and
>> >> >> >>>> >>>>>> picking up only subjects or objects.
>> >> >> >>>> >>>>>>
>> >> >> >>>> >>>>>> At this point I'd like to know if this kind of logic
>> would
>> >> be
>> >> >> >>>> useful
>> >> >> >>>> >>>>>> as a
>> >> >> >>>> >>>>>> separate Enhancement Engine (in case the precision and
>> >> recall
>> >> >> are
>> >> >> >>>> good
>> >> >> >>>> >>>>>> enough) in Stanbol?
>> >> >> >>>> >>>>>>
>> >> >> >>>> >>>>>> Thanks,
>> >> >> >>>> >>>>>> Cristian
>> >> >> >>>> >>>>>>
>> >> >> >>>> >>>>>>
>> >> >> >>>> >>>>>>
>> >> >> >>>> >>
>> >> >> >>>>
>> >> >> >>>>
>> >> >> >>>>
>> >> >> >>>> --
>> >> >> >>>> | Rupert Westenthaler             rupert.westentha...@gmail.com
>> >> >> >>>> | Bodenlehenstraße 11
>> >> ++43-699-11108907
>> >> >> >>>> | A-5500 Bischofshofen
>> >> >> >>>>
>> >> >> >>>
>> >> >> >>>
>> >> >> >>
>> >> >>
>> >> >>
>> >> >>
>> >> >> --
>> >> >> | Rupert Westenthaler             rupert.westentha...@gmail.com
>> >> >> | Bodenlehenstraße 11                             ++43-699-11108907
>> >> >> | A-5500 Bischofshofen
>> >> >>
>> >>
>> >>
>> >>
>> >> --
>> >> | Rupert Westenthaler             rupert.westentha...@gmail.com
>> >> | Bodenlehenstraße 11                             ++43-699-11108907
>> >> | A-5500 Bischofshofen
>> >>
>>
>>
>>
>> --
>> | Rupert Westenthaler             rupert.westentha...@gmail.com
>> | Bodenlehenstraße 11                             ++43-699-11108907
>> | A-5500 Bischofshofen
>>



-- 
| Rupert Westenthaler             rupert.westentha...@gmail.com
| Bodenlehenstraße 11                             ++43-699-11108907
| A-5500 Bischofshofen

Reply via email to