Hi Rupert, This is a reminder in case you missed this e-mail.
Cristian 2013/9/3 Cristian Petroaca <cristian.petro...@gmail.com> > Ok, then to sum it up we would have : > > 1. Coref > > "stanbol.enhancer.nlp.coref" { > "isRepresentative" : true/false, // whether this token or chunk is the > representative mention in the chain > "mentions" : [ { "type" : "Token", // type of element which refers to > this token/chunk > "start": 123 , // start index of the mentioning element > "end": 130 // end index of the mentioning element > }, ... > ], > "class" : ""class" : "org.apache.stanbol.enhancer.nlp.coref.CorefTag" > } > > > 2. Dependency tree > > "stanbol.enhancer.nlp.dependency" : { > "relations" : [ { "tag" : "nsubj", //type of relation - Stanford NLP > notation > "dep" : 12, // type of relation - Stanbol NLP > mapped value - ordinal number in enum Dependency > "role" : "gov/dep", // whether this token is the depender or the dependee > "type" : "Token", // type of element with which this token is in relation > "start" : 123, // start index of the relating token > "end" : 130 // end index of the relating token > }, > ... > ] > "class" : "org.apache.stanbol.enhancer.nlp.dependency.DependencyTag" > } > > > 2013/9/2 Rupert Westenthaler <rupert.westentha...@gmail.com> > >> Hi Cristian, >> >> let me provide some feedback to your proposals: >> >> ### Referring other Spans >> >> Both suggested annotations require to link other spans (Sentence, >> Chunk or Token). For that we should introduce a JSON element used for >> referring those elements and use it for all usages. >> >> In the java model this would allow you to have a reference to the >> other Span (Sentence, Chunk, Token). In the serialized form you would >> have JSON elements with the "type", "start" and "end" attributes as >> those three uniquely identify any span. >> >> Here an example based on the "mention" attribute as defined by the >> proposed "org.apache.stanbol.enhancer.nlp.coref.CorefTag" >> >> ... >> "mentions" : [ { >> "type" : "Token", >> "start": 123 , >> "end": 130 } ,{ >> "type" : "Token", >> "start": 157 , >> "end": 165 }], >> ... >> >> Similar token links in >> "org.apache.stanbol.enhancer.nlp.dependency.DependencyTag" should also >> use this model. >> >> ### Usage of Controlled Vocabularies >> >> In addition the DependencyTag also seams to use a controlled >> vocabulary (e.g. 'nsubj', 'conj_and' ...). In such cases the Stanbol >> NLP module tries to define those in some kind of Ontology. For POS >> tags we use OLIA ontology [1]. This is important as most NLP >> frameworks will use different strings and we need to unify those to >> commons IDs so that component that consume those data do not depend on >> a specific NLP tool. >> >> Because the usage of Ontologies within Java is not well supported. The >> Stanbol NLP module defines Java Enumerations for those Ontologies such >> as the POS type enumeration [2]. >> >> Both the Java Model as well as the JSON serialization do support both >> (1) the lexical tag as used by the NLP tool and (2) the mapped >> concept. In the Java API via two different methods and in the JSON >> serialization via two separate keys. >> >> To make this more clear here an example for a POS annotation of a proper >> noun. >> >> "stanbol.enhancer.nlp.pos" : { >> "tag" : "PN", >> "pos" : 53, >> "class" : "org.apache.stanbol.enhancer.nlp.pos.PosTag", >> "prob" : 0.95 >> } >> >> where >> >> "tag" : "PN" >> >> is the lexical form as used by the NLP tool and >> >> "pos" : 53 >> >> refers to the ordinal number of the entry "ProperNoun" in the POS >> enumeration >> >> IMO the "type" property of DependencyTag should use a similar design. >> >> best >> Rupert >> >> [1] http://olia.nlp2rdf.org/ >> [2] >> http://svn.apache.org/repos/asf/stanbol/trunk/enhancer/generic/nlp/src/main/java/org/apache/stanbol/enhancer/nlp/pos/Pos.java >> >> On Sun, Sep 1, 2013 at 8:09 PM, Cristian Petroaca >> <cristian.petro...@gmail.com> wrote: >> > Sorry, pressed sent too soon :). >> > >> > Continued : >> > >> > nsubj(met-4, Mary-1), conj_and(Mary-1, Tom-3), nsubj(met-4, Tom-3), >> > root(ROOT-0, met-4), nn(today-6, Danny-5), tmod(met-4, today-6)] >> > >> > Given this, we can have for each "Token" an additional dependency >> > annotation : >> > >> > "stanbol.enhancer.nlp.dependency" : { >> > "tag" : //is it necessary? >> > "relations" : [ { "type" : "nsubj", //type of relation >> > "role" : "gov/dep", //whether it is depender or the dependee >> > "dependencyValue" : "met", // the word with which the token has a >> relation >> > "dependencyIndexInSentence" : "2" //the index of the dependency in the >> > current sentence >> > } >> > ... >> > ] >> > "class" : >> > "org.apache.stanbol.enhancer.nlp.dependency.DependencyTag" >> > } >> > >> > 2013/9/1 Cristian Petroaca <cristian.petro...@gmail.com> >> > >> >> Related to the Stanford Dependency Tree Feature, this is the way the >> >> output from the tool looks like for this sentence : "Mary and Tom met >> Danny >> >> today" : >> >> >> >> >> >> 2013/8/30 Cristian Petroaca <cristian.petro...@gmail.com> >> >> >> >>> Hi Rupert, >> >>> >> >>> Ok, so after looking at the JSON output from the Stanford NLP Server >> and >> >>> the coref module I'm thinking I can represent the coreference >> information >> >>> this way: >> >>> Each "Token" or "Chunk" will contain an additional coref annotation >> with >> >>> the following structure : >> >>> >> >>> "stanbol.enhancer.nlp.coref" { >> >>> "tag" : //does this need to exist? >> >>> "isRepresentative" : true/false, // whether this token or chunk is >> >>> the representative mention in the chain >> >>> "mentions" : [ { "sentenceNo" : 1 //the sentence in which the >> mention >> >>> is found >> >>> "startWord" : 2 //the first word making up >> the >> >>> mention >> >>> "endWord" : 3 //the last word making up the >> >>> mention >> >>> }, ... >> >>> ], >> >>> "class" : ""class" : >> "org.apache.stanbol.enhancer.nlp.coref.CorefTag" >> >>> } >> >>> >> >>> The CorefTag should resemble this model. >> >>> >> >>> What do you think? >> >>> >> >>> Cristian >> >>> >> >>> >> >>> 2013/8/24 Rupert Westenthaler <rupert.westentha...@gmail.com> >> >>> >> >>>> Hi Cristian, >> >>>> >> >>>> you can not directly call StanfordNLP components from Stanbol, but >> you >> >>>> have to extend the RESTful service to include the information you >> >>>> need. The main reason for that is that the license of StanfordNLP is >> >>>> not compatible with the Apache Software License. So Stanbol can not >> >>>> directly link to the StanfordNLP API. >> >>>> >> >>>> You will need to >> >>>> >> >>>> 1. define an additional class {yourTag} extends Tag<{yourType}> class >> >>>> in the o.a.s.enhancer.nlp module >> >>>> 2. add JSON parsing and serialization support for this tag to the >> >>>> o.a.s.enhancer.nlp.json module (see e.g. PosTagSupport as an example) >> >>>> >> >>>> As (1) would be necessary anyway the only additional thing you need >> to >> >>>> develop is (2). After that you can add {yourTag} instance to the >> >>>> AnalyzedText in the StanfornNLP integration. The >> >>>> RestfulNlpAnalysisEngine will parse them from the response. All >> >>>> engines executed after the RestfulNlpAnalysisEngine will have access >> >>>> to your annotations. >> >>>> >> >>>> If you have a design for {yourTag} - the model you would like to use >> >>>> to represent your data - I can help with (1) and (2). >> >>>> >> >>>> best >> >>>> Rupert >> >>>> >> >>>> >> >>>> On Fri, Aug 23, 2013 at 5:11 PM, Cristian Petroaca >> >>>> <cristian.petro...@gmail.com> wrote: >> >>>> > Hi Rupert, >> >>>> > >> >>>> > Thanks for the info. Looking at the standbol-stanfordnlp project I >> see >> >>>> that >> >>>> > the stanford nlp is not implemented as an EnhancementEngine but >> rather >> >>>> it >> >>>> > is used directly in a Jetty Server instance. How does that fit >> into the >> >>>> > Stanbol stack? For example how can I call the StanfordNlpAnalyzer's >> >>>> routine >> >>>> > from my TripleExtractionEnhancementEngine which lives in the >> Stanbol >> >>>> stack? >> >>>> > >> >>>> > Thanks, >> >>>> > Cristian >> >>>> > >> >>>> > >> >>>> > 2013/8/12 Rupert Westenthaler <rupert.westentha...@gmail.com> >> >>>> > >> >>>> >> Hi Cristian, >> >>>> >> >> >>>> >> Sorry for the late response, but I was offline for the last two >> weeks >> >>>> >> >> >>>> >> On Fri, Aug 2, 2013 at 9:19 PM, Cristian Petroaca >> >>>> >> <cristian.petro...@gmail.com> wrote: >> >>>> >> > Hi Rupert, >> >>>> >> > >> >>>> >> > After doing some tests it seems that the Stanford NLP >> coreference >> >>>> module >> >>>> >> is >> >>>> >> > much more accurate than the Open NLP one.So I decided to extend >> >>>> Stanford >> >>>> >> > NLP to add coreference there. >> >>>> >> >> >>>> >> The Stanford NLP integration is not part of the Stanbol codebase >> >>>> >> because the licenses are not compatible. >> >>>> >> >> >>>> >> You can find the Stanford NLP integration on >> >>>> >> >> >>>> >> https://github.com/westei/stanbol-stanfordnlp >> >>>> >> >> >>>> >> just create a fork and send pull requests. >> >>>> >> >> >>>> >> >> >>>> >> > Could you add the necessary projects on the branch? And also >> remove >> >>>> the >> >>>> >> > Open NLP ones? >> >>>> >> > >> >>>> >> >> >>>> >> Currently the branch >> >>>> >> >> >>>> >> >> >>>> >> >> >>>> >> http://svn.apache.org/repos/asf/stanbol/branches/nlp-dep-tree-and-co-ref/ >> >>>> >> >> >>>> >> only contains the "nlp" and the "nlp-json" modules. IMO those >> should >> >>>> >> be enough for adding coreference support. >> >>>> >> >> >>>> >> IMO you will need to >> >>>> >> >> >>>> >> * add an model for representing coreference to the nlp module >> >>>> >> * add parsing and serializing support to the nlp-json module >> >>>> >> * add the implementation to your fork of the stanbol-stanfordnlp >> >>>> project >> >>>> >> >> >>>> >> best >> >>>> >> Rupert >> >>>> >> >> >>>> >> >> >>>> >> >> >>>> >> > Thanks, >> >>>> >> > Cristian >> >>>> >> > >> >>>> >> > >> >>>> >> > 2013/7/5 Rupert Westenthaler <rupert.westentha...@gmail.com> >> >>>> >> > >> >>>> >> >> Hi Cristian, >> >>>> >> >> >> >>>> >> >> I created the branch at >> >>>> >> >> >> >>>> >> >> >> >>>> >> >> >> >>>> >> >> >>>> >> http://svn.apache.org/repos/asf/stanbol/branches/nlp-dep-tree-and-co-ref/ >> >>>> >> >> >> >>>> >> >> ATM in contains only the "nlp" and "nlp-json" module. Let me >> know >> >>>> if >> >>>> >> >> you would like to have more >> >>>> >> >> >> >>>> >> >> best >> >>>> >> >> Rupert >> >>>> >> >> >> >>>> >> >> >> >>>> >> >> >> >>>> >> >> On Thu, Jul 4, 2013 at 10:14 AM, Cristian Petroaca >> >>>> >> >> <cristian.petro...@gmail.com> wrote: >> >>>> >> >> > Hi Rupert, >> >>>> >> >> > >> >>>> >> >> > I created jiras : >> >>>> https://issues.apache.org/jira/browse/STANBOL-1132and >> >>>> >> >> > https://issues.apache.org/jira/browse/STANBOL-1133. The >> >>>> original one >> >>>> >> in >> >>>> >> >> > dependent upon these. >> >>>> >> >> > Please let me know when I can start using the branch. >> >>>> >> >> > >> >>>> >> >> > Thanks, >> >>>> >> >> > Cristian >> >>>> >> >> > >> >>>> >> >> > >> >>>> >> >> > 2013/6/27 Cristian Petroaca <cristian.petro...@gmail.com> >> >>>> >> >> > >> >>>> >> >> >> >> >>>> >> >> >> >> >>>> >> >> >> >> >>>> >> >> >> 2013/6/27 Rupert Westenthaler < >> rupert.westentha...@gmail.com> >> >>>> >> >> >> >> >>>> >> >> >>> On Thu, Jun 27, 2013 at 3:12 PM, Cristian Petroaca >> >>>> >> >> >>> <cristian.petro...@gmail.com> wrote: >> >>>> >> >> >>> > Sorry, I meant the Stanbol NLP API, not Stanford in my >> >>>> previous >> >>>> >> >> e-mail. >> >>>> >> >> >>> By >> >>>> >> >> >>> > the way, does Open NLP have the ability to build >> dependency >> >>>> trees? >> >>>> >> >> >>> > >> >>>> >> >> >>> >> >>>> >> >> >>> AFAIK OpenNLP does not provide this feature. >> >>>> >> >> >>> >> >>>> >> >> >> >> >>>> >> >> >> Then , since the Stanford NLP lib is also integrated into >> >>>> Stanbol, >> >>>> >> I'll >> >>>> >> >> >> take a look at how I can extend its integration to include >> the >> >>>> >> >> dependency >> >>>> >> >> >> tree feature. >> >>>> >> >> >> >> >>>> >> >> >>> >> >>>> >> >> >>> >> >>>> >> >> >> > >> >>>> >> >> >>> > 2013/6/23 Cristian Petroaca <cristian.petro...@gmail.com >> > >> >>>> >> >> >>> > >> >>>> >> >> >>> >> Hi Rupert, >> >>>> >> >> >>> >> >> >>>> >> >> >>> >> I created jira >> >>>> >> https://issues.apache.org/jira/browse/STANBOL-1121. >> >>>> >> >> >>> >> As you suggested I would start with extending the >> Stanford >> >>>> NLP >> >>>> >> with >> >>>> >> >> >>> >> co-reference resolution but I think also with dependency >> >>>> trees >> >>>> >> >> because >> >>>> >> >> >>> I >> >>>> >> >> >>> >> also need to know the Subject of the sentence and the >> object >> >>>> >> that it >> >>>> >> >> >>> >> affects, right? >> >>>> >> >> >>> >> >> >>>> >> >> >>> >> Given that I need to extend the Stanford NLP API in >> Stanbol >> >>>> for >> >>>> >> >> >>> >> co-reference and dependency trees, how do I proceed with >> >>>> this? >> >>>> >> Do I >> >>>> >> >> >>> create >> >>>> >> >> >>> >> 2 new sub-tasks to the already opened Jira? After that >> can I >> >>>> >> start >> >>>> >> >> >>> >> implementing on my local copy of Stanbol and when I'm >> done >> >>>> I'll >> >>>> >> send >> >>>> >> >> >>> you >> >>>> >> >> >>> >> guys the patch fo review? >> >>>> >> >> >>> >> >> >>>> >> >> >>> >> >>>> >> >> >>> I would create two "New Feature" type Issues one for adding >> >>>> support >> >>>> >> >> >>> for "dependency trees" and the other for "co-reference" >> >>>> support. You >> >>>> >> >> >>> should also define "depends on" relations between >> STANBOL-1121 >> >>>> and >> >>>> >> >> >>> those two new issues. >> >>>> >> >> >>> >> >>>> >> >> >>> Sub-task could also work, but as adding those features >> would >> >>>> be also >> >>>> >> >> >>> interesting for other things I would rather define them as >> >>>> separate >> >>>> >> >> >>> issues. >> >>>> >> >> >>> >> >>>> >> >> >>> >> >>>> >> >> >> 2 New Features connected with the original jira it is then. >> >>>> >> >> >> >> >>>> >> >> >> >> >>>> >> >> >>> If you would prefer to work in an own branch please tell >> me. >> >>>> This >> >>>> >> >> >>> could have the advantage that patches would not be >> affected by >> >>>> >> changes >> >>>> >> >> >>> in the trunk. >> >>>> >> >> >>> >> >>>> >> >> >>> Yes, a separate branch sounds good. >> >>>> >> >> >> >> >>>> >> >> >> best >> >>>> >> >> >>> Rupert >> >>>> >> >> >>> >> >>>> >> >> >>> >> Regards, >> >>>> >> >> >>> >> Cristian >> >>>> >> >> >>> >> >> >>>> >> >> >>> >> >> >>>> >> >> >>> >> 2013/6/18 Rupert Westenthaler < >> >>>> rupert.westentha...@gmail.com> >> >>>> >> >> >>> >> >> >>>> >> >> >>> >>> On Mon, Jun 17, 2013 at 10:18 PM, Cristian Petroaca >> >>>> >> >> >>> >>> <cristian.petro...@gmail.com> wrote: >> >>>> >> >> >>> >>> > Hi Rupert, >> >>>> >> >> >>> >>> > >> >>>> >> >> >>> >>> > Agreed on the >> >>>> >> >> >>> SettingAnnotation/ParticipantAnnotation/OccurentAnnotation >> >>>> >> >> >>> >>> > data structure. >> >>>> >> >> >>> >>> > >> >>>> >> >> >>> >>> > Should I open up a Jira for all of this in order to >> >>>> >> encapsulate >> >>>> >> >> this >> >>>> >> >> >>> >>> > information and establish the goals and these initial >> >>>> steps >> >>>> >> >> towards >> >>>> >> >> >>> >>> these >> >>>> >> >> >>> >>> > goals? >> >>>> >> >> >>> >>> >> >>>> >> >> >>> >>> Yes please. A JIRA issue for this work would be great. >> >>>> >> >> >>> >>> >> >>>> >> >> >>> >>> > How should I proceed further? Should I create some >> design >> >>>> >> >> documents >> >>>> >> >> >>> that >> >>>> >> >> >>> >>> > need to be reviewed? >> >>>> >> >> >>> >>> >> >>>> >> >> >>> >>> Usually it is the best to write design related text >> >>>> directly in >> >>>> >> >> JIRA >> >>>> >> >> >>> >>> by using Markdown [1] syntax. This will allow us later >> to >> >>>> use >> >>>> >> this >> >>>> >> >> >>> >>> text directly for the documentation on the Stanbol >> Webpage. >> >>>> >> >> >>> >>> >> >>>> >> >> >>> >>> best >> >>>> >> >> >>> >>> Rupert >> >>>> >> >> >>> >>> >> >>>> >> >> >>> >>> >> >>>> >> >> >>> >>> [1] http://daringfireball.net/projects/markdown/ >> >>>> >> >> >>> >>> > >> >>>> >> >> >>> >>> > Regards, >> >>>> >> >> >>> >>> > Cristian >> >>>> >> >> >>> >>> > >> >>>> >> >> >>> >>> > >> >>>> >> >> >>> >>> > 2013/6/17 Rupert Westenthaler < >> >>>> rupert.westentha...@gmail.com> >> >>>> >> >> >>> >>> > >> >>>> >> >> >>> >>> >> On Thu, Jun 13, 2013 at 8:22 PM, Cristian Petroaca >> >>>> >> >> >>> >>> >> <cristian.petro...@gmail.com> wrote: >> >>>> >> >> >>> >>> >> > HI Rupert, >> >>>> >> >> >>> >>> >> > >> >>>> >> >> >>> >>> >> > First of all thanks for the detailed suggestions. >> >>>> >> >> >>> >>> >> > >> >>>> >> >> >>> >>> >> > 2013/6/12 Rupert Westenthaler < >> >>>> >> rupert.westentha...@gmail.com> >> >>>> >> >> >>> >>> >> > >> >>>> >> >> >>> >>> >> >> Hi Cristian, all >> >>>> >> >> >>> >>> >> >> >> >>>> >> >> >>> >>> >> >> really interesting use case! >> >>>> >> >> >>> >>> >> >> >> >>>> >> >> >>> >>> >> >> In this mail I will try to give some suggestions >> on >> >>>> how >> >>>> >> this >> >>>> >> >> >>> could >> >>>> >> >> >>> >>> >> >> work out. This suggestions are mainly based on >> >>>> experiences >> >>>> >> >> and >> >>>> >> >> >>> >>> lessons >> >>>> >> >> >>> >>> >> >> learned in the LIVE [2] project where we built an >> >>>> >> information >> >>>> >> >> >>> system >> >>>> >> >> >>> >>> >> >> for the Olympic Games in Peking. While this >> Project >> >>>> >> excluded >> >>>> >> >> the >> >>>> >> >> >>> >>> >> >> extraction of Events from unstructured text >> (because >> >>>> the >> >>>> >> >> Olympic >> >>>> >> >> >>> >>> >> >> Information System was already providing event >> data >> >>>> as XML >> >>>> >> >> >>> messages) >> >>>> >> >> >>> >>> >> >> the semantic search capabilities of this system >> >>>> where very >> >>>> >> >> >>> similar >> >>>> >> >> >>> >>> as >> >>>> >> >> >>> >>> >> >> the one described by your use case. >> >>>> >> >> >>> >>> >> >> >> >>>> >> >> >>> >>> >> >> IMHO you are not only trying to extract >> relations, >> >>>> but a >> >>>> >> >> formal >> >>>> >> >> >>> >>> >> >> representation of the situation described by the >> >>>> text. So >> >>>> >> >> lets >> >>>> >> >> >>> >>> assume >> >>>> >> >> >>> >>> >> >> that the goal is to Annotate a Setting (or >> Situation) >> >>>> >> >> described >> >>>> >> >> >>> in >> >>>> >> >> >>> >>> the >> >>>> >> >> >>> >>> >> >> text - a fise:SettingAnnotation. >> >>>> >> >> >>> >>> >> >> >> >>>> >> >> >>> >>> >> >> The DOLCE foundational ontology [1] gives some >> >>>> advices on >> >>>> >> >> how to >> >>>> >> >> >>> >>> model >> >>>> >> >> >>> >>> >> >> those. The important relation for modeling this >> >>>> >> >> Participation: >> >>>> >> >> >>> >>> >> >> >> >>>> >> >> >>> >>> >> >> PC(x, y, t) → (ED(x) ∧ PD(y) ∧ T(t)) >> >>>> >> >> >>> >>> >> >> >> >>>> >> >> >>> >>> >> >> where .. >> >>>> >> >> >>> >>> >> >> >> >>>> >> >> >>> >>> >> >> * ED are Endurants (continuants): Endurants do >> have >> >>>> an >> >>>> >> >> >>> identity so >> >>>> >> >> >>> >>> we >> >>>> >> >> >>> >>> >> >> would typically refer to them as Entities >> referenced >> >>>> by a >> >>>> >> >> >>> setting. >> >>>> >> >> >>> >>> >> >> Note that this includes physical, non-physical as >> >>>> well as >> >>>> >> >> >>> >>> >> >> social-objects. >> >>>> >> >> >>> >>> >> >> * PD are Perdurants (occurrents): Perdurants >> are >> >>>> >> entities >> >>>> >> >> that >> >>>> >> >> >>> >>> >> >> happen in time. This refers to Events, >> Activities ... >> >>>> >> >> >>> >>> >> >> * PC are Participation: It is an time indexed >> >>>> relation >> >>>> >> where >> >>>> >> >> >>> >>> >> >> Endurants participate in Perdurants >> >>>> >> >> >>> >>> >> >> >> >>>> >> >> >>> >>> >> >> Modeling this in RDF requires to define some >> >>>> intermediate >> >>>> >> >> >>> resources >> >>>> >> >> >>> >>> >> >> because RDF does not allow for n-ary relations. >> >>>> >> >> >>> >>> >> >> >> >>>> >> >> >>> >>> >> >> * fise:SettingAnnotation: It is really handy to >> >>>> define >> >>>> >> one >> >>>> >> >> >>> resource >> >>>> >> >> >>> >>> >> >> being the context for all described data. I would >> >>>> call >> >>>> >> this >> >>>> >> >> >>> >>> >> >> "fise:SettingAnnotation" and define it as a >> >>>> sub-concept to >> >>>> >> >> >>> >>> >> >> fise:Enhancement. All further enhancement about >> the >> >>>> >> extracted >> >>>> >> >> >>> >>> Setting >> >>>> >> >> >>> >>> >> >> would define a "fise:in-setting" relation to it. >> >>>> >> >> >>> >>> >> >> >> >>>> >> >> >>> >>> >> >> * fise:ParticipantAnnotation: Is used to >> annotate >> >>>> that >> >>>> >> >> >>> Endurant is >> >>>> >> >> >>> >>> >> >> participating on a setting (fise:in-setting >> >>>> >> >> >>> fise:SettingAnnotation). >> >>>> >> >> >>> >>> >> >> The Endurant itself is described by existing >> >>>> >> >> fise:TextAnnotaion >> >>>> >> >> >>> (the >> >>>> >> >> >>> >>> >> >> mentions) and fise:EntityAnnotation (suggested >> >>>> Entities). >> >>>> >> >> >>> Basically >> >>>> >> >> >>> >>> >> >> the fise:ParticipantAnnotation will allow an >> >>>> >> >> EnhancementEngine >> >>>> >> >> >>> to >> >>>> >> >> >>> >>> >> >> state that several mentions (in possible >> different >> >>>> >> >> sentences) do >> >>>> >> >> >>> >>> >> >> represent the same Endurant as participating in >> the >> >>>> >> Setting. >> >>>> >> >> In >> >>>> >> >> >>> >>> >> >> addition it would be possible to use the dc:type >> >>>> property >> >>>> >> >> >>> (similar >> >>>> >> >> >>> >>> as >> >>>> >> >> >>> >>> >> >> for fise:TextAnnotation) to refer to the role(s) >> of >> >>>> an >> >>>> >> >> >>> participant >> >>>> >> >> >>> >>> >> >> (e.g. the set: Agent (intensionally performs an >> >>>> action) >> >>>> >> Cause >> >>>> >> >> >>> >>> >> >> (unintentionally e.g. a mud slide), Patient (a >> >>>> passive >> >>>> >> role >> >>>> >> >> in >> >>>> >> >> >>> an >> >>>> >> >> >>> >>> >> >> activity) and Instrument (aids an process)), but >> I am >> >>>> >> >> wondering >> >>>> >> >> >>> if >> >>>> >> >> >>> >>> one >> >>>> >> >> >>> >>> >> >> could extract those information. >> >>>> >> >> >>> >>> >> >> >> >>>> >> >> >>> >>> >> >> * fise:OccurrentAnnotation: is used to annotate a >> >>>> >> Perdurant >> >>>> >> >> in >> >>>> >> >> >>> the >> >>>> >> >> >>> >>> >> >> context of the Setting. Also >> >>>> fise:OccurrentAnnotation can >> >>>> >> >> link >> >>>> >> >> >>> to >> >>>> >> >> >>> >>> >> >> fise:TextAnnotaion (typically verbs in the text >> >>>> defining >> >>>> >> the >> >>>> >> >> >>> >>> >> >> perdurant) as well as fise:EntityAnnotation >> >>>> suggesting >> >>>> >> well >> >>>> >> >> >>> known >> >>>> >> >> >>> >>> >> >> Events in a knowledge base (e.g. a Election in a >> >>>> country, >> >>>> >> or >> >>>> >> >> an >> >>>> >> >> >>> >>> >> >> upraising ...). In addition >> fise:OccurrentAnnotation >> >>>> can >> >>>> >> >> define >> >>>> >> >> >>> >>> >> >> dc:has-participant links to >> >>>> fise:ParticipantAnnotation. In >> >>>> >> >> this >> >>>> >> >> >>> case >> >>>> >> >> >>> >>> >> >> it is explicitly stated hat an Endurant (the >> >>>> >> >> >>> >>> >> >> fise:ParticipantAnnotation) involved in this >> >>>> Perturant >> >>>> >> (the >> >>>> >> >> >>> >>> >> >> fise:OccurrentAnnotation). As Occurrences are >> >>>> temporal >> >>>> >> >> indexed >> >>>> >> >> >>> this >> >>>> >> >> >>> >>> >> >> annotation should also support properties for >> >>>> defining the >> >>>> >> >> >>> >>> >> >> xsd:dateTime for the start/end. >> >>>> >> >> >>> >>> >> >> >> >>>> >> >> >>> >>> >> >> >> >>>> >> >> >>> >>> >> >> Indeed, an event based data structure makes a >> lot of >> >>>> sense >> >>>> >> >> with >> >>>> >> >> >>> the >> >>>> >> >> >>> >>> >> remark >> >>>> >> >> >>> >>> >> > that you probably won't be able to always extract >> the >> >>>> date >> >>>> >> >> for a >> >>>> >> >> >>> >>> given >> >>>> >> >> >>> >>> >> > setting(situation). >> >>>> >> >> >>> >>> >> > There are 2 thing which are unclear though. >> >>>> >> >> >>> >>> >> > >> >>>> >> >> >>> >>> >> > 1. Perdurant : You could have situations in which >> the >> >>>> >> object >> >>>> >> >> upon >> >>>> >> >> >>> >>> which >> >>>> >> >> >>> >>> >> the >> >>>> >> >> >>> >>> >> > Subject ( or Endurant ) is acting is not a >> transitory >> >>>> >> object ( >> >>>> >> >> >>> such >> >>>> >> >> >>> >>> as an >> >>>> >> >> >>> >>> >> > event, activity ) but rather another Endurant. For >> >>>> example >> >>>> >> we >> >>>> >> >> can >> >>>> >> >> >>> >>> have >> >>>> >> >> >>> >>> >> the >> >>>> >> >> >>> >>> >> > phrase "USA invades Irak" where "USA" is the >> Endurant >> >>>> ( >> >>>> >> >> Subject ) >> >>>> >> >> >>> >>> which >> >>>> >> >> >>> >>> >> > performs the action of "invading" on another >> >>>> Eundurant, >> >>>> >> namely >> >>>> >> >> >>> >>> "Irak". >> >>>> >> >> >>> >>> >> > >> >>>> >> >> >>> >>> >> >> >>>> >> >> >>> >>> >> By using CAOS, USA would be the Agent and Iraq the >> >>>> Patient. >> >>>> >> Both >> >>>> >> >> >>> are >> >>>> >> >> >>> >>> >> Endurants. The activity "invading" would be the >> >>>> Perdurant. So >> >>>> >> >> >>> ideally >> >>>> >> >> >>> >>> >> you would have a "fise:SettingAnnotation" with: >> >>>> >> >> >>> >>> >> >> >>>> >> >> >>> >>> >> * fise:ParticipantAnnotation for USA with the >> dc:type >> >>>> >> >> caos:Agent, >> >>>> >> >> >>> >>> >> linking to a fise:TextAnnotation for "USA" and a >> >>>> >> >> >>> fise:EntityAnnotation >> >>>> >> >> >>> >>> >> linking to dbpedia:United_States >> >>>> >> >> >>> >>> >> * fise:ParticipantAnnotation for Iraq with the >> dc:type >> >>>> >> >> >>> caos:Patient, >> >>>> >> >> >>> >>> >> linking to a fise:TextAnnotation for "Irak" and a >> >>>> >> >> >>> >>> >> fise:EntityAnnotation linking to dbpedia:Iraq >> >>>> >> >> >>> >>> >> * fise:OccurrentAnnotation for "invades" with the >> >>>> dc:type >> >>>> >> >> >>> >>> >> caos:Activity, linking to a fise:TextAnnotation for >> >>>> "invades" >> >>>> >> >> >>> >>> >> >> >>>> >> >> >>> >>> >> > 2. Where does the verb, which links the Subject >> and >> >>>> the >> >>>> >> Object >> >>>> >> >> >>> come >> >>>> >> >> >>> >>> into >> >>>> >> >> >>> >>> >> > this? I imagined that the Endurant would have a >> >>>> >> dc:"property" >> >>>> >> >> >>> where >> >>>> >> >> >>> >>> the >> >>>> >> >> >>> >>> >> > property = verb which links to the Object in noun >> >>>> form. For >> >>>> >> >> >>> example >> >>>> >> >> >>> >>> take >> >>>> >> >> >>> >>> >> > again the sentence "USA invades Irak". You would >> have >> >>>> the >> >>>> >> >> "USA" >> >>>> >> >> >>> >>> Entity >> >>>> >> >> >>> >>> >> with >> >>>> >> >> >>> >>> >> > dc:invader which points to the Object "Irak". The >> >>>> Endurant >> >>>> >> >> would >> >>>> >> >> >>> >>> have as >> >>>> >> >> >>> >>> >> > many dc:"property" elements as there are verbs >> which >> >>>> link >> >>>> >> it >> >>>> >> >> to >> >>>> >> >> >>> an >> >>>> >> >> >>> >>> >> Object. >> >>>> >> >> >>> >>> >> >> >>>> >> >> >>> >>> >> As explained above you would have a >> >>>> fise:OccurrentAnnotation >> >>>> >> >> that >> >>>> >> >> >>> >>> >> represents the Perdurant. The information that the >> >>>> activity >> >>>> >> >> >>> mention in >> >>>> >> >> >>> >>> >> the text is "invades" would be by linking to a >> >>>> >> >> >>> fise:TextAnnotation. If >> >>>> >> >> >>> >>> >> you can also provide an Ontology for Tasks that >> defines >> >>>> >> >> >>> >>> >> "myTasks:invade" the fise:OccurrentAnnotation could >> >>>> also link >> >>>> >> >> to an >> >>>> >> >> >>> >>> >> fise:EntityAnnotation for this concept. >> >>>> >> >> >>> >>> >> >> >>>> >> >> >>> >>> >> best >> >>>> >> >> >>> >>> >> Rupert >> >>>> >> >> >>> >>> >> >> >>>> >> >> >>> >>> >> > >> >>>> >> >> >>> >>> >> > ### Consuming the data: >> >>>> >> >> >>> >>> >> >> >> >>>> >> >> >>> >>> >> >> I think this model should be sufficient for >> >>>> use-cases as >> >>>> >> >> >>> described >> >>>> >> >> >>> >>> by >> >>>> >> >> >>> >>> >> you. >> >>>> >> >> >>> >>> >> >> >> >>>> >> >> >>> >>> >> >> Users would be able to consume data on the >> setting >> >>>> level. >> >>>> >> >> This >> >>>> >> >> >>> can >> >>>> >> >> >>> >>> be >> >>>> >> >> >>> >>> >> >> done my simple retrieving all >> >>>> fise:ParticipantAnnotation >> >>>> >> as >> >>>> >> >> >>> well as >> >>>> >> >> >>> >>> >> >> fise:OccurrentAnnotation linked with a setting. >> BTW >> >>>> this >> >>>> >> was >> >>>> >> >> the >> >>>> >> >> >>> >>> >> >> approach used in LIVE [2] for semantic search. It >> >>>> allows >> >>>> >> >> >>> queries for >> >>>> >> >> >>> >>> >> >> Settings that involve specific Entities e.g. you >> >>>> could >> >>>> >> filter >> >>>> >> >> >>> for >> >>>> >> >> >>> >>> >> >> Settings that involve a {Person}, >> >>>> activities:Arrested and >> >>>> >> a >> >>>> >> >> >>> specific >> >>>> >> >> >>> >>> >> >> {Upraising}. However note that with this approach >> >>>> you will >> >>>> >> >> get >> >>>> >> >> >>> >>> results >> >>>> >> >> >>> >>> >> >> for Setting where the {Person} participated and >> an >> >>>> other >> >>>> >> >> person >> >>>> >> >> >>> was >> >>>> >> >> >>> >>> >> >> arrested. >> >>>> >> >> >>> >>> >> >> >> >>>> >> >> >>> >>> >> >> An other possibility would be to process >> enhancement >> >>>> >> results >> >>>> >> >> on >> >>>> >> >> >>> the >> >>>> >> >> >>> >>> >> >> fise:OccurrentAnnotation. This would allow to a >> much >> >>>> >> higher >> >>>> >> >> >>> >>> >> >> granularity level (e.g. it would allow to >> correctly >> >>>> answer >> >>>> >> >> the >> >>>> >> >> >>> query >> >>>> >> >> >>> >>> >> >> used as an example above). But I am wondering if >> the >> >>>> >> quality >> >>>> >> >> of >> >>>> >> >> >>> the >> >>>> >> >> >>> >>> >> >> Setting extraction will be sufficient for this. I >> >>>> have >> >>>> >> also >> >>>> >> >> >>> doubts >> >>>> >> >> >>> >>> if >> >>>> >> >> >>> >>> >> >> this can be still realized by using semantic >> >>>> indexing to >> >>>> >> >> Apache >> >>>> >> >> >>> Solr >> >>>> >> >> >>> >>> >> >> or if it would be better/necessary to store >> results >> >>>> in a >> >>>> >> >> >>> TripleStore >> >>>> >> >> >>> >>> >> >> and using SPARQL for retrieval. >> >>>> >> >> >>> >>> >> >> >> >>>> >> >> >>> >>> >> >> The methodology and query language used by YAGO >> [3] >> >>>> is >> >>>> >> also >> >>>> >> >> very >> >>>> >> >> >>> >>> >> >> relevant for this (especially note chapter 7 >> SPOTL(X) >> >>>> >> >> >>> >>> Representation). >> >>>> >> >> >>> >>> >> >> >> >>>> >> >> >>> >>> >> >> An other related Topic is the enrichment of >> Entities >> >>>> >> >> (especially >> >>>> >> >> >>> >>> >> >> Events) in knowledge bases based on Settings >> >>>> extracted >> >>>> >> form >> >>>> >> >> >>> >>> Documents. >> >>>> >> >> >>> >>> >> >> As per definition - in DOLCE - Perdurants are >> >>>> temporal >> >>>> >> >> indexed. >> >>>> >> >> >>> That >> >>>> >> >> >>> >>> >> >> means that at the time when added to a knowledge >> >>>> base they >> >>>> >> >> might >> >>>> >> >> >>> >>> still >> >>>> >> >> >>> >>> >> >> be in process. So the creation, enriching and >> >>>> refinement >> >>>> >> of >> >>>> >> >> such >> >>>> >> >> >>> >>> >> >> Entities in a the knowledge base seams to be >> >>>> critical for >> >>>> >> a >> >>>> >> >> >>> System >> >>>> >> >> >>> >>> >> >> like described in your use-case. >> >>>> >> >> >>> >>> >> >> >> >>>> >> >> >>> >>> >> >> On Tue, Jun 11, 2013 at 9:09 PM, Cristian >> Petroaca >> >>>> >> >> >>> >>> >> >> <cristian.petro...@gmail.com> wrote: >> >>>> >> >> >>> >>> >> >> > >> >>>> >> >> >>> >>> >> >> > First of all I have to mention that I am new >> in the >> >>>> >> field >> >>>> >> >> of >> >>>> >> >> >>> >>> semantic >> >>>> >> >> >>> >>> >> >> > technologies, I've started to read about them >> in >> >>>> the >> >>>> >> last >> >>>> >> >> 4-5 >> >>>> >> >> >>> >>> >> >> months.Having >> >>>> >> >> >>> >>> >> >> > said that I have a high level overview of what >> is >> >>>> a good >> >>>> >> >> >>> approach >> >>>> >> >> >>> >>> to >> >>>> >> >> >>> >>> >> >> solve >> >>>> >> >> >>> >>> >> >> > this problem. There are a number of papers on >> the >> >>>> >> internet >> >>>> >> >> >>> which >> >>>> >> >> >>> >>> >> describe >> >>>> >> >> >>> >>> >> >> > what steps need to be taken such as : named >> entity >> >>>> >> >> >>> recognition, >> >>>> >> >> >>> >>> >> >> > co-reference resolution, pos tagging and >> others. >> >>>> >> >> >>> >>> >> >> >> >>>> >> >> >>> >>> >> >> The Stanbol NLP processing module currently only >> >>>> supports >> >>>> >> >> >>> sentence >> >>>> >> >> >>> >>> >> >> detection, tokenization, POS tagging, Chunking, >> NER >> >>>> and >> >>>> >> >> lemma. >> >>>> >> >> >>> >>> support >> >>>> >> >> >>> >>> >> >> for co-reference resolution and dependency trees >> is >> >>>> >> currently >> >>>> >> >> >>> >>> missing. >> >>>> >> >> >>> >>> >> >> >> >>>> >> >> >>> >>> >> >> Stanford NLP is already integrated with Stanbol >> [4]. >> >>>> At >> >>>> >> the >> >>>> >> >> >>> moment >> >>>> >> >> >>> >>> it >> >>>> >> >> >>> >>> >> >> only supports English, but I do already work to >> >>>> include >> >>>> >> the >> >>>> >> >> >>> other >> >>>> >> >> >>> >>> >> >> supported languages. Other NLP framework that is >> >>>> already >> >>>> >> >> >>> integrated >> >>>> >> >> >>> >>> >> >> with Stanbol are Freeling [5] and Talismane [6]. >> But >> >>>> note >> >>>> >> >> that >> >>>> >> >> >>> for >> >>>> >> >> >>> >>> all >> >>>> >> >> >>> >>> >> >> those the integration excludes support for >> >>>> co-reference >> >>>> >> and >> >>>> >> >> >>> >>> dependency >> >>>> >> >> >>> >>> >> >> trees. >> >>>> >> >> >>> >>> >> >> >> >>>> >> >> >>> >>> >> >> Anyways I am confident that one can implement a >> first >> >>>> >> >> prototype >> >>>> >> >> >>> by >> >>>> >> >> >>> >>> >> >> only using Sentences and POS tags and - if >> available >> >>>> - >> >>>> >> Chunks >> >>>> >> >> >>> (e.g. >> >>>> >> >> >>> >>> >> >> Noun phrases). >> >>>> >> >> >>> >>> >> >> >> >>>> >> >> >>> >>> >> >> >> >>>> >> >> >>> >>> >> > I assume that in the Stanbol context, a feature >> like >> >>>> >> Relation >> >>>> >> >> >>> >>> extraction >> >>>> >> >> >>> >>> >> > would be implemented as an EnhancementEngine? >> >>>> >> >> >>> >>> >> > What kind of effort would be required for a >> >>>> co-reference >> >>>> >> >> >>> resolution >> >>>> >> >> >>> >>> tool >> >>>> >> >> >>> >>> >> > integration into Stanbol? >> >>>> >> >> >>> >>> >> > >> >>>> >> >> >>> >>> >> >> >>>> >> >> >>> >>> >> Yes in the end it would be an EnhancementEngine. But >> >>>> before >> >>>> >> we >> >>>> >> >> can >> >>>> >> >> >>> >>> >> build such an engine we would need to >> >>>> >> >> >>> >>> >> >> >>>> >> >> >>> >>> >> * extend the Stanbol NLP processing API with >> >>>> Annotations for >> >>>> >> >> >>> >>> co-reference >> >>>> >> >> >>> >>> >> * add support for JSON Serialisation/Parsing for >> those >> >>>> >> >> annotation >> >>>> >> >> >>> so >> >>>> >> >> >>> >>> >> that the RESTful NLP Analysis Service can provide >> >>>> >> co-reference >> >>>> >> >> >>> >>> >> information >> >>>> >> >> >>> >>> >> >> >>>> >> >> >>> >>> >> > At this moment I'll be focusing on 2 aspects: >> >>>> >> >> >>> >>> >> > >> >>>> >> >> >>> >>> >> > 1. Determine the best data structure to >> encapsulate >> >>>> the >> >>>> >> >> extracted >> >>>> >> >> >>> >>> >> > information. I'll take a closer look at Dolce. >> >>>> >> >> >>> >>> >> >> >>>> >> >> >>> >>> >> Don't make to to complex. Defining a proper >> structure to >> >>>> >> >> represent >> >>>> >> >> >>> >>> >> Events will only pay-off if we can also successfully >> >>>> extract >> >>>> >> >> such >> >>>> >> >> >>> >>> >> information form processed texts. >> >>>> >> >> >>> >>> >> >> >>>> >> >> >>> >>> >> I would start with >> >>>> >> >> >>> >>> >> >> >>>> >> >> >>> >>> >> * fise:SettingAnnotation >> >>>> >> >> >>> >>> >> * {fise:Enhancement} metadata >> >>>> >> >> >>> >>> >> >> >>>> >> >> >>> >>> >> * fise:ParticipantAnnotation >> >>>> >> >> >>> >>> >> * {fise:Enhancement} metadata >> >>>> >> >> >>> >>> >> * fise:inSetting {settingAnnotation} >> >>>> >> >> >>> >>> >> * fise:hasMention {textAnnotation} >> >>>> >> >> >>> >>> >> * fise:suggestion {entityAnnotation} (multiple >> if >> >>>> there >> >>>> >> are >> >>>> >> >> >>> more >> >>>> >> >> >>> >>> >> suggestions) >> >>>> >> >> >>> >>> >> * dc:type one of fise:Agent, fise:Patient, >> >>>> >> fise:Instrument, >> >>>> >> >> >>> >>> fise:Cause >> >>>> >> >> >>> >>> >> >> >>>> >> >> >>> >>> >> * fise:OccurrentAnnotation >> >>>> >> >> >>> >>> >> * {fise:Enhancement} metadata >> >>>> >> >> >>> >>> >> * fise:inSetting {settingAnnotation} >> >>>> >> >> >>> >>> >> * fise:hasMention {textAnnotation} >> >>>> >> >> >>> >>> >> * dc:type set to fise:Activity >> >>>> >> >> >>> >>> >> >> >>>> >> >> >>> >>> >> If it turns out that we can extract more, we can add >> >>>> more >> >>>> >> >> >>> structure to >> >>>> >> >> >>> >>> >> those annotations. We might also think about using >> an >> >>>> own >> >>>> >> >> namespace >> >>>> >> >> >>> >>> >> for those extensions to the annotation structure. >> >>>> >> >> >>> >>> >> >> >>>> >> >> >>> >>> >> > 2. Determine how should all of this be integrated >> into >> >>>> >> >> Stanbol. >> >>>> >> >> >>> >>> >> >> >>>> >> >> >>> >>> >> Just create an EventExtractionEngine and configure a >> >>>> >> enhancement >> >>>> >> >> >>> chain >> >>>> >> >> >>> >>> >> that does NLP processing and EntityLinking. >> >>>> >> >> >>> >>> >> >> >>>> >> >> >>> >>> >> You should have a look at >> >>>> >> >> >>> >>> >> >> >>>> >> >> >>> >>> >> * SentimentSummarizationEngine [1] as it does a lot >> of >> >>>> things >> >>>> >> >> with >> >>>> >> >> >>> NLP >> >>>> >> >> >>> >>> >> processing results (e.g. connecting adjectives (via >> >>>> verbs) to >> >>>> >> >> >>> >>> >> nouns/pronouns. So as long we can not use explicit >> >>>> dependency >> >>>> >> >> trees >> >>>> >> >> >>> >>> >> you code will need to do similar things with Nouns, >> >>>> Pronouns >> >>>> >> and >> >>>> >> >> >>> >>> >> Verbs. >> >>>> >> >> >>> >>> >> >> >>>> >> >> >>> >>> >> * Disambigutation-MLT engine, as it creates a Java >> >>>> >> >> representation >> >>>> >> >> >>> of >> >>>> >> >> >>> >>> >> present fise:TextAnnotation and >> fise:EntityAnnotation >> >>>> [2]. >> >>>> >> >> >>> Something >> >>>> >> >> >>> >>> >> similar will also be required by the >> >>>> EventExtractionEngine >> >>>> >> for >> >>>> >> >> fast >> >>>> >> >> >>> >>> >> access to such annotations while iterating over the >> >>>> >> Sentences of >> >>>> >> >> >>> the >> >>>> >> >> >>> >>> >> text. >> >>>> >> >> >>> >>> >> >> >>>> >> >> >>> >>> >> >> >>>> >> >> >>> >>> >> best >> >>>> >> >> >>> >>> >> Rupert >> >>>> >> >> >>> >>> >> >> >>>> >> >> >>> >>> >> [1] >> >>>> >> >> >>> >>> >> >> >>>> >> >> >>> >>> >> >>>> >> >> >>> >> >>>> >> >> >> >>>> >> >> >>>> >> https://svn.apache.org/repos/asf/stanbol/trunk/enhancement-engines/sentiment-summarization/src/main/java/org/apache/stanbol/enhancer/engines/sentiment/summarize/SentimentSummarizationEngine.java >> >>>> >> >> >>> >>> >> [2] >> >>>> >> >> >>> >>> >> >> >>>> >> >> >>> >>> >> >>>> >> >> >>> >> >>>> >> >> >> >>>> >> >> >>>> >> https://svn.apache.org/repos/asf/stanbol/trunk/enhancement-engines/disambiguation-mlt/src/main/java/org/apache/stanbol/enhancer/engine/disambiguation/mlt/DisambiguationData.java >> >>>> >> >> >>> >>> >> >> >>>> >> >> >>> >>> >> > >> >>>> >> >> >>> >>> >> > Thanks >> >>>> >> >> >>> >>> >> > >> >>>> >> >> >>> >>> >> > Hope this helps to bootstrap this discussion >> >>>> >> >> >>> >>> >> >> best >> >>>> >> >> >>> >>> >> >> Rupert >> >>>> >> >> >>> >>> >> >> >> >>>> >> >> >>> >>> >> >> -- >> >>>> >> >> >>> >>> >> >> | Rupert Westenthaler >> >>>> >> >> rupert.westentha...@gmail.com >> >>>> >> >> >>> >>> >> >> | Bodenlehenstraße 11 >> >>>> >> >> >>> ++43-699-11108907 >> >>>> >> >> >>> >>> >> >> | A-5500 Bischofshofen >> >>>> >> >> >>> >>> >> >> >> >>>> >> >> >>> >>> >> >> >>>> >> >> >>> >>> >> >> >>>> >> >> >>> >>> >> >> >>>> >> >> >>> >>> >> -- >> >>>> >> >> >>> >>> >> | Rupert Westenthaler >> >>>> >> rupert.westentha...@gmail.com >> >>>> >> >> >>> >>> >> | Bodenlehenstraße 11 >> >>>> >> >> >>> ++43-699-11108907 >> >>>> >> >> >>> >>> >> | A-5500 Bischofshofen >> >>>> >> >> >>> >>> >> >> >>>> >> >> >>> >>> >> >>>> >> >> >>> >>> >> >>>> >> >> >>> >>> >> >>>> >> >> >>> >>> -- >> >>>> >> >> >>> >>> | Rupert Westenthaler >> >>>> rupert.westentha...@gmail.com >> >>>> >> >> >>> >>> | Bodenlehenstraße 11 >> >>>> >> >> ++43-699-11108907 >> >>>> >> >> >>> >>> | A-5500 Bischofshofen >> >>>> >> >> >>> >>> >> >>>> >> >> >>> >> >> >>>> >> >> >>> >> >> >>>> >> >> >>> >> >>>> >> >> >>> >> >>>> >> >> >>> >> >>>> >> >> >>> -- >> >>>> >> >> >>> | Rupert Westenthaler >> >>>> rupert.westentha...@gmail.com >> >>>> >> >> >>> | Bodenlehenstraße 11 >> >>>> ++43-699-11108907 >> >>>> >> >> >>> | A-5500 Bischofshofen >> >>>> >> >> >>> >> >>>> >> >> >> >> >>>> >> >> >> >> >>>> >> >> >> >>>> >> >> >> >>>> >> >> >> >>>> >> >> -- >> >>>> >> >> | Rupert Westenthaler >> rupert.westentha...@gmail.com >> >>>> >> >> | Bodenlehenstraße 11 >> >>>> ++43-699-11108907 >> >>>> >> >> | A-5500 Bischofshofen >> >>>> >> >> >> >>>> >> >> >>>> >> >> >>>> >> >> >>>> >> -- >> >>>> >> | Rupert Westenthaler rupert.westentha...@gmail.com >> >>>> >> | Bodenlehenstraße 11 >> ++43-699-11108907 >> >>>> >> | A-5500 Bischofshofen >> >>>> >> >> >>>> >> >>>> >> >>>> >> >>>> -- >> >>>> | Rupert Westenthaler rupert.westentha...@gmail.com >> >>>> | Bodenlehenstraße 11 ++43-699-11108907 >> >>>> | A-5500 Bischofshofen >> >>>> >> >>> >> >>> >> >> >> >> >> >> -- >> | Rupert Westenthaler rupert.westentha...@gmail.com >> | Bodenlehenstraße 11 ++43-699-11108907 >> | A-5500 Bischofshofen >> > >