Hi Rupert! Nice to see great feedback on this topic!. I wanted to comment part of your previous email:
En 24 de octubre de 2014 en 12:31:39, Rupert Westenthaler (rupert.westentha...@gmail.com) escrito: I fear that each disambiguation approach will come with its own data model. Mainly because the way data is kept is central for performance. Also based on what I have seen up to now keeping everything in-memory is the way to go. For Aida-Light I suggest to keep the current solutions. The requirement of about 50GByte RAM for Yago is anyways quite OK. If we can add support for more focused datasets one will often end up with far less entities. This is also a way to keep resource requirements down. I completely understand your point but I partially disagree. I find the memory consumption requirement quite tough. It may prevent a lot of people to give it try or experiment with it. Actually, Chalitha had serious problems to find a machine for testing it. I agree that probably it is difficult to find a valid architecture for any disambiguation approach, but focusing only in Aida-Light I think that it is worth to provide also an other data management solution. It will affect the performance for sure, but at least it will let people use it without a super machine. I’m actually thinking in alternatives that can “simulate” memory access using the disk like LevelDB or similar. Makes sense? Cheers, Rafa