Hi Andrey
The tables in the db contain data ranging from simple key value
objects to serialized objects. The following is a snapshot of the types of
tables in the db
*Type 1*. Uniquely indexed key value tables in which I try to persist an
object whose key field is uniquely indexed.
The variables in the object have the following types
a) primitives
b) Strings
c) Serialized fields
*Type 2*. Objects which have both Uniquely Indexed fields and Non
Uniquely Indexed fields
The variables in the object have the following types
a) primitives
b) Strings
c) Serialized fields
*Type 3*. Key Value tables with a unique key and a serialized value.
In all the above types of tables,
The db has around *78 tables* out of which* 19 tables fall under Type 2*. *9
tables fall under Type 3*. and the *rest Type 1*.
Every *read* call happens within *one single transaction*.
The reads involve fetching data from across all these types of
tables(almost all 78) and the data gets fetched using Unique keys and Non
Unique Keys. All this within a single transaction.
When read happens for the very *first time*, fetching data for *50,000
keys* which are a combination of uniquely indexed and non-uniquely indexed
keys within one transaction, the total time taken for fetching all the data
for the 50,000 keys is *361053 ms*.
Time for fetching data the *second time* for the *same* set of* 50,000
keys* is *276512* *ms*
Time for fetching data the *third time* for the *same* set of *50,000 keys*
is *57768 ms*.
Time for fetching data the *fourth time* for the *same* set of *50,000
keys* is *26626 ms*.
After some time say *half an hour* when i *repeat the same exercise,* the
*same
pattern* is repeated.
The basic understanding I get when I observe this pattern is that *speed
of fetch works on an LRU* basis.(Correct me if I am wrong)
*The application I work on, has strict time constraints and something has
to be done to improve the read performance.* So my question is
*Are there any Configurations available to improve the entire fetch
process? Increasing the DISK_CACKE_BUFFER_SIZE will be of any use? Or is
there anything else I can do?*
On Friday, January 9, 2015 at 12:05:53 PM UTC+5:30, Mandark13 wrote:
>
>
> Hi,
>
>
> I am a new Orient DB user. I have created a database which has
> a combination of Document data and Serialized data. I have a total of 78
> tables out of which 17 tables have non unique indices and the remaining
> have unique indices. The total number of records across all tables is
> pretty huge and amounts to around 5 crores. I am finding the fetch response
> slow for the first few times when I try to fetch around 50 k items from
> across all the tables within a single transaction. When I repeatedly try to
> fetch the same 50 K items, I am observing the fetch time is faster after
> say the fourth try. I understand this behaviour is because of the LRU
> implementation. But my concern is, *is there any way I can speed up the
> fetch the first few times? Any configurations can accelerate stuff?*
>
--
---
You received this message because you are subscribed to the Google Groups
"OrientDB" group.
To unsubscribe from this group and stop receiving emails from it, send an email
to [email protected].
For more options, visit https://groups.google.com/d/optout.