Thank you for your reply. I'll try it.

在 2019/7/12 下午6:14, Ilya Kasnacheev 写道:
Hello!

The recommendation here is to disable WAL before ingesting data in a table. You can do that by issuing ALTER TABLE tbl NOLOGGING;

After data is loaded, you should turn it back on by ALTER TABLE tbl LOGGING.

Regards,
--
Ilya Kasnacheev


пт, 12 июл. 2019 г. в 11:33, liyuj <[email protected] <mailto:[email protected]>>:

    Hi,

    The CSV file is about 250 GB, with about 1 billion rows of data.
    Persistence is on and there is enough memory.
    It has been successfully imported, but it takes a long time.

    The problem at present is that the data of this large table is
    imported
    successfully, and then 50 million tables are imported. The speed
    of data
    writing is significantly slowed down.

    We have four hosts in total. The cache configuration is as follows:
    <property name="backups" value="1"/>
    <property name="partitionLossPolicy" value="READ_ONLY_SAFE"/>

    With persistence enabled, the other parameters are nothing special.

    在 2019/7/12 下午1:47, Павлухин Иван 写道:
    > Hi,
    >
    > Currently COPY is a mechanism designed for the fastest data
    load. Yes,
    > you can try separate your data in chunks and execute COPY in
    parallel.
    > By the way, where is your input located and what is it size in bytes
    > (Gb)? Is persistence enabled? Does a DataRegion have enough
    memory to
    > keep all data?
    >
    > ср, 10 июл. 2019 г. в 05:02, 18624049226 <[email protected]
    <mailto:[email protected]>>:
    >> If the COPY command is used to import a large amount of data,
    the execution time is a little long.
    >> In the current test environment, the performance is about more
    than 10,000/s, so if it is 100 million data, it will take several
    hours.
    >>
    >> Is there a faster way to import, or is COPY working in parallel?
    >>
    >> thanks!
    >>
    >

Reply via email to