On 04/26/2017 01:25 AM, lixi...@cmss.chinamobile.com wrote:
> for_each_sg(data_sg, sg, data_nents, i) {
> @@ -275,22 +371,26 @@ static void alloc_and_scatter_data_area(struct tcmu_dev
> *udev,
> from = kmap_atomic(sg_page(sg)) + sg->offset;
> while (sg_remaining
On 04/26/2017 01:25 AM, lixi...@cmss.chinamobile.com wrote:
> for_each_sg(data_sg, sg, data_nents, i) {
> @@ -275,22 +371,26 @@ static void alloc_and_scatter_data_area(struct tcmu_dev
> *udev,
> from = kmap_atomic(sg_page(sg)) + sg->offset;
> while (sg_remaining
From: Xiubo Li
Currently for the TCMU, the ring buffer size is fixed to 64K cmd
area + 1M data area, and this will be bottlenecks for high iops.
The struct tcmu_cmd_entry {} size is fixed about 112 bytes with
iovec[N] & N <= 4, and the size of struct iovec is about
From: Xiubo Li
Currently for the TCMU, the ring buffer size is fixed to 64K cmd
area + 1M data area, and this will be bottlenecks for high iops.
The struct tcmu_cmd_entry {} size is fixed about 112 bytes with
iovec[N] & N <= 4, and the size of struct iovec is about 16 bytes.
If N == 0, the
4 matches
Mail list logo