On 16-11-30 06:30 AM, Jakub Kicinski wrote:
> [add MST]
> 

Thanks sorry MST. I did a cut'n'paste of an old list of CC's and missed
you were not on the list.

[...]

>> +    memcpy(page_address(page) + page_off, page_address(p) + offset, *len);
>> +    while (--num_buf) {
>> +            unsigned int buflen;
>> +            unsigned long ctx;
>> +            void *buf;
>> +            int off;
>> +
>> +            ctx = (unsigned long)virtqueue_get_buf(rq->vq, &buflen);
>> +            if (unlikely(!ctx))
>> +                    goto err_buf;
>> +
>> +            buf = mergeable_ctx_to_buf_address(ctx);
>> +            p = virt_to_head_page(buf);
>> +            off = buf - page_address(p);
>> +
>> +            memcpy(page_address(page) + page_off,
>> +                   page_address(p) + off, buflen);
>> +            page_off += buflen;
> 
> Could malicious user potentially submit a frame bigger than MTU?

Well presumably if the MTU is greater than PAGE_SIZE the xdp program
would not have been loaded. And the malicious user in this case would
have to be qemu which seems like everything is already lost if qemu
is trying to attack its VM.

But this is a good point because it looks like there is nothing in
virtio or qemu that drops frames with MTU greater than the virtio
configured setting. Maybe Michael can confirm this or I'll poke at it
more. I think qemu should drop these frames in general.

So I think adding a guard here is sensible I'll go ahead and do that.
Also the MTU guard at set_xdp time needs to account for header length.

Thanks nice catch.

> 
>> +    }
>> +
>> +    *len = page_off;
>> +    return page;
>> +err_buf:
>> +    __free_pages(page, 0);
>> +    return NULL;
>> +}
>> +
>>  static struct sk_buff *receive_mergeable(struct net_device *dev,
>>                                       struct virtnet_info *vi,
>>                                       struct receive_queue *rq,
>> @@ -469,21 +519,37 @@ static struct sk_buff *receive_mergeable(struct 
>> net_device *dev,
>>      rcu_read_lock();
>>      xdp_prog = rcu_dereference(rq->xdp_prog);
>>      if (xdp_prog) {
>> +            struct page *xdp_page;
>>              u32 act;
>>  
>>              if (num_buf > 1) {
>>                      bpf_warn_invalid_xdp_buffer();
>> -                    goto err_xdp;
>> +
>> +                    /* linearize data for XDP */
>> +                    xdp_page = xdp_linearize_page(rq, num_buf,
>> +                                                  page, offset, &len);
>> +                    if (!xdp_page)
>> +                            goto err_xdp;
>> +                    offset = len;
>> +            } else {
>> +                    xdp_page = page;
>>              }
>>  
>> -            act = do_xdp_prog(vi, xdp_prog, page, offset, len);
>> +            act = do_xdp_prog(vi, xdp_prog, xdp_page, offset, len);
>>              switch (act) {
>>              case XDP_PASS:
>> +                    if (unlikely(xdp_page != page))
>> +                            __free_pages(xdp_page, 0);
>>                      break;
>>              case XDP_TX:
>> +                    if (unlikely(xdp_page != page))
>> +                            goto err_xdp;
>> +                    rcu_read_unlock();
> 
> Only if there is a reason for v4 - this unlock could go to the previous
> patch.
> 

Sure will do this.

Reply via email to