Re: [RFC PATCH 3/3] x86: Create dma_mark_dirty to dirty pages used for DMA by VM guest
On Mon, Dec 14, 2015 at 12:52 PM, Michael S. Tsirkin wrote: > On Mon, Dec 14, 2015 at 09:59:13AM -0800, Alexander Duyck wrote: >> On Mon, Dec 14, 2015 at 9:20 AM, Michael S. Tsirkin wrote: >> > On Mon, Dec 14, 2015 at 08:34:00AM -0800, Alexander Duyck wrote: >> >> > This way distro can use a guest agent to disable >> >> > dirtying until before migration starts. >> >> >> >> Right. For a v2 version I would definitely want to have some way to >> >> limit the scope of this. My main reason for putting this out here is >> >> to start altering the course of discussions since it seems like were >> >> weren't getting anywhere with the ixgbevf migration changes that were >> >> being proposed. >> > >> > Absolutely, thanks for working on this. >> > >> >> >> + unsigned long pg_addr, start; >> >> >> + >> >> >> + start = (unsigned long)addr; >> >> >> + pg_addr = PAGE_ALIGN(start + size); >> >> >> + start &= ~(sizeof(atomic_t) - 1); >> >> >> + >> >> >> + /* trigger a write fault on each page, excluding first page */ >> >> >> + while ((pg_addr -= PAGE_SIZE) > start) >> >> >> + atomic_add(0, (atomic_t *)pg_addr); >> >> >> + >> >> >> + /* trigger a write fault on first word of DMA */ >> >> >> + atomic_add(0, (atomic_t *)start); > > Actually, I have second thoughts about using atomic_add here, > especially for _sync. > > Many architectures do > > #define ATOMIC_OP_RETURN(op, c_op) \ > static inline int atomic_##op##_return(int i, atomic_t *v) \ > { \ > unsigned long flags;\ > int ret;\ > \ > raw_local_irq_save(flags); \ > ret = (v->counter = v->counter c_op i); \ > raw_local_irq_restore(flags); \ > \ > return ret; \ > } > > and this is not safe if device is still doing DMA to/from > this memory. > > Generally, atomic_t is there for SMP effects, not for sync > with devices. > > This is why I said you should do > cmpxchg(pg_addr, 0xdead, 0xdead); > > Yes, we probably never actually want to run m68k within a VM, > but let's not misuse interfaces like this. Right now this implementation is for x86 only. Any other architecture currently reports dma_mark_dirty as an empty inline function. The reason why I chose the atomic_add for x86 is simply because it is guaranteed dirty the cache line with relatively few instructions and operands as all I have to have is the pointer and 0. For the m68k we could implement it as a cmpxchg instead. The general thought here is that each architecture is probably going to have to do it a little bit differently. >> >> > >> >> > start might not be aligned correctly for a cast to atomic_t. >> >> > It's harmless to do this for any memory, so I think you should >> >> > just do this for 1st byte of all pages including the first one. >> >> >> >> You may not have noticed it but I actually aligned start in the line >> >> after pg_addr. >> > >> > Yes you did. alignof would make it a bit more noticeable. >> > >> >> However instead of aligning to the start of the next >> >> atomic_t I just masked off the lower bits so that we start at the >> >> DWORD that contains the first byte of the starting address. The >> >> assumption here is that I cannot trigger any sort of fault since if I >> >> have access to a given byte within a DWORD I will have access to the >> >> entire DWORD. >> > >> > I'm curious where does this come from. Isn't it true that access is >> > controlled at page granularity normally, so you can touch beginning of >> > page just as well? >> >> Yeah, I am pretty sure it probably is page granularity. However my >> thought was to try and stick to the start of the DMA as the last >> access. That way we don't pull in any more cache lines than we need >> to in order to dirty the pages. Usually the start of the DMA region >> will contain some sort of headers or something that needs to be >> accessed with the highest priority so I wanted to make certain that we >> were forcing usable data into the L1 cache rather than just the first >> cache line of the page where the DMA started. If however the start of >> a DMA was the start of the page there is nothing there to prevent >> that. > > OK, maybe this helps. You should document all these tricks > in code comments. I'll try to get that taken care of for v2. >> >> I coded this up so that the spots where we touch the >> >> memory should match up with addresses provided by the hardware to >> >> perform the DMA over the PCI bus. >> > >> > Yes but
Re: [RFC PATCH 3/3] x86: Create dma_mark_dirty to dirty pages used for DMA by VM guest
On Mon, Dec 14, 2015 at 09:59:13AM -0800, Alexander Duyck wrote: > On Mon, Dec 14, 2015 at 9:20 AM, Michael S. Tsirkin wrote: > > On Mon, Dec 14, 2015 at 08:34:00AM -0800, Alexander Duyck wrote: > >> > This way distro can use a guest agent to disable > >> > dirtying until before migration starts. > >> > >> Right. For a v2 version I would definitely want to have some way to > >> limit the scope of this. My main reason for putting this out here is > >> to start altering the course of discussions since it seems like were > >> weren't getting anywhere with the ixgbevf migration changes that were > >> being proposed. > > > > Absolutely, thanks for working on this. > > > >> >> + unsigned long pg_addr, start; > >> >> + > >> >> + start = (unsigned long)addr; > >> >> + pg_addr = PAGE_ALIGN(start + size); > >> >> + start &= ~(sizeof(atomic_t) - 1); > >> >> + > >> >> + /* trigger a write fault on each page, excluding first page */ > >> >> + while ((pg_addr -= PAGE_SIZE) > start) > >> >> + atomic_add(0, (atomic_t *)pg_addr); > >> >> + > >> >> + /* trigger a write fault on first word of DMA */ > >> >> + atomic_add(0, (atomic_t *)start); Actually, I have second thoughts about using atomic_add here, especially for _sync. Many architectures do #define ATOMIC_OP_RETURN(op, c_op) \ static inline int atomic_##op##_return(int i, atomic_t *v) \ { \ unsigned long flags;\ int ret;\ \ raw_local_irq_save(flags); \ ret = (v->counter = v->counter c_op i); \ raw_local_irq_restore(flags); \ \ return ret; \ } and this is not safe if device is still doing DMA to/from this memory. Generally, atomic_t is there for SMP effects, not for sync with devices. This is why I said you should do cmpxchg(pg_addr, 0xdead, 0xdead); Yes, we probably never actually want to run m68k within a VM, but let's not misuse interfaces like this. > >> > > >> > start might not be aligned correctly for a cast to atomic_t. > >> > It's harmless to do this for any memory, so I think you should > >> > just do this for 1st byte of all pages including the first one. > >> > >> You may not have noticed it but I actually aligned start in the line > >> after pg_addr. > > > > Yes you did. alignof would make it a bit more noticeable. > > > >> However instead of aligning to the start of the next > >> atomic_t I just masked off the lower bits so that we start at the > >> DWORD that contains the first byte of the starting address. The > >> assumption here is that I cannot trigger any sort of fault since if I > >> have access to a given byte within a DWORD I will have access to the > >> entire DWORD. > > > > I'm curious where does this come from. Isn't it true that access is > > controlled at page granularity normally, so you can touch beginning of > > page just as well? > > Yeah, I am pretty sure it probably is page granularity. However my > thought was to try and stick to the start of the DMA as the last > access. That way we don't pull in any more cache lines than we need > to in order to dirty the pages. Usually the start of the DMA region > will contain some sort of headers or something that needs to be > accessed with the highest priority so I wanted to make certain that we > were forcing usable data into the L1 cache rather than just the first > cache line of the page where the DMA started. If however the start of > a DMA was the start of the page there is nothing there to prevent > that. OK, maybe this helps. You should document all these tricks in code comments. > >> I coded this up so that the spots where we touch the > >> memory should match up with addresses provided by the hardware to > >> perform the DMA over the PCI bus. > > > > Yes but there's no requirement to do it like this from > > virt POV. You just need to touch each page. > > I know, but at the same time if we match up with the DMA then it is > more likely that we avoid grabbing unneeded cache lines. In the case > of most drivers the data for headers and start is at the start of the > DMA. So if we dirty the cache line associated with the start of the > DMA it will be pulled into the L1 cache and there is a greater chance > that it may already be prefetched as well. > > >> Also I intentionally ran from highest address to lowest since that way > >> we don't risk pushing the first cache line of the DMA buffer out of > >> the L1 cache due to the PAGE_SIZE stride. > > > > Interesting.
Re: [RFC PATCH 3/3] x86: Create dma_mark_dirty to dirty pages used for DMA by VM guest
On Mon, Dec 14, 2015 at 9:20 AM, Michael S. Tsirkin wrote: > On Mon, Dec 14, 2015 at 08:34:00AM -0800, Alexander Duyck wrote: >> > This way distro can use a guest agent to disable >> > dirtying until before migration starts. >> >> Right. For a v2 version I would definitely want to have some way to >> limit the scope of this. My main reason for putting this out here is >> to start altering the course of discussions since it seems like were >> weren't getting anywhere with the ixgbevf migration changes that were >> being proposed. > > Absolutely, thanks for working on this. > >> >> + unsigned long pg_addr, start; >> >> + >> >> + start = (unsigned long)addr; >> >> + pg_addr = PAGE_ALIGN(start + size); >> >> + start &= ~(sizeof(atomic_t) - 1); >> >> + >> >> + /* trigger a write fault on each page, excluding first page */ >> >> + while ((pg_addr -= PAGE_SIZE) > start) >> >> + atomic_add(0, (atomic_t *)pg_addr); >> >> + >> >> + /* trigger a write fault on first word of DMA */ >> >> + atomic_add(0, (atomic_t *)start); >> > >> > start might not be aligned correctly for a cast to atomic_t. >> > It's harmless to do this for any memory, so I think you should >> > just do this for 1st byte of all pages including the first one. >> >> You may not have noticed it but I actually aligned start in the line >> after pg_addr. > > Yes you did. alignof would make it a bit more noticeable. > >> However instead of aligning to the start of the next >> atomic_t I just masked off the lower bits so that we start at the >> DWORD that contains the first byte of the starting address. The >> assumption here is that I cannot trigger any sort of fault since if I >> have access to a given byte within a DWORD I will have access to the >> entire DWORD. > > I'm curious where does this come from. Isn't it true that access is > controlled at page granularity normally, so you can touch beginning of > page just as well? Yeah, I am pretty sure it probably is page granularity. However my thought was to try and stick to the start of the DMA as the last access. That way we don't pull in any more cache lines than we need to in order to dirty the pages. Usually the start of the DMA region will contain some sort of headers or something that needs to be accessed with the highest priority so I wanted to make certain that we were forcing usable data into the L1 cache rather than just the first cache line of the page where the DMA started. If however the start of a DMA was the start of the page there is nothing there to prevent that. >> I coded this up so that the spots where we touch the >> memory should match up with addresses provided by the hardware to >> perform the DMA over the PCI bus. > > Yes but there's no requirement to do it like this from > virt POV. You just need to touch each page. I know, but at the same time if we match up with the DMA then it is more likely that we avoid grabbing unneeded cache lines. In the case of most drivers the data for headers and start is at the start of the DMA. So if we dirty the cache line associated with the start of the DMA it will be pulled into the L1 cache and there is a greater chance that it may already be prefetched as well. >> Also I intentionally ran from highest address to lowest since that way >> we don't risk pushing the first cache line of the DMA buffer out of >> the L1 cache due to the PAGE_SIZE stride. > > Interesting. How does order of access help with this? If you use a PAGE_SIZE stride you will start evicting things from L1 cache after something like 8 accesses on an x86 processor as most of the recent ones have a 32K 8 way associative L1 cache. So if I go from back to front then I evict the stuff that would likely be in the data portion of a buffer instead of headers which are usually located at the front. > By the way, if you are into these micro-optimizations you might want to > limit prefetch, to this end you want to access the last line of the > page. And it's probably worth benchmarking a bit and not doing it all just > based on theory, keep code simple in v1 otherwise. My main goal for now is functional code over high performance code. That is why I have kept this code fairly simple. I might have done some optimization but it was as much about the optimization as keeping the code simple. For example by using the start of the page instead of the end I could easily do the comparison against start and avoid doing more than one write per page. The issue for me doing performance testing is that I don't have anything that uses DMA blocks that are actually big enough to make use of the PAGE_SIZE stride. That is why the PAGE_SIZE stride portion is mostly just theoretical. I just have a few NICs and most of them only allocate 1 page or so for DMA buffers. What little benchmarking I have done with netperf only showed a ~1% CPU penalty for the page dirtying code. For setups where we did more with the DMA such as small packet ha
Re: [RFC PATCH 3/3] x86: Create dma_mark_dirty to dirty pages used for DMA by VM guest
On Mon, Dec 14, 2015 at 08:34:00AM -0800, Alexander Duyck wrote: > > This way distro can use a guest agent to disable > > dirtying until before migration starts. > > Right. For a v2 version I would definitely want to have some way to > limit the scope of this. My main reason for putting this out here is > to start altering the course of discussions since it seems like were > weren't getting anywhere with the ixgbevf migration changes that were > being proposed. Absolutely, thanks for working on this. > >> + unsigned long pg_addr, start; > >> + > >> + start = (unsigned long)addr; > >> + pg_addr = PAGE_ALIGN(start + size); > >> + start &= ~(sizeof(atomic_t) - 1); > >> + > >> + /* trigger a write fault on each page, excluding first page */ > >> + while ((pg_addr -= PAGE_SIZE) > start) > >> + atomic_add(0, (atomic_t *)pg_addr); > >> + > >> + /* trigger a write fault on first word of DMA */ > >> + atomic_add(0, (atomic_t *)start); > > > > start might not be aligned correctly for a cast to atomic_t. > > It's harmless to do this for any memory, so I think you should > > just do this for 1st byte of all pages including the first one. > > You may not have noticed it but I actually aligned start in the line > after pg_addr. Yes you did. alignof would make it a bit more noticeable. > However instead of aligning to the start of the next > atomic_t I just masked off the lower bits so that we start at the > DWORD that contains the first byte of the starting address. The > assumption here is that I cannot trigger any sort of fault since if I > have access to a given byte within a DWORD I will have access to the > entire DWORD. I'm curious where does this come from. Isn't it true that access is controlled at page granularity normally, so you can touch beginning of page just as well? > I coded this up so that the spots where we touch the > memory should match up with addresses provided by the hardware to > perform the DMA over the PCI bus. Yes but there's no requirement to do it like this from virt POV. You just need to touch each page. > Also I intentionally ran from highest address to lowest since that way > we don't risk pushing the first cache line of the DMA buffer out of > the L1 cache due to the PAGE_SIZE stride. > > - Alex Interesting. How does order of access help with this? By the way, if you are into these micro-optimizations you might want to limit prefetch, to this end you want to access the last line of the page. And it's probably worth benchmarking a bit and not doing it all just based on theory, keep code simple in v1 otherwise. -- MST -- To unsubscribe from this list: send the line "unsubscribe kvm" in the body of a message to majord...@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html
Re: [RFC PATCH 3/3] x86: Create dma_mark_dirty to dirty pages used for DMA by VM guest
On Mon, Dec 14, 2015 at 6:00 AM, Michael S. Tsirkin wrote: > On Sun, Dec 13, 2015 at 01:28:31PM -0800, Alexander Duyck wrote: >> This patch is meant to provide the guest with a way of flagging DMA pages >> as being dirty to the host when using a direct-assign device within a >> guest. The advantage to this approach is that it is fairly simple, however >> It currently has a singificant impact on device performance in all the >> scenerios where it won't be needed. >> >> As such this is really meant only as a proof of concept and to get the ball >> rolling in terms of figuring out how best to approach the issue of dirty >> page tracking for a guest that is using a direct assigned device. In >> addition with just this patch it should be possible to modify current >> migration approaches so that instead of having to hot-remove the device >> before starting the migration this can instead be delayed until the period >> before the final stop and copy. >> >> Signed-off-by: Alexander Duyck >> --- >> arch/arm/include/asm/dma-mapping.h |3 ++- >> arch/arm64/include/asm/dma-mapping.h |5 ++--- >> arch/ia64/include/asm/dma.h |1 + >> arch/mips/include/asm/dma-mapping.h |1 + >> arch/powerpc/include/asm/swiotlb.h |1 + >> arch/tile/include/asm/dma-mapping.h |1 + >> arch/unicore32/include/asm/dma-mapping.h |1 + >> arch/x86/Kconfig | 11 +++ >> arch/x86/include/asm/swiotlb.h | 26 ++ >> drivers/xen/swiotlb-xen.c|6 ++ >> lib/swiotlb.c|6 ++ >> 11 files changed, 58 insertions(+), 4 deletions(-) >> >> diff --git a/arch/arm/include/asm/dma-mapping.h >> b/arch/arm/include/asm/dma-mapping.h >> index ccb3aa64640d..1962d7b471c7 100644 >> --- a/arch/arm/include/asm/dma-mapping.h >> +++ b/arch/arm/include/asm/dma-mapping.h >> @@ -167,7 +167,8 @@ static inline bool dma_capable(struct device *dev, >> dma_addr_t addr, size_t size) >> return 1; >> } >> >> -static inline void dma_mark_clean(void *addr, size_t size) { } >> +static inline void dma_mark_clean(void *addr, size_t size) {} >> +static inline void dma_mark_dirty(void *addr, size_t size) {} >> >> extern int arm_dma_set_mask(struct device *dev, u64 dma_mask); >> >> diff --git a/arch/arm64/include/asm/dma-mapping.h >> b/arch/arm64/include/asm/dma-mapping.h >> index 61e08f360e31..8d24fe11c8a3 100644 >> --- a/arch/arm64/include/asm/dma-mapping.h >> +++ b/arch/arm64/include/asm/dma-mapping.h >> @@ -84,9 +84,8 @@ static inline bool dma_capable(struct device *dev, >> dma_addr_t addr, size_t size) >> return addr + size - 1 <= *dev->dma_mask; >> } >> >> -static inline void dma_mark_clean(void *addr, size_t size) >> -{ >> -} >> +static inline void dma_mark_clean(void *addr, size_t size) {} >> +static inline void dma_mark_dirty(void *addr, size_t size) {} >> >> #endif /* __KERNEL__ */ >> #endif /* __ASM_DMA_MAPPING_H */ >> diff --git a/arch/ia64/include/asm/dma.h b/arch/ia64/include/asm/dma.h >> index 4d97f60f1ef5..d92ebeb2758e 100644 >> --- a/arch/ia64/include/asm/dma.h >> +++ b/arch/ia64/include/asm/dma.h >> @@ -20,5 +20,6 @@ extern unsigned long MAX_DMA_ADDRESS; >> #define free_dma(x) >> >> void dma_mark_clean(void *addr, size_t size); >> +static inline void dma_mark_dirty(void *addr, size_t size) {} >> >> #endif /* _ASM_IA64_DMA_H */ >> diff --git a/arch/mips/include/asm/dma-mapping.h >> b/arch/mips/include/asm/dma-mapping.h >> index e604f760c4a0..567f6e03e337 100644 >> --- a/arch/mips/include/asm/dma-mapping.h >> +++ b/arch/mips/include/asm/dma-mapping.h >> @@ -28,6 +28,7 @@ static inline bool dma_capable(struct device *dev, >> dma_addr_t addr, size_t size) >> } >> >> static inline void dma_mark_clean(void *addr, size_t size) {} >> +static inline void dma_mark_dirty(void *addr, size_t size) {} >> >> #include >> >> diff --git a/arch/powerpc/include/asm/swiotlb.h >> b/arch/powerpc/include/asm/swiotlb.h >> index de99d6e29430..b694e8399e28 100644 >> --- a/arch/powerpc/include/asm/swiotlb.h >> +++ b/arch/powerpc/include/asm/swiotlb.h >> @@ -16,6 +16,7 @@ >> extern struct dma_map_ops swiotlb_dma_ops; >> >> static inline void dma_mark_clean(void *addr, size_t size) {} >> +static inline void dma_mark_dirty(void *addr, size_t size) {} >> >> extern unsigned int ppc_swiotlb_enable; >> int __init swiotlb_setup_bus_notifier(void); >> diff --git a/arch/tile/include/asm/dma-mapping.h >> b/arch/tile/include/asm/dma-mapping.h >> index 96ac6cce4a32..79953f09e938 100644 >> --- a/arch/tile/include/asm/dma-mapping.h >> +++ b/arch/tile/include/asm/dma-mapping.h >> @@ -58,6 +58,7 @@ static inline phys_addr_t dma_to_phys(struct device *dev, >> dma_addr_t daddr) >> } >> >> static inline void dma_mark_clean(void *addr, size_t size) {} >> +static inline void dma_mark_dirty(void *addr, size_t size) {} >> >> static inline void set_dma_ops(struct device *dev, str
Re: [RFC PATCH 3/3] x86: Create dma_mark_dirty to dirty pages used for DMA by VM guest
On Sun, Dec 13, 2015 at 01:28:31PM -0800, Alexander Duyck wrote: > This patch is meant to provide the guest with a way of flagging DMA pages > as being dirty to the host when using a direct-assign device within a > guest. The advantage to this approach is that it is fairly simple, however > It currently has a singificant impact on device performance in all the > scenerios where it won't be needed. > > As such this is really meant only as a proof of concept and to get the ball > rolling in terms of figuring out how best to approach the issue of dirty > page tracking for a guest that is using a direct assigned device. In > addition with just this patch it should be possible to modify current > migration approaches so that instead of having to hot-remove the device > before starting the migration this can instead be delayed until the period > before the final stop and copy. > > Signed-off-by: Alexander Duyck > --- > arch/arm/include/asm/dma-mapping.h |3 ++- > arch/arm64/include/asm/dma-mapping.h |5 ++--- > arch/ia64/include/asm/dma.h |1 + > arch/mips/include/asm/dma-mapping.h |1 + > arch/powerpc/include/asm/swiotlb.h |1 + > arch/tile/include/asm/dma-mapping.h |1 + > arch/unicore32/include/asm/dma-mapping.h |1 + > arch/x86/Kconfig | 11 +++ > arch/x86/include/asm/swiotlb.h | 26 ++ > drivers/xen/swiotlb-xen.c|6 ++ > lib/swiotlb.c|6 ++ > 11 files changed, 58 insertions(+), 4 deletions(-) > > diff --git a/arch/arm/include/asm/dma-mapping.h > b/arch/arm/include/asm/dma-mapping.h > index ccb3aa64640d..1962d7b471c7 100644 > --- a/arch/arm/include/asm/dma-mapping.h > +++ b/arch/arm/include/asm/dma-mapping.h > @@ -167,7 +167,8 @@ static inline bool dma_capable(struct device *dev, > dma_addr_t addr, size_t size) > return 1; > } > > -static inline void dma_mark_clean(void *addr, size_t size) { } > +static inline void dma_mark_clean(void *addr, size_t size) {} > +static inline void dma_mark_dirty(void *addr, size_t size) {} > > extern int arm_dma_set_mask(struct device *dev, u64 dma_mask); > > diff --git a/arch/arm64/include/asm/dma-mapping.h > b/arch/arm64/include/asm/dma-mapping.h > index 61e08f360e31..8d24fe11c8a3 100644 > --- a/arch/arm64/include/asm/dma-mapping.h > +++ b/arch/arm64/include/asm/dma-mapping.h > @@ -84,9 +84,8 @@ static inline bool dma_capable(struct device *dev, > dma_addr_t addr, size_t size) > return addr + size - 1 <= *dev->dma_mask; > } > > -static inline void dma_mark_clean(void *addr, size_t size) > -{ > -} > +static inline void dma_mark_clean(void *addr, size_t size) {} > +static inline void dma_mark_dirty(void *addr, size_t size) {} > > #endif /* __KERNEL__ */ > #endif /* __ASM_DMA_MAPPING_H */ > diff --git a/arch/ia64/include/asm/dma.h b/arch/ia64/include/asm/dma.h > index 4d97f60f1ef5..d92ebeb2758e 100644 > --- a/arch/ia64/include/asm/dma.h > +++ b/arch/ia64/include/asm/dma.h > @@ -20,5 +20,6 @@ extern unsigned long MAX_DMA_ADDRESS; > #define free_dma(x) > > void dma_mark_clean(void *addr, size_t size); > +static inline void dma_mark_dirty(void *addr, size_t size) {} > > #endif /* _ASM_IA64_DMA_H */ > diff --git a/arch/mips/include/asm/dma-mapping.h > b/arch/mips/include/asm/dma-mapping.h > index e604f760c4a0..567f6e03e337 100644 > --- a/arch/mips/include/asm/dma-mapping.h > +++ b/arch/mips/include/asm/dma-mapping.h > @@ -28,6 +28,7 @@ static inline bool dma_capable(struct device *dev, > dma_addr_t addr, size_t size) > } > > static inline void dma_mark_clean(void *addr, size_t size) {} > +static inline void dma_mark_dirty(void *addr, size_t size) {} > > #include > > diff --git a/arch/powerpc/include/asm/swiotlb.h > b/arch/powerpc/include/asm/swiotlb.h > index de99d6e29430..b694e8399e28 100644 > --- a/arch/powerpc/include/asm/swiotlb.h > +++ b/arch/powerpc/include/asm/swiotlb.h > @@ -16,6 +16,7 @@ > extern struct dma_map_ops swiotlb_dma_ops; > > static inline void dma_mark_clean(void *addr, size_t size) {} > +static inline void dma_mark_dirty(void *addr, size_t size) {} > > extern unsigned int ppc_swiotlb_enable; > int __init swiotlb_setup_bus_notifier(void); > diff --git a/arch/tile/include/asm/dma-mapping.h > b/arch/tile/include/asm/dma-mapping.h > index 96ac6cce4a32..79953f09e938 100644 > --- a/arch/tile/include/asm/dma-mapping.h > +++ b/arch/tile/include/asm/dma-mapping.h > @@ -58,6 +58,7 @@ static inline phys_addr_t dma_to_phys(struct device *dev, > dma_addr_t daddr) > } > > static inline void dma_mark_clean(void *addr, size_t size) {} > +static inline void dma_mark_dirty(void *addr, size_t size) {} > > static inline void set_dma_ops(struct device *dev, struct dma_map_ops *ops) > { > diff --git a/arch/unicore32/include/asm/dma-mapping.h > b/arch/unicore32/include/asm/dma-mapping.h > index 8140e053