On Tue, Jan 29, 2008 at 07:20:12PM +0200, Avi Kivity wrote: > Here's a rough sketch of my proposal: > > - For every memory slot, allocate an array containing one int for every > potential large page included within that memory slot. Each entry in > the array contains the number of write-protected 4KB pages within the > large page frame corresponding to that entry. > > For example, if we have a memory slot for gpas 1MB-1GB, we'd have an > array of size 511, corresponding to the 511 2MB pages from 2MB upwards. > If we shadow a pagetable at address 4MB+8KB, we'd increment the entry > corresponding to the large page at 4MB. When we unshadow that page, > decrement the entry.
You need to take care the the 2MB gpa is aligned 2 MB host physical to be able to map it correctly with a large pte. So maybe we need two memslots for 1MB-1GB. One for 1MB-2MB using normal 4kb pages and one from 2MB-1GB which can be allocated using HugeTLBfs. > - If we attempt to shadow a large page (either a guest pse pte, or a > real-mode pseudo pte), we check if the host page is a large page. If > so, we also check the write-protect count array. If the result is zero, > we create a shadow pse pte. > > - Whenever we write-protect a page, also zap any large-page mappings for > that page. This means rmap will need some extension to handle pde rmaps > in addition to pte rmaps. This sounds straight forward to me. All you need is a short value for every potential large page and initialize it with -1 if the host page is a large page and with 0 otherwise. Every time this value reaches -1 we can map the page with a large pte (and the guest maps with large pte). > - qemu is extended to have a command-line option to use large pages to > back guest memory. > > Large pages should improve performance significantly, both with > traditional shadow and npt/ept. Yes, I think that too. But with shadow paging it really depends on the guest if the performance increasement is long-term. In a Linux guest, for example, the direct mapped memory will become fragmented over time (together with the location of the page tables). So the number of potential large page mappings will likely decrease over time. But the situation is different when the Linux guest uses HugeTLBfs in its userspace. We will always be able to map these pages using large ptes if the guest physical memory is correctly aligned. With Nested Paging (and EPT also) we will always have the benefit because we don't need to write protect anything. I really look forward to large page support in KVM. Maybe we reach the 95% VCPU performance mark compared to native performance with it :-) > Hopefully we will have transparent Linux support for them one day. Unlikely. As far as I know Linus doesn't like the idea... Joerg -- | AMD Saxony Limited Liability Company & Co. KG Operating | Wilschdorfer Landstr. 101, 01109 Dresden, Germany System | Register Court Dresden: HRA 4896 Research | General Partner authorized to represent: Center | AMD Saxony LLC (Wilmington, Delaware, US) | General Manager of AMD Saxony LLC: Dr. Hans-R. Deppe, Thomas McCoy ------------------------------------------------------------------------- This SF.net email is sponsored by: Microsoft Defy all challenges. Microsoft(R) Visual Studio 2008. http://clk.atdmt.com/MRT/go/vse0120000070mrt/direct/01/ _______________________________________________ kvm-devel mailing list kvm-devel@lists.sourceforge.net https://lists.sourceforge.net/lists/listinfo/kvm-devel