On Wed, 30 Nov 2016 16:45:33 -0700
Ross Zwisler <ross.zwis...@linux.intel.com> wrote:

> diff --git a/include/linux/pfn_t.h b/include/linux/pfn_t.h
> index a3d90b9..033fc7b 100644
> --- a/include/linux/pfn_t.h
> +++ b/include/linux/pfn_t.h
> @@ -15,6 +15,12 @@
>  #define PFN_DEV (1ULL << (BITS_PER_LONG_LONG - 3))
>  #define PFN_MAP (1ULL << (BITS_PER_LONG_LONG - 4))
>  
> +#define PFN_FLAGS_TRACE \
> +     { PFN_SG_CHAIN, "SG_CHAIN" }, \
> +     { PFN_SG_LAST,  "SG_LAST" }, \
> +     { PFN_DEV,      "DEV" }, \
> +     { PFN_MAP,      "MAP" }
> +
>  static inline pfn_t __pfn_to_pfn_t(unsigned long pfn, u64 flags)
>  {
>       pfn_t pfn_t = { .val = pfn | (flags & PFN_FLAGS_MASK), };
> diff --git a/include/trace/events/fs_dax.h b/include/trace/events/fs_dax.h
> index 9f0a455..7d0ea33 100644
> --- a/include/trace/events/fs_dax.h
> +++ b/include/trace/events/fs_dax.h
> @@ -104,6 +104,57 @@ DEFINE_EVENT(dax_pmd_load_hole_class, name, \
>  DEFINE_PMD_LOAD_HOLE_EVENT(dax_pmd_load_hole);
>  DEFINE_PMD_LOAD_HOLE_EVENT(dax_pmd_load_hole_fallback);
>  
> +DECLARE_EVENT_CLASS(dax_pmd_insert_mapping_class,
> +     TP_PROTO(struct inode *inode, struct vm_area_struct *vma,
> +             unsigned long address, int write, long length, pfn_t pfn,
> +             void *radix_entry),
> +     TP_ARGS(inode, vma, address, write, length, pfn, radix_entry),
> +     TP_STRUCT__entry(
> +             __field(dev_t, dev)
> +             __field(unsigned long, ino)
> +             __field(unsigned long, vm_flags)
> +             __field(unsigned long, address)
> +             __field(int, write)

Place "write" at the end. The ring buffer is 4 byte aligned, so on
archs that can access 8 bytes on 4 byte alignment, this will be packed
tighter. Otherwise, you'll get 4 empty bytes after "write".

-- Steve

> +             __field(long, length)
> +             __field(u64, pfn_val)
> +             __field(void *, radix_entry)
> +     ),
> +     TP_fast_assign(
> +             __entry->dev = inode->i_sb->s_dev;
> +             __entry->ino = inode->i_ino;
> +             __entry->vm_flags = vma->vm_flags;
> +             __entry->address = address;
> +             __entry->write = write;
> +             __entry->length = length;
> +             __entry->pfn_val = pfn.val;
> +             __entry->radix_entry = radix_entry;
> +     ),
> +     TP_printk("dev %d:%d ino %#lx %s %s address %#lx length %#lx "
> +                     "pfn %#llx %s radix_entry %#lx",
> +             MAJOR(__entry->dev),
> +             MINOR(__entry->dev),
> +             __entry->ino,
> +             __entry->vm_flags & VM_SHARED ? "shared" : "private",
> +             __entry->write ? "write" : "read",
> +             __entry->address,
> +             __entry->length,
> +             __entry->pfn_val & ~PFN_FLAGS_MASK,
> +             __print_flags_u64(__entry->pfn_val & PFN_FLAGS_MASK, "|",
> +                     PFN_FLAGS_TRACE),
> +             (unsigned long)__entry->radix_entry
> +     )
> +)
> +
> +#define DEFINE_PMD_INSERT_MAPPING_EVENT(name) \
> +DEFINE_EVENT(dax_pmd_insert_mapping_class, name, \
> +     TP_PROTO(struct inode *inode, struct vm_area_struct *vma, \
> +             unsigned long address, int write, long length, pfn_t pfn, \
> +             void *radix_entry), \
> +     TP_ARGS(inode, vma, address, write, length, pfn, radix_entry))
> +
> +DEFINE_PMD_INSERT_MAPPING_EVENT(dax_pmd_insert_mapping);
> +DEFINE_PMD_INSERT_MAPPING_EVENT(dax_pmd_insert_mapping_fallback);
> +
>  #endif /* _TRACE_FS_DAX_H */
>  
>  /* This part must be outside protection */

_______________________________________________
Linux-nvdimm mailing list
Linux-nvdimm@lists.01.org
https://lists.01.org/mailman/listinfo/linux-nvdimm

Reply via email to