When Shared Virtual Memory is exposed to a guest via vIOMMU, extended
IOTLB invalidation may be passed down from outside IOMMU subsystems.
This patch adds invalidation functions that can be used for additional
translation cache types.
Signed-off-by: Jacob Pan
---
drivers/iommu/dmar.c|
Shared virtual address (SVA), a.k.a, Shared virtual memory (SVM) on Intel
platforms allow address space sharing between device DMA and applications.
SVA can reduce programming complexity and enhance security. To enable SVA
in the guest, i.e. shared guest application address space and physical
DMA faults can be detected by IOMMU at device level. Adding a pointer
to struct device allows IOMMU subsystem to report relevant faults
back to the device driver for further handling.
For direct assigned device (or user space drivers), guest OS holds
responsibility to handle and respond per device
Device faults detected by IOMMU can be reported outside IOMMU
subsystem for further processing. This patch intends to provide
a generic device fault data such that device drivers can be
communicated with IOMMU faults without model specific knowledge.
The proposed format is the result of
DMA faults can be detected by IOMMU at device level. Adding a pointer
to struct device allows IOMMU subsystem to report relevant faults
back to the device driver for further handling.
For direct assigned device (or user space drivers), guest OS holds
responsibility to handle and respond per device
Device faults detected by IOMMU can be reported outside IOMMU
subsystem for further processing. This patch intends to provide
a generic device fault data such that device drivers can be
communicated with IOMMU faults without model specific knowledge.
The proposed format is the result of
Allow both intel-iommu.c and dmar.c to access device_domain_info.
Prepare for additional per device arch data used in TLB flush function
Signed-off-by: Jacob Pan
---
drivers/iommu/intel-iommu.c | 18 --
include/linux/intel-iommu.h | 19
Allow both intel-iommu.c and dmar.c to access device_domain_info.
Prepare for additional per device arch data used in TLB flush function
Signed-off-by: Jacob Pan
---
drivers/iommu/intel-iommu.c | 18 --
include/linux/intel-iommu.h | 19 +++
2 files changed, 19
When Shared Virtual Address (SVA) is enabled for a guest OS via
vIOMMU, we need to provide invalidation support at IOMMU API and driver
level. This patch adds Intel VT-d specific function to implement
iommu passdown invalidate API for shared virtual address.
The use case is for supporting caching
When Shared Virtual Address (SVA) is enabled for a guest OS via
vIOMMU, we need to provide invalidation support at IOMMU API and driver
level. This patch adds Intel VT-d specific function to implement
iommu passdown invalidate API for shared virtual address.
The use case is for supporting caching
When an IO page request is processed outside IOMMU subsystem, response can be
delayed or lost. Add a tunable setup parameter such that user can chooose
the timeout for IOMMU to track pending page requests.
This timeout mechanism is a basic safty net which can be implemented in
conjunction
with
When an IO page request is processed outside IOMMU subsystem, response can be
delayed or lost. Add a tunable setup parameter such that user can chooose
the timeout for IOMMU to track pending page requests.
This timeout mechanism is a basic safty net which can be implemented in
conjunction
with
On Fri, May 11, 2018 at 01:17:04PM -0300, Paul Cercueil wrote:
> Le 11 mai 2018 11:52, James Hogan a écrit :
> > Otherwise
> > Acked-by: James Hogan
> >
> > I'm happy to apply for 4.18 with that change if you want it to go
> > through the MIPS tree.
>
>
Intel VT-d interrupts come from both IRQ remapping and DMA remapping.
In order to report non-recoverable faults back to device driver, we
need to have access to IOMMU fault reporting APIs. This patch adds
build depenency to DMAR code where fault IRQ handlers can selectively
report faults.
On Fri, May 11, 2018 at 01:17:04PM -0300, Paul Cercueil wrote:
> Le 11 mai 2018 11:52, James Hogan a écrit :
> > Otherwise
> > Acked-by: James Hogan
> >
> > I'm happy to apply for 4.18 with that change if you want it to go
> > through the MIPS tree.
>
> Yes please!
Done
Thanks
James
Intel VT-d interrupts come from both IRQ remapping and DMA remapping.
In order to report non-recoverable faults back to device driver, we
need to have access to IOMMU fault reporting APIs. This patch adds
build depenency to DMAR code where fault IRQ handlers can selectively
report faults.
When IO page faults are reported outside IOMMU subsystem, the page
request handler may fail for various reasons. E.g. a guest received
page requests but did not have a chance to run for a long time. The
irresponsive behavior could hold off limited resources on the pending
device.
There can be
When IO page faults are reported outside IOMMU subsystem, the page
request handler may fail for various reasons. E.g. a guest received
page requests but did not have a chance to run for a long time. The
irresponsive behavior could hold off limited resources on the pending
device.
There can be
Traditionally, device specific faults are detected and handled within
their own device drivers. When IOMMU is enabled, faults such as DMA
related transactions are detected by IOMMU. There is no generic
reporting mechanism to report faults back to the in-kernel device
driver or the guest OS in case
Traditionally, device specific faults are detected and handled within
their own device drivers. When IOMMU is enabled, faults such as DMA
related transactions are detected by IOMMU. There is no generic
reporting mechanism to report faults back to the in-kernel device
driver or the guest OS in case
Currently, dmar fault IRQ handler does nothing more than rate
limited printk, no critical hardware handling need to be done
in IRQ context.
For some use case such as vIOMMU, it might be useful to report
non-recoverable faults outside host IOMMU subsystem. DMAR fault
can come from both DMA and
IO page faults can be handled outside IOMMU subsystem. For an example,
when nested translation is turned on and guest owns the
first level page tables, device page request can be forwared
to the guest for handling faults. As the page response returns
by the guest, IOMMU driver on the host need to
Currently, dmar fault IRQ handler does nothing more than rate
limited printk, no critical hardware handling need to be done
in IRQ context.
For some use case such as vIOMMU, it might be useful to report
non-recoverable faults outside host IOMMU subsystem. DMAR fault
can come from both DMA and
IO page faults can be handled outside IOMMU subsystem. For an example,
when nested translation is turned on and guest owns the
first level page tables, device page request can be forwared
to the guest for handling faults. As the page response returns
by the guest, IOMMU driver on the host need to
With the introduction of generic IOMMU device fault reporting API, we
can replace the private fault callback functions with standard function
and event data.
Signed-off-by: Jacob Pan
---
drivers/iommu/intel-svm.c | 7 +--
include/linux/intel-svm.h | 20
With the introduction of generic IOMMU device fault reporting API, we
can replace the private fault callback functions with standard function
and event data.
Signed-off-by: Jacob Pan
---
drivers/iommu/intel-svm.c | 7 +--
include/linux/intel-svm.h | 20 +++-
2 files
Signed-off-by: Jacob Pan
---
include/trace/events/iommu.h | 112 +++
1 file changed, 112 insertions(+)
diff --git a/include/trace/events/iommu.h b/include/trace/events/iommu.h
index 72b4582..e64eb29 100644
---
When SRIOV VF device IOTLB is invalidated, we need to provide
the PF source ID such that IOMMU hardware can gauge the depth
of invalidation queue which is shared among VFs. This is needed
when device invalidation throttle (DIT) capability is supported.
This patch adds bit definitions for checking
When SRIOV VF device IOTLB is invalidated, we need to provide
the PF source ID such that IOMMU hardware can gauge the depth
of invalidation queue which is shared among VFs. This is needed
when device invalidation throttle (DIT) capability is supported.
This patch adds bit definitions for checking
Signed-off-by: Jacob Pan
---
include/trace/events/iommu.h | 112 +++
1 file changed, 112 insertions(+)
diff --git a/include/trace/events/iommu.h b/include/trace/events/iommu.h
index 72b4582..e64eb29 100644
--- a/include/trace/events/iommu.h
+++
If the source device of a page request has its PASID table pointer
bound to a guest, the first level page tables are owned by the guest.
In this case, we shall let guest OS to manage page fault.
This patch uses the IOMMU fault reporting API to send fault events,
possibly via VFIO, to the guest
For performance and debugging purposes, these trace events help
analyzing device faults and passdown invalidations that interact
with IOMMU subsystem.
E.g.
IOMMU::00:0a.0 type=2 reason=0 addr=0x007ff000 pasid=1
group=1 last=0 prot=1
Signed-off-by: Jacob Pan
This patch adds page response support for Intel VT-d.
Generic response data is taken from the IOMMU API
then parsed into VT-d specific response descriptor format.
Signed-off-by: Jacob Pan
---
drivers/iommu/intel-iommu.c | 47
For performance and debugging purposes, these trace events help
analyzing device faults and passdown invalidations that interact
with IOMMU subsystem.
E.g.
IOMMU::00:0a.0 type=2 reason=0 addr=0x007ff000 pasid=1
group=1 last=0 prot=1
Signed-off-by: Jacob Pan
---
drivers/iommu/iommu.c
This patch adds page response support for Intel VT-d.
Generic response data is taken from the IOMMU API
then parsed into VT-d specific response descriptor format.
Signed-off-by: Jacob Pan
---
drivers/iommu/intel-iommu.c | 47 +
If the source device of a page request has its PASID table pointer
bound to a guest, the first level page tables are owned by the guest.
In this case, we shall let guest OS to manage page fault.
This patch uses the IOMMU fault reporting API to send fault events,
possibly via VFIO, to the guest
vIOMMU passdown invalidation will be inclusive, PASID cache invalidation
includes TLBs. See Intel VT-d Specification Ch 6.5.2.2 for details.
Signed-off-by: Jacob Pan
---
drivers/iommu/intel-svm.c | 4 +++-
1 file changed, 3 insertions(+), 1 deletion(-)
diff --git
vIOMMU passdown invalidation will be inclusive, PASID cache invalidation
includes TLBs. See Intel VT-d Specification Ch 6.5.2.2 for details.
Signed-off-by: Jacob Pan
---
drivers/iommu/intel-svm.c | 4 +++-
1 file changed, 3 insertions(+), 1 deletion(-)
diff --git a/drivers/iommu/intel-svm.c
PFSID should be used in the invalidation descriptor for flushing
device IOTLBs on SRIOV VFs.
Signed-off-by: Jacob Pan
---
drivers/iommu/dmar.c| 6 +++---
drivers/iommu/intel-iommu.c | 16 +++-
include/linux/intel-iommu.h | 5 ++---
3 files
From: "Liu, Yi L"
When an SVM capable device is assigned to a guest, the first level page
tables are owned by the guest and the guest PASID table pointer is
linked to the device context entry of the physical IOMMU.
Host IOMMU driver has no knowledge of caching
PFSID should be used in the invalidation descriptor for flushing
device IOTLBs on SRIOV VFs.
Signed-off-by: Jacob Pan
---
drivers/iommu/dmar.c| 6 +++---
drivers/iommu/intel-iommu.c | 16 +++-
include/linux/intel-iommu.h | 5 ++---
3 files changed, 20 insertions(+), 7
From: "Liu, Yi L"
When an SVM capable device is assigned to a guest, the first level page
tables are owned by the guest and the guest PASID table pointer is
linked to the device context entry of the physical IOMMU.
Host IOMMU driver has no knowledge of caching structure updates unless
the guest
On Fri, May 11, 2018 at 12:08:33PM -0700, Eric Dumazet wrote:
>
>
> On 05/11/2018 11:41 AM, Marcelo Ricardo Leitner wrote:
>
> > But calling ip6_xmit with rcu_read_lock is expected. tcp stack also
> > does it.
> > Thus I think this is more of an issue with IPv6 stack. If a host has
> > an
On Fri, May 11, 2018 at 12:08:33PM -0700, Eric Dumazet wrote:
>
>
> On 05/11/2018 11:41 AM, Marcelo Ricardo Leitner wrote:
>
> > But calling ip6_xmit with rcu_read_lock is expected. tcp stack also
> > does it.
> > Thus I think this is more of an issue with IPv6 stack. If a host has
> > an
On Wed, May 09, 2018 at 04:05:24PM +0530, Viresh Kumar wrote:
> The schedutil driver sets sg_policy->next_freq to UINT_MAX on certain
> occasions to discard the cached value of next freq:
> - In sugov_start(), when the schedutil governor is started for a group
> of CPUs.
> - And whenever we need
On Wed, May 09, 2018 at 04:05:24PM +0530, Viresh Kumar wrote:
> The schedutil driver sets sg_policy->next_freq to UINT_MAX on certain
> occasions to discard the cached value of next freq:
> - In sugov_start(), when the schedutil governor is started for a group
> of CPUs.
> - And whenever we need
From: Dan Murphy
Date: Fri, 11 May 2018 13:08:19 -0500
> Add support for the DP83811 phy.
>
> The DP83811 supports both rgmii and sgmii interfaces.
> There are 2 part numbers for this the DP83TC811R does not
> reliably support the SGMII interface but the DP83TC811S will.
>
>
From: Dan Murphy
Date: Fri, 11 May 2018 13:08:19 -0500
> Add support for the DP83811 phy.
>
> The DP83811 supports both rgmii and sgmii interfaces.
> There are 2 part numbers for this the DP83TC811R does not
> reliably support the SGMII interface but the DP83TC811S will.
>
> There is not a way
Dan,
On 05/11/2018 02:12 PM, Dan Murphy wrote:
Jacek
Thanks for the review
On 05/10/2018 03:17 PM, Jacek Anaszewski wrote:
Hi Dan,
On 05/10/2018 09:10 PM, Dan Murphy wrote:
On 05/10/2018 02:06 PM, Dan Murphy wrote:
Pavel
On 05/10/2018 01:54 PM, Pavel Machek wrote:
Hi!
Introduce the
Dan,
On 05/11/2018 02:12 PM, Dan Murphy wrote:
Jacek
Thanks for the review
On 05/10/2018 03:17 PM, Jacek Anaszewski wrote:
Hi Dan,
On 05/10/2018 09:10 PM, Dan Murphy wrote:
On 05/10/2018 02:06 PM, Dan Murphy wrote:
Pavel
On 05/10/2018 01:54 PM, Pavel Machek wrote:
Hi!
Introduce the
Hi Dan,
On 05/11/2018 01:56 PM, Dan Murphy wrote:
Jacek
Thanks for the review
On 05/10/2018 03:48 PM, Jacek Anaszewski wrote:
Hi Dan,
Thank you for the patch.
On 05/10/2018 07:47 PM, Dan Murphy wrote:
Introduce the family of LED devices that can
drive a torch, strobe or IR LED.
The LED
Hi Dan,
On 05/11/2018 01:56 PM, Dan Murphy wrote:
Jacek
Thanks for the review
On 05/10/2018 03:48 PM, Jacek Anaszewski wrote:
Hi Dan,
Thank you for the patch.
On 05/10/2018 07:47 PM, Dan Murphy wrote:
Introduce the family of LED devices that can
drive a torch, strobe or IR LED.
The LED
From: Casey Schaufler
Date: Fri, 11 May 2018 13:18:11 -0700
Subject: [PATCH 24/23] LSM: Functions for deling with struct secids
These are the functions that mainipulate the collection
of secids.
Signed-off-by: Casey Schaufler
---
From: Casey Schaufler
Date: Fri, 11 May 2018 13:18:11 -0700
Subject: [PATCH 24/23] LSM: Functions for deling with struct secids
These are the functions that mainipulate the collection
of secids.
Signed-off-by: Casey Schaufler
---
security/stacking.c | 119
On 05/11/2018 02:02 PM, Tejun Heo wrote:
> On Fri, May 11, 2018 at 01:33:26PM -0600, Shuah Khan wrote:
>> Yeah. I see that. You have a switch for the KSFT_ values. Since there is no
>> dependency on the cgroup tree, I would recommend having this patch go through
>> kselftest tree which is the
On 05/11/2018 02:02 PM, Tejun Heo wrote:
> On Fri, May 11, 2018 at 01:33:26PM -0600, Shuah Khan wrote:
>> Yeah. I see that. You have a switch for the KSFT_ values. Since there is no
>> dependency on the cgroup tree, I would recommend having this patch go through
>> kselftest tree which is the
Hi,
On Wed, May 9, 2018 at 10:01 AM, Lina Iyer wrote:
> /**
> @@ -77,12 +82,14 @@ struct rpmh_request {
> * @cache: the list of cached requests
> * @lock: synchronize access to the controller data
> * @dirty: was the cache updated since flush
> + * @batch_cache:
Hi,
On Wed, May 9, 2018 at 10:01 AM, Lina Iyer wrote:
> /**
> @@ -77,12 +82,14 @@ struct rpmh_request {
> * @cache: the list of cached requests
> * @lock: synchronize access to the controller data
> * @dirty: was the cache updated since flush
> + * @batch_cache: Cache sleep and wake
Hi,
On Wed, May 9, 2018 at 10:01 AM, Lina Iyer wrote:
> /**
> * struct rpmh_request: the message to be sent to rpmh-rsc
> *
> @@ -54,9 +71,15 @@ struct rpmh_request {
> * struct rpmh_ctrlr: our representation of the controller
> *
> * @drv: the controller
Hi,
On Wed, May 9, 2018 at 10:01 AM, Lina Iyer wrote:
> /**
> * struct rpmh_request: the message to be sent to rpmh-rsc
> *
> @@ -54,9 +71,15 @@ struct rpmh_request {
> * struct rpmh_ctrlr: our representation of the controller
> *
> * @drv: the controller instance
> + * @cache: the
Hi,
On Wed, May 9, 2018 at 10:01 AM, Lina Iyer wrote:
> Some RSCs may only have sleep and wake TCS, i.e, there is no dedicated
> TCS for active mode request, but drivers may still want to make active
> requests from these RSCs. In such cases re-purpose the wake TCS to send
Hi,
On Wed, May 9, 2018 at 10:01 AM, Lina Iyer wrote:
> Some RSCs may only have sleep and wake TCS, i.e, there is no dedicated
> TCS for active mode request, but drivers may still want to make active
> requests from these RSCs. In such cases re-purpose the wake TCS to send
> active state
Hi,
On Wed, May 9, 2018 at 10:01 AM, Lina Iyer wrote:
> +int rpmh_write(const struct device *dev, enum rpmh_state state,
> + const struct tcs_cmd *cmd, u32 n)
> +{
> + DECLARE_COMPLETION_ONSTACK(compl);
> + DEFINE_RPMH_MSG_ONSTACK(dev, state, ,
Hi,
On Wed, May 9, 2018 at 10:01 AM, Lina Iyer wrote:
> +int rpmh_write(const struct device *dev, enum rpmh_state state,
> + const struct tcs_cmd *cmd, u32 n)
> +{
> + DECLARE_COMPLETION_ONSTACK(compl);
> + DEFINE_RPMH_MSG_ONSTACK(dev, state, , rpm_msg);
> + int
Hi,
On Wed, May 9, 2018 at 10:01 AM, Lina Iyer wrote:
> /**
> @@ -137,6 +140,8 @@ void rpmh_tx_done(const struct tcs_request *msg, int r)
> dev_err(rpm_msg->dev, "RPMH TX fail in msg addr=%#x,
> err=%d\n",
>
Hi,
On Wed, May 9, 2018 at 10:01 AM, Lina Iyer wrote:
> /**
> @@ -137,6 +140,8 @@ void rpmh_tx_done(const struct tcs_request *msg, int r)
> dev_err(rpm_msg->dev, "RPMH TX fail in msg addr=%#x,
> err=%d\n",
> rpm_msg->msg.cmds[0].addr, r);
>
> +
Hi,
On Wed, May 9, 2018 at 10:01 AM, Lina Iyer wrote:
> +int rpmh_rsc_send_data(struct rsc_drv *drv, const struct tcs_request *msg)
> +{
> + int ret;
> +
> + if (!msg || !msg->cmds || !msg->num_cmds ||
> + msg->num_cmds > MAX_RPMH_PAYLOAD) {
> +
Hi,
On Wed, May 9, 2018 at 10:01 AM, Lina Iyer wrote:
> +int rpmh_rsc_send_data(struct rsc_drv *drv, const struct tcs_request *msg)
> +{
> + int ret;
> +
> + if (!msg || !msg->cmds || !msg->num_cmds ||
> + msg->num_cmds > MAX_RPMH_PAYLOAD) {
> + WARN_ON(1);
>
Hi,
On Fri, May 11, 2018 at 8:06 AM, Lina Iyer wrote:
>> As I've said I haven't reviewed RPMh in any amount of detail and so
>> perhaps I don't understand something.
>>
>> OK, I dug a little more and coded up something for you. Basically
>> you're doing a whole bunch of
Hi,
On Fri, May 11, 2018 at 8:06 AM, Lina Iyer wrote:
>> As I've said I haven't reviewed RPMh in any amount of detail and so
>> perhaps I don't understand something.
>>
>> OK, I dug a little more and coded up something for you. Basically
>> you're doing a whole bunch of iteration / extra work
On Fri, May 11, 2018 at 11:20 AM, Bartosz Golaszewski wrote:
> This series is a follow-up to the RFC[1] posted a couple days ago.
>
> NOTE: this series applies on top of my recent patches[2] that move the
> previous
> implementation of early platform devices to arch/sh.
>
>
On Fri, May 11, 2018 at 11:20 AM, Bartosz Golaszewski wrote:
> This series is a follow-up to the RFC[1] posted a couple days ago.
>
> NOTE: this series applies on top of my recent patches[2] that move the
> previous
> implementation of early platform devices to arch/sh.
>
> Problem:
>
> Certain
On 11.05.2018 16:02, Robin Murphy wrote:
> On 08/05/18 19:16, Dmitry Osipenko wrote:
>> Introduce iotlb_sync_map() callback that is invoked in the end of
>> iommu_map(). This new callback allows IOMMU drivers to avoid syncing
>> on mapping of each contiguous chunk and sync only when whole mapping
On 11.05.2018 15:32, Robin Murphy wrote:
> On 08/05/18 19:16, Dmitry Osipenko wrote:
>> GART aperture is shared by all devices, hence there is a single IOMMU
>> domain and group shared by these devices. Allocation of a group per
>> device only wastes resources and allowance of having more than one
On 11.05.2018 16:02, Robin Murphy wrote:
> On 08/05/18 19:16, Dmitry Osipenko wrote:
>> Introduce iotlb_sync_map() callback that is invoked in the end of
>> iommu_map(). This new callback allows IOMMU drivers to avoid syncing
>> on mapping of each contiguous chunk and sync only when whole mapping
On 11.05.2018 15:32, Robin Murphy wrote:
> On 08/05/18 19:16, Dmitry Osipenko wrote:
>> GART aperture is shared by all devices, hence there is a single IOMMU
>> domain and group shared by these devices. Allocation of a group per
>> device only wastes resources and allowance of having more than one
On Tue, 24 Apr 2018, Tycho Andersen wrote:
> We're interested in getting rid of all of the stack allocated arrays in the
> kernel [1]. This patch simply hardcodes the iv length to match that of the
> hardcoded cipher.
>
> [1]: https://lkml.org/lkml/2018/3/7/621
>
> v2: hardcode the length of
On Tue, 24 Apr 2018, Tycho Andersen wrote:
> We're interested in getting rid of all of the stack allocated arrays in the
> kernel [1]. This patch simply hardcodes the iv length to match that of the
> hardcoded cipher.
>
> [1]: https://lkml.org/lkml/2018/3/7/621
>
> v2: hardcode the length of
Hi Robin,
On 11.05.2018 14:34, Robin Murphy wrote:
> Hi Dmitry,
>
> On 08/05/18 19:16, Dmitry Osipenko wrote:
>> GART can't handle all devices, ignore devices that aren't related to GART.
>> Device tree must explicitly assign GART IOMMU to the devices.
>>
>> Signed-off-by: Dmitry Osipenko
Hi Robin,
On 11.05.2018 14:34, Robin Murphy wrote:
> Hi Dmitry,
>
> On 08/05/18 19:16, Dmitry Osipenko wrote:
>> GART can't handle all devices, ignore devices that aren't related to GART.
>> Device tree must explicitly assign GART IOMMU to the devices.
>>
>> Signed-off-by: Dmitry Osipenko
>>
On Fri, May 11, 2018 at 01:33:26PM -0600, Shuah Khan wrote:
> Yeah. I see that. You have a switch for the KSFT_ values. Since there is no
> dependency on the cgroup tree, I would recommend having this patch go through
> kselftest tree which is the normal process for tests anyway.
>
> This version
On Fri, May 11, 2018 at 01:33:26PM -0600, Shuah Khan wrote:
> Yeah. I see that. You have a switch for the KSFT_ values. Since there is no
> dependency on the cgroup tree, I would recommend having this patch go through
> kselftest tree which is the normal process for tests anyway.
>
> This version
From: David Howells
Date: Thu, 10 May 2018 23:45:17 +0100
> Here are three fixes for AF_RXRPC and two tracepoints that were useful for
> finding them:
...
> The patches are tagged here:
>
> git://git.kernel.org/pub/scm/linux/kernel/git/dhowells/linux-fs.git
>
From: David Howells
Date: Thu, 10 May 2018 23:45:17 +0100
> Here are three fixes for AF_RXRPC and two tracepoints that were useful for
> finding them:
...
> The patches are tagged here:
>
> git://git.kernel.org/pub/scm/linux/kernel/git/dhowells/linux-fs.git
> rxrpc-fixes-20180510
From: "Steven Rostedt (VMware)"
Now that trace_marker can have triggers, including a histogram triggers, the
onmatch() and onmax() access the trace event. To do so, the search routine
to find the event file needs to use the raw __find_event_file() that does
not filter out
From: "Steven Rostedt (VMware)"
Now that trace_marker can have triggers, including a histogram triggers, the
onmatch() and onmax() access the trace event. To do so, the search routine
to find the event file needs to use the raw __find_event_file() that does
not filter out ftrace events.
From: "Steven Rostedt (VMware)"
The filter file in the ftrace internal events, like in
/sys/kernel/tracing/events/ftrace/function/filter is not attached to any
functionality. Do not create them as they are meaningless.
In the future, if an ftrace internal event gets filter
From: "Steven Rostedt (VMware)"
The dynamic arrays defined for ftrace internal events, such as the buf field
for trace_marker (ftrace/print) did not have brackets which makes the filter
code not accept it as a string. This is not currently an issues because the
filter code
From: "Steven Rostedt (VMware)"
The dynamic arrays defined for ftrace internal events, such as the buf field
for trace_marker (ftrace/print) did not have brackets which makes the filter
code not accept it as a string. This is not currently an issues because the
filter code doesn't do anything
From: "Steven Rostedt (VMware)"
The filter file in the ftrace internal events, like in
/sys/kernel/tracing/events/ftrace/function/filter is not attached to any
functionality. Do not create them as they are meaningless.
In the future, if an ftrace internal event gets filter functionality, then
From: "Steven Rostedt (VMware)"
As strings in trace events may not have a nul terminating character, the
filter string compares use the defined string length for the field for the
compares.
The trace_marker records data slightly different than do normal events. It's
size is
From: "Steven Rostedt (VMware)"
As strings in trace events may not have a nul terminating character, the
filter string compares use the defined string length for the field for the
compares.
The trace_marker records data slightly different than do normal events. It's
size is zero, meaning that
From: "Steven Rostedt (VMware)"
Allow writing to the trace_markers file initiate triggers defined in
tracefs/ftrace/print/trigger file. This will allow of user space to trigger
the same type of triggers (including histograms) that the trace events use.
Cc: Tom Zanussi
From: "Steven Rostedt (VMware)"
Allow writing to the trace_markers file initiate triggers defined in
tracefs/ftrace/print/trigger file. This will allow of user space to trigger
the same type of triggers (including histograms) that the trace events use.
Cc: Tom Zanussi
Cc: Clark Williams
Cc:
From: "Steven Rostedt (VMware)"
A zero size static array has special meaning in the ftrace infrastructure.
Trace events are for recording data in the trace buffers that is normally
difficult to obtain via probes or function tracing. There is no reason for
any trace event to
From: "Steven Rostedt (VMware)"
A zero size static array has special meaning in the ftrace infrastructure.
Trace events are for recording data in the trace buffers that is normally
difficult to obtain via probes or function tracing. There is no reason for
any trace event to declare a zero size
From: "Steven Rostedt (VMware)"
Add documentation and an example on how to use trace_marker triggers.
Signed-off-by: Steven Rostedt (VMware)
---
Documentation/trace/events.rst| 6 +-
Documentation/trace/ftrace.rst| 5 +
From: "Steven Rostedt (VMware)"
Add documentation and an example on how to use trace_marker triggers.
Signed-off-by: Steven Rostedt (VMware)
---
Documentation/trace/events.rst| 6 +-
Documentation/trace/ftrace.rst| 5 +
Documentation/trace/histogram.txt | 546
From: "Steven Rostedt (VMware)"
Trace event triggers can be called before or after the event has been
committed. If it has been called after the commit, there's a possibility
that the event no longer exists. Currently, the two post callers is the
trigger to disable tracing
From: "Steven Rostedt (VMware)"
Trace event triggers can be called before or after the event has been
committed. If it has been called after the commit, there's a possibility
that the event no longer exists. Currently, the two post callers is the
trigger to disable tracing (traceoff) and the one
201 - 300 of 1384 matches
Mail list logo