From: Steven Rostedt <[email protected]>

The ftrace_lock is held throughout cache_mod(), use guard to simplify the
error paths.

Signed-off-by: Steven Rostedt (Google) <[email protected]>
---
 kernel/trace/ftrace.c | 17 ++++++-----------
 1 file changed, 6 insertions(+), 11 deletions(-)

diff --git a/kernel/trace/ftrace.c b/kernel/trace/ftrace.c
index 44adc34643c9..64997416415e 100644
--- a/kernel/trace/ftrace.c
+++ b/kernel/trace/ftrace.c
@@ -4947,14 +4947,14 @@ static int cache_mod(struct trace_array *tr,
 {
        struct ftrace_mod_load *ftrace_mod, *n;
        struct list_head *head = enable ? &tr->mod_trace : &tr->mod_notrace;
-       int ret;
 
-       mutex_lock(&ftrace_lock);
+       guard(mutex)(&ftrace_lock);
 
        /* We do not cache inverse filters */
        if (func[0] == '!') {
+               int ret = -EINVAL;
+
                func++;
-               ret = -EINVAL;
 
                /* Look to remove this hash */
                list_for_each_entry_safe(ftrace_mod, n, head, list) {
@@ -4970,20 +4970,15 @@ static int cache_mod(struct trace_array *tr,
                                continue;
                        }
                }
-               goto out;
+               return ret;
        }
 
-       ret = -EINVAL;
        /* We only care about modules that have not been loaded yet */
        if (module_exists(module))
-               goto out;
+               return -EINVAL;
 
        /* Save this string off, and execute it when the module is loaded */
-       ret = ftrace_add_mod(tr, func, module, enable);
- out:
-       mutex_unlock(&ftrace_lock);
-
-       return ret;
+       return ftrace_add_mod(tr, func, module, enable);
 }
 
 static int
-- 
2.45.2



Reply via email to