blob: d0309de2f84fea5bfce5cbe4a8321be92c547057 [file] [log] [blame]
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00001// SPDX-License-Identifier: GPL-2.0
2/*
3 * trace_events_trigger - trace event triggers
4 *
5 * Copyright (C) 2013 Tom Zanussi <tom.zanussi@linux.intel.com>
6 */
7
David Brazdil0f672f62019-12-10 10:32:29 +00008#include <linux/security.h>
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00009#include <linux/module.h>
10#include <linux/ctype.h>
11#include <linux/mutex.h>
12#include <linux/slab.h>
13#include <linux/rculist.h>
14
15#include "trace.h"
16
17static LIST_HEAD(trigger_commands);
18static DEFINE_MUTEX(trigger_cmd_mutex);
19
20void trigger_data_free(struct event_trigger_data *data)
21{
22 if (data->cmd_ops->set_filter)
23 data->cmd_ops->set_filter(NULL, data, NULL);
24
25 /* make sure current triggers exit before free */
26 tracepoint_synchronize_unregister();
27
28 kfree(data);
29}
30
31/**
32 * event_triggers_call - Call triggers associated with a trace event
33 * @file: The trace_event_file associated with the event
34 * @rec: The trace entry for the event, NULL for unconditional invocation
35 *
36 * For each trigger associated with an event, invoke the trigger
37 * function registered with the associated trigger command. If rec is
38 * non-NULL, it means that the trigger requires further processing and
39 * shouldn't be unconditionally invoked. If rec is non-NULL and the
40 * trigger has a filter associated with it, rec will checked against
41 * the filter and if the record matches the trigger will be invoked.
42 * If the trigger is a 'post_trigger', meaning it shouldn't be invoked
43 * in any case until the current event is written, the trigger
44 * function isn't invoked but the bit associated with the deferred
45 * trigger is set in the return value.
46 *
47 * Returns an enum event_trigger_type value containing a set bit for
48 * any trigger that should be deferred, ETT_NONE if nothing to defer.
49 *
50 * Called from tracepoint handlers (with rcu_read_lock_sched() held).
51 *
52 * Return: an enum event_trigger_type value containing a set bit for
53 * any trigger that should be deferred, ETT_NONE if nothing to defer.
54 */
55enum event_trigger_type
56event_triggers_call(struct trace_event_file *file, void *rec,
57 struct ring_buffer_event *event)
58{
59 struct event_trigger_data *data;
60 enum event_trigger_type tt = ETT_NONE;
61 struct event_filter *filter;
62
63 if (list_empty(&file->triggers))
64 return tt;
65
66 list_for_each_entry_rcu(data, &file->triggers, list) {
67 if (data->paused)
68 continue;
69 if (!rec) {
70 data->ops->func(data, rec, event);
71 continue;
72 }
73 filter = rcu_dereference_sched(data->filter);
74 if (filter && !filter_match_preds(filter, rec))
75 continue;
76 if (event_command_post_trigger(data->cmd_ops)) {
77 tt |= data->cmd_ops->trigger_type;
78 continue;
79 }
80 data->ops->func(data, rec, event);
81 }
82 return tt;
83}
84EXPORT_SYMBOL_GPL(event_triggers_call);
85
86/**
87 * event_triggers_post_call - Call 'post_triggers' for a trace event
88 * @file: The trace_event_file associated with the event
89 * @tt: enum event_trigger_type containing a set bit for each trigger to invoke
90 *
91 * For each trigger associated with an event, invoke the trigger
92 * function registered with the associated trigger command, if the
93 * corresponding bit is set in the tt enum passed into this function.
94 * See @event_triggers_call for details on how those bits are set.
95 *
96 * Called from tracepoint handlers (with rcu_read_lock_sched() held).
97 */
98void
99event_triggers_post_call(struct trace_event_file *file,
100 enum event_trigger_type tt)
101{
102 struct event_trigger_data *data;
103
104 list_for_each_entry_rcu(data, &file->triggers, list) {
105 if (data->paused)
106 continue;
107 if (data->cmd_ops->trigger_type & tt)
108 data->ops->func(data, NULL, NULL);
109 }
110}
111EXPORT_SYMBOL_GPL(event_triggers_post_call);
112
113#define SHOW_AVAILABLE_TRIGGERS (void *)(1UL)
114
115static void *trigger_next(struct seq_file *m, void *t, loff_t *pos)
116{
117 struct trace_event_file *event_file = event_file_data(m->private);
118
Olivier Deprez0e641232021-09-23 10:07:05 +0200119 if (t == SHOW_AVAILABLE_TRIGGERS) {
120 (*pos)++;
Andrew Scullb4b6d4a2019-01-02 15:54:55 +0000121 return NULL;
Olivier Deprez0e641232021-09-23 10:07:05 +0200122 }
Andrew Scullb4b6d4a2019-01-02 15:54:55 +0000123 return seq_list_next(t, &event_file->triggers, pos);
124}
125
126static void *trigger_start(struct seq_file *m, loff_t *pos)
127{
128 struct trace_event_file *event_file;
129
130 /* ->stop() is called even if ->start() fails */
131 mutex_lock(&event_mutex);
132 event_file = event_file_data(m->private);
133 if (unlikely(!event_file))
134 return ERR_PTR(-ENODEV);
135
136 if (list_empty(&event_file->triggers))
137 return *pos == 0 ? SHOW_AVAILABLE_TRIGGERS : NULL;
138
139 return seq_list_start(&event_file->triggers, *pos);
140}
141
142static void trigger_stop(struct seq_file *m, void *t)
143{
144 mutex_unlock(&event_mutex);
145}
146
147static int trigger_show(struct seq_file *m, void *v)
148{
149 struct event_trigger_data *data;
150 struct event_command *p;
151
152 if (v == SHOW_AVAILABLE_TRIGGERS) {
153 seq_puts(m, "# Available triggers:\n");
154 seq_putc(m, '#');
155 mutex_lock(&trigger_cmd_mutex);
156 list_for_each_entry_reverse(p, &trigger_commands, list)
157 seq_printf(m, " %s", p->name);
158 seq_putc(m, '\n');
159 mutex_unlock(&trigger_cmd_mutex);
160 return 0;
161 }
162
163 data = list_entry(v, struct event_trigger_data, list);
164 data->ops->print(m, data->ops, data);
165
166 return 0;
167}
168
169static const struct seq_operations event_triggers_seq_ops = {
170 .start = trigger_start,
171 .next = trigger_next,
172 .stop = trigger_stop,
173 .show = trigger_show,
174};
175
176static int event_trigger_regex_open(struct inode *inode, struct file *file)
177{
David Brazdil0f672f62019-12-10 10:32:29 +0000178 int ret;
179
180 ret = security_locked_down(LOCKDOWN_TRACEFS);
181 if (ret)
182 return ret;
Andrew Scullb4b6d4a2019-01-02 15:54:55 +0000183
184 mutex_lock(&event_mutex);
185
186 if (unlikely(!event_file_data(file))) {
187 mutex_unlock(&event_mutex);
188 return -ENODEV;
189 }
190
191 if ((file->f_mode & FMODE_WRITE) &&
192 (file->f_flags & O_TRUNC)) {
193 struct trace_event_file *event_file;
194 struct event_command *p;
195
196 event_file = event_file_data(file);
197
198 list_for_each_entry(p, &trigger_commands, list) {
199 if (p->unreg_all)
200 p->unreg_all(event_file);
201 }
202 }
203
204 if (file->f_mode & FMODE_READ) {
205 ret = seq_open(file, &event_triggers_seq_ops);
206 if (!ret) {
207 struct seq_file *m = file->private_data;
208 m->private = file;
209 }
210 }
211
212 mutex_unlock(&event_mutex);
213
214 return ret;
215}
216
Olivier Deprez157378f2022-04-04 15:47:50 +0200217int trigger_process_regex(struct trace_event_file *file, char *buff)
Andrew Scullb4b6d4a2019-01-02 15:54:55 +0000218{
Olivier Deprez0e641232021-09-23 10:07:05 +0200219 char *command, *next;
Andrew Scullb4b6d4a2019-01-02 15:54:55 +0000220 struct event_command *p;
221 int ret = -EINVAL;
222
Olivier Deprez0e641232021-09-23 10:07:05 +0200223 next = buff = skip_spaces(buff);
Andrew Scullb4b6d4a2019-01-02 15:54:55 +0000224 command = strsep(&next, ": \t");
Olivier Deprez0e641232021-09-23 10:07:05 +0200225 if (next) {
226 next = skip_spaces(next);
227 if (!*next)
228 next = NULL;
229 }
Andrew Scullb4b6d4a2019-01-02 15:54:55 +0000230 command = (command[0] != '!') ? command : command + 1;
231
232 mutex_lock(&trigger_cmd_mutex);
233 list_for_each_entry(p, &trigger_commands, list) {
234 if (strcmp(p->name, command) == 0) {
235 ret = p->func(p, file, buff, command, next);
236 goto out_unlock;
237 }
238 }
239 out_unlock:
240 mutex_unlock(&trigger_cmd_mutex);
241
242 return ret;
243}
244
245static ssize_t event_trigger_regex_write(struct file *file,
246 const char __user *ubuf,
247 size_t cnt, loff_t *ppos)
248{
249 struct trace_event_file *event_file;
250 ssize_t ret;
251 char *buf;
252
253 if (!cnt)
254 return 0;
255
256 if (cnt >= PAGE_SIZE)
257 return -EINVAL;
258
259 buf = memdup_user_nul(ubuf, cnt);
260 if (IS_ERR(buf))
261 return PTR_ERR(buf);
262
263 strim(buf);
264
265 mutex_lock(&event_mutex);
266 event_file = event_file_data(file);
267 if (unlikely(!event_file)) {
268 mutex_unlock(&event_mutex);
269 kfree(buf);
270 return -ENODEV;
271 }
272 ret = trigger_process_regex(event_file, buf);
273 mutex_unlock(&event_mutex);
274
275 kfree(buf);
276 if (ret < 0)
277 goto out;
278
279 *ppos += cnt;
280 ret = cnt;
281 out:
282 return ret;
283}
284
285static int event_trigger_regex_release(struct inode *inode, struct file *file)
286{
287 mutex_lock(&event_mutex);
288
289 if (file->f_mode & FMODE_READ)
290 seq_release(inode, file);
291
292 mutex_unlock(&event_mutex);
293
294 return 0;
295}
296
297static ssize_t
298event_trigger_write(struct file *filp, const char __user *ubuf,
299 size_t cnt, loff_t *ppos)
300{
301 return event_trigger_regex_write(filp, ubuf, cnt, ppos);
302}
303
304static int
305event_trigger_open(struct inode *inode, struct file *filp)
306{
David Brazdil0f672f62019-12-10 10:32:29 +0000307 /* Checks for tracefs lockdown */
Andrew Scullb4b6d4a2019-01-02 15:54:55 +0000308 return event_trigger_regex_open(inode, filp);
309}
310
311static int
312event_trigger_release(struct inode *inode, struct file *file)
313{
314 return event_trigger_regex_release(inode, file);
315}
316
317const struct file_operations event_trigger_fops = {
318 .open = event_trigger_open,
319 .read = seq_read,
320 .write = event_trigger_write,
321 .llseek = tracing_lseek,
322 .release = event_trigger_release,
323};
324
325/*
326 * Currently we only register event commands from __init, so mark this
327 * __init too.
328 */
329__init int register_event_command(struct event_command *cmd)
330{
331 struct event_command *p;
332 int ret = 0;
333
334 mutex_lock(&trigger_cmd_mutex);
335 list_for_each_entry(p, &trigger_commands, list) {
336 if (strcmp(cmd->name, p->name) == 0) {
337 ret = -EBUSY;
338 goto out_unlock;
339 }
340 }
341 list_add(&cmd->list, &trigger_commands);
342 out_unlock:
343 mutex_unlock(&trigger_cmd_mutex);
344
345 return ret;
346}
347
348/*
349 * Currently we only unregister event commands from __init, so mark
350 * this __init too.
351 */
352__init int unregister_event_command(struct event_command *cmd)
353{
354 struct event_command *p, *n;
355 int ret = -ENODEV;
356
357 mutex_lock(&trigger_cmd_mutex);
358 list_for_each_entry_safe(p, n, &trigger_commands, list) {
359 if (strcmp(cmd->name, p->name) == 0) {
360 ret = 0;
361 list_del_init(&p->list);
362 goto out_unlock;
363 }
364 }
365 out_unlock:
366 mutex_unlock(&trigger_cmd_mutex);
367
368 return ret;
369}
370
371/**
372 * event_trigger_print - Generic event_trigger_ops @print implementation
373 * @name: The name of the event trigger
374 * @m: The seq_file being printed to
375 * @data: Trigger-specific data
376 * @filter_str: filter_str to print, if present
377 *
378 * Common implementation for event triggers to print themselves.
379 *
380 * Usually wrapped by a function that simply sets the @name of the
381 * trigger command and then invokes this.
382 *
383 * Return: 0 on success, errno otherwise
384 */
385static int
386event_trigger_print(const char *name, struct seq_file *m,
387 void *data, char *filter_str)
388{
389 long count = (long)data;
390
391 seq_puts(m, name);
392
393 if (count == -1)
394 seq_puts(m, ":unlimited");
395 else
396 seq_printf(m, ":count=%ld", count);
397
398 if (filter_str)
399 seq_printf(m, " if %s\n", filter_str);
400 else
401 seq_putc(m, '\n');
402
403 return 0;
404}
405
406/**
407 * event_trigger_init - Generic event_trigger_ops @init implementation
408 * @ops: The trigger ops associated with the trigger
409 * @data: Trigger-specific data
410 *
411 * Common implementation of event trigger initialization.
412 *
413 * Usually used directly as the @init method in event trigger
414 * implementations.
415 *
416 * Return: 0 on success, errno otherwise
417 */
418int event_trigger_init(struct event_trigger_ops *ops,
419 struct event_trigger_data *data)
420{
421 data->ref++;
422 return 0;
423}
424
425/**
426 * event_trigger_free - Generic event_trigger_ops @free implementation
427 * @ops: The trigger ops associated with the trigger
428 * @data: Trigger-specific data
429 *
430 * Common implementation of event trigger de-initialization.
431 *
432 * Usually used directly as the @free method in event trigger
433 * implementations.
434 */
435static void
436event_trigger_free(struct event_trigger_ops *ops,
437 struct event_trigger_data *data)
438{
439 if (WARN_ON_ONCE(data->ref <= 0))
440 return;
441
442 data->ref--;
443 if (!data->ref)
444 trigger_data_free(data);
445}
446
447int trace_event_trigger_enable_disable(struct trace_event_file *file,
448 int trigger_enable)
449{
450 int ret = 0;
451
452 if (trigger_enable) {
453 if (atomic_inc_return(&file->tm_ref) > 1)
454 return ret;
455 set_bit(EVENT_FILE_FL_TRIGGER_MODE_BIT, &file->flags);
456 ret = trace_event_enable_disable(file, 1, 1);
457 } else {
458 if (atomic_dec_return(&file->tm_ref) > 0)
459 return ret;
460 clear_bit(EVENT_FILE_FL_TRIGGER_MODE_BIT, &file->flags);
461 ret = trace_event_enable_disable(file, 0, 1);
462 }
463
464 return ret;
465}
466
467/**
468 * clear_event_triggers - Clear all triggers associated with a trace array
469 * @tr: The trace array to clear
470 *
471 * For each trigger, the triggering event has its tm_ref decremented
472 * via trace_event_trigger_enable_disable(), and any associated event
473 * (in the case of enable/disable_event triggers) will have its sm_ref
474 * decremented via free()->trace_event_enable_disable(). That
475 * combination effectively reverses the soft-mode/trigger state added
476 * by trigger registration.
477 *
478 * Must be called with event_mutex held.
479 */
480void
481clear_event_triggers(struct trace_array *tr)
482{
483 struct trace_event_file *file;
484
485 list_for_each_entry(file, &tr->events, list) {
486 struct event_trigger_data *data, *n;
487 list_for_each_entry_safe(data, n, &file->triggers, list) {
488 trace_event_trigger_enable_disable(file, 0);
489 list_del_rcu(&data->list);
490 if (data->ops->free)
491 data->ops->free(data->ops, data);
492 }
493 }
494}
495
496/**
497 * update_cond_flag - Set or reset the TRIGGER_COND bit
498 * @file: The trace_event_file associated with the event
499 *
500 * If an event has triggers and any of those triggers has a filter or
501 * a post_trigger, trigger invocation needs to be deferred until after
502 * the current event has logged its data, and the event should have
503 * its TRIGGER_COND bit set, otherwise the TRIGGER_COND bit should be
504 * cleared.
505 */
506void update_cond_flag(struct trace_event_file *file)
507{
508 struct event_trigger_data *data;
509 bool set_cond = false;
510
Olivier Deprez0e641232021-09-23 10:07:05 +0200511 lockdep_assert_held(&event_mutex);
512
513 list_for_each_entry(data, &file->triggers, list) {
Andrew Scullb4b6d4a2019-01-02 15:54:55 +0000514 if (data->filter || event_command_post_trigger(data->cmd_ops) ||
515 event_command_needs_rec(data->cmd_ops)) {
516 set_cond = true;
517 break;
518 }
519 }
520
521 if (set_cond)
522 set_bit(EVENT_FILE_FL_TRIGGER_COND_BIT, &file->flags);
523 else
524 clear_bit(EVENT_FILE_FL_TRIGGER_COND_BIT, &file->flags);
525}
526
527/**
528 * register_trigger - Generic event_command @reg implementation
529 * @glob: The raw string used to register the trigger
530 * @ops: The trigger ops associated with the trigger
531 * @data: Trigger-specific data to associate with the trigger
532 * @file: The trace_event_file associated with the event
533 *
534 * Common implementation for event trigger registration.
535 *
536 * Usually used directly as the @reg method in event command
537 * implementations.
538 *
539 * Return: 0 on success, errno otherwise
540 */
541static int register_trigger(char *glob, struct event_trigger_ops *ops,
542 struct event_trigger_data *data,
543 struct trace_event_file *file)
544{
545 struct event_trigger_data *test;
546 int ret = 0;
547
Olivier Deprez0e641232021-09-23 10:07:05 +0200548 lockdep_assert_held(&event_mutex);
549
550 list_for_each_entry(test, &file->triggers, list) {
Andrew Scullb4b6d4a2019-01-02 15:54:55 +0000551 if (test->cmd_ops->trigger_type == data->cmd_ops->trigger_type) {
552 ret = -EEXIST;
553 goto out;
554 }
555 }
556
557 if (data->ops->init) {
558 ret = data->ops->init(data->ops, data);
559 if (ret < 0)
560 goto out;
561 }
562
563 list_add_rcu(&data->list, &file->triggers);
564 ret++;
565
566 update_cond_flag(file);
567 if (trace_event_trigger_enable_disable(file, 1) < 0) {
568 list_del_rcu(&data->list);
569 update_cond_flag(file);
570 ret--;
571 }
572out:
573 return ret;
574}
575
576/**
577 * unregister_trigger - Generic event_command @unreg implementation
578 * @glob: The raw string used to register the trigger
579 * @ops: The trigger ops associated with the trigger
580 * @test: Trigger-specific data used to find the trigger to remove
581 * @file: The trace_event_file associated with the event
582 *
583 * Common implementation for event trigger unregistration.
584 *
585 * Usually used directly as the @unreg method in event command
586 * implementations.
587 */
588static void unregister_trigger(char *glob, struct event_trigger_ops *ops,
589 struct event_trigger_data *test,
590 struct trace_event_file *file)
591{
592 struct event_trigger_data *data;
593 bool unregistered = false;
594
Olivier Deprez0e641232021-09-23 10:07:05 +0200595 lockdep_assert_held(&event_mutex);
596
597 list_for_each_entry(data, &file->triggers, list) {
Andrew Scullb4b6d4a2019-01-02 15:54:55 +0000598 if (data->cmd_ops->trigger_type == test->cmd_ops->trigger_type) {
599 unregistered = true;
600 list_del_rcu(&data->list);
601 trace_event_trigger_enable_disable(file, 0);
602 update_cond_flag(file);
603 break;
604 }
605 }
606
607 if (unregistered && data->ops->free)
608 data->ops->free(data->ops, data);
609}
610
611/**
612 * event_trigger_callback - Generic event_command @func implementation
613 * @cmd_ops: The command ops, used for trigger registration
614 * @file: The trace_event_file associated with the event
615 * @glob: The raw string used to register the trigger
616 * @cmd: The cmd portion of the string used to register the trigger
617 * @param: The params portion of the string used to register the trigger
618 *
619 * Common implementation for event command parsing and trigger
620 * instantiation.
621 *
622 * Usually used directly as the @func method in event command
623 * implementations.
624 *
625 * Return: 0 on success, errno otherwise
626 */
627static int
628event_trigger_callback(struct event_command *cmd_ops,
629 struct trace_event_file *file,
630 char *glob, char *cmd, char *param)
631{
632 struct event_trigger_data *trigger_data;
633 struct event_trigger_ops *trigger_ops;
634 char *trigger = NULL;
635 char *number;
636 int ret;
637
638 /* separate the trigger from the filter (t:n [if filter]) */
Olivier Deprez0e641232021-09-23 10:07:05 +0200639 if (param && isdigit(param[0])) {
Andrew Scullb4b6d4a2019-01-02 15:54:55 +0000640 trigger = strsep(&param, " \t");
Olivier Deprez0e641232021-09-23 10:07:05 +0200641 if (param) {
642 param = skip_spaces(param);
643 if (!*param)
644 param = NULL;
645 }
646 }
Andrew Scullb4b6d4a2019-01-02 15:54:55 +0000647
648 trigger_ops = cmd_ops->get_trigger_ops(cmd, trigger);
649
650 ret = -ENOMEM;
651 trigger_data = kzalloc(sizeof(*trigger_data), GFP_KERNEL);
652 if (!trigger_data)
653 goto out;
654
655 trigger_data->count = -1;
656 trigger_data->ops = trigger_ops;
657 trigger_data->cmd_ops = cmd_ops;
658 trigger_data->private_data = file;
659 INIT_LIST_HEAD(&trigger_data->list);
660 INIT_LIST_HEAD(&trigger_data->named_list);
661
662 if (glob[0] == '!') {
663 cmd_ops->unreg(glob+1, trigger_ops, trigger_data, file);
664 kfree(trigger_data);
665 ret = 0;
666 goto out;
667 }
668
669 if (trigger) {
670 number = strsep(&trigger, ":");
671
672 ret = -EINVAL;
673 if (!strlen(number))
674 goto out_free;
675
676 /*
677 * We use the callback data field (which is a pointer)
678 * as our counter.
679 */
680 ret = kstrtoul(number, 0, &trigger_data->count);
681 if (ret)
682 goto out_free;
683 }
684
685 if (!param) /* if param is non-empty, it's supposed to be a filter */
686 goto out_reg;
687
688 if (!cmd_ops->set_filter)
689 goto out_reg;
690
691 ret = cmd_ops->set_filter(param, trigger_data, file);
692 if (ret < 0)
693 goto out_free;
694
695 out_reg:
696 /* Up the trigger_data count to make sure reg doesn't free it on failure */
697 event_trigger_init(trigger_ops, trigger_data);
698 ret = cmd_ops->reg(glob, trigger_ops, trigger_data, file);
699 /*
700 * The above returns on success the # of functions enabled,
701 * but if it didn't find any functions it returns zero.
702 * Consider no functions a failure too.
703 */
704 if (!ret) {
705 cmd_ops->unreg(glob, trigger_ops, trigger_data, file);
706 ret = -ENOENT;
707 } else if (ret > 0)
708 ret = 0;
709
710 /* Down the counter of trigger_data or free it if not used anymore */
711 event_trigger_free(trigger_ops, trigger_data);
712 out:
713 return ret;
714
715 out_free:
716 if (cmd_ops->set_filter)
717 cmd_ops->set_filter(NULL, trigger_data, NULL);
718 kfree(trigger_data);
719 goto out;
720}
721
722/**
723 * set_trigger_filter - Generic event_command @set_filter implementation
724 * @filter_str: The filter string for the trigger, NULL to remove filter
725 * @trigger_data: Trigger-specific data
726 * @file: The trace_event_file associated with the event
727 *
728 * Common implementation for event command filter parsing and filter
729 * instantiation.
730 *
731 * Usually used directly as the @set_filter method in event command
732 * implementations.
733 *
734 * Also used to remove a filter (if filter_str = NULL).
735 *
736 * Return: 0 on success, errno otherwise
737 */
738int set_trigger_filter(char *filter_str,
739 struct event_trigger_data *trigger_data,
740 struct trace_event_file *file)
741{
742 struct event_trigger_data *data = trigger_data;
743 struct event_filter *filter = NULL, *tmp;
744 int ret = -EINVAL;
745 char *s;
746
747 if (!filter_str) /* clear the current filter */
748 goto assign;
749
750 s = strsep(&filter_str, " \t");
751
752 if (!strlen(s) || strcmp(s, "if") != 0)
753 goto out;
754
755 if (!filter_str)
756 goto out;
757
758 /* The filter is for the 'trigger' event, not the triggered event */
David Brazdil0f672f62019-12-10 10:32:29 +0000759 ret = create_event_filter(file->tr, file->event_call,
760 filter_str, false, &filter);
Andrew Scullb4b6d4a2019-01-02 15:54:55 +0000761 /*
762 * If create_event_filter() fails, filter still needs to be freed.
763 * Which the calling code will do with data->filter.
764 */
765 assign:
766 tmp = rcu_access_pointer(data->filter);
767
768 rcu_assign_pointer(data->filter, filter);
769
770 if (tmp) {
771 /* Make sure the call is done with the filter */
772 tracepoint_synchronize_unregister();
773 free_event_filter(tmp);
774 }
775
776 kfree(data->filter_str);
777 data->filter_str = NULL;
778
779 if (filter_str) {
780 data->filter_str = kstrdup(filter_str, GFP_KERNEL);
781 if (!data->filter_str) {
782 free_event_filter(rcu_access_pointer(data->filter));
783 data->filter = NULL;
784 ret = -ENOMEM;
785 }
786 }
787 out:
788 return ret;
789}
790
791static LIST_HEAD(named_triggers);
792
793/**
794 * find_named_trigger - Find the common named trigger associated with @name
795 * @name: The name of the set of named triggers to find the common data for
796 *
797 * Named triggers are sets of triggers that share a common set of
798 * trigger data. The first named trigger registered with a given name
799 * owns the common trigger data that the others subsequently
800 * registered with the same name will reference. This function
801 * returns the common trigger data associated with that first
802 * registered instance.
803 *
804 * Return: the common trigger data for the given named trigger on
805 * success, NULL otherwise.
806 */
807struct event_trigger_data *find_named_trigger(const char *name)
808{
809 struct event_trigger_data *data;
810
811 if (!name)
812 return NULL;
813
814 list_for_each_entry(data, &named_triggers, named_list) {
815 if (data->named_data)
816 continue;
817 if (strcmp(data->name, name) == 0)
818 return data;
819 }
820
821 return NULL;
822}
823
824/**
825 * is_named_trigger - determine if a given trigger is a named trigger
826 * @test: The trigger data to test
827 *
828 * Return: true if 'test' is a named trigger, false otherwise.
829 */
830bool is_named_trigger(struct event_trigger_data *test)
831{
832 struct event_trigger_data *data;
833
834 list_for_each_entry(data, &named_triggers, named_list) {
835 if (test == data)
836 return true;
837 }
838
839 return false;
840}
841
842/**
843 * save_named_trigger - save the trigger in the named trigger list
844 * @name: The name of the named trigger set
845 * @data: The trigger data to save
846 *
847 * Return: 0 if successful, negative error otherwise.
848 */
849int save_named_trigger(const char *name, struct event_trigger_data *data)
850{
851 data->name = kstrdup(name, GFP_KERNEL);
852 if (!data->name)
853 return -ENOMEM;
854
855 list_add(&data->named_list, &named_triggers);
856
857 return 0;
858}
859
860/**
861 * del_named_trigger - delete a trigger from the named trigger list
862 * @data: The trigger data to delete
863 */
864void del_named_trigger(struct event_trigger_data *data)
865{
866 kfree(data->name);
867 data->name = NULL;
868
869 list_del(&data->named_list);
870}
871
872static void __pause_named_trigger(struct event_trigger_data *data, bool pause)
873{
874 struct event_trigger_data *test;
875
876 list_for_each_entry(test, &named_triggers, named_list) {
877 if (strcmp(test->name, data->name) == 0) {
878 if (pause) {
879 test->paused_tmp = test->paused;
880 test->paused = true;
881 } else {
882 test->paused = test->paused_tmp;
883 }
884 }
885 }
886}
887
888/**
889 * pause_named_trigger - Pause all named triggers with the same name
890 * @data: The trigger data of a named trigger to pause
891 *
892 * Pauses a named trigger along with all other triggers having the
893 * same name. Because named triggers share a common set of data,
894 * pausing only one is meaningless, so pausing one named trigger needs
895 * to pause all triggers with the same name.
896 */
897void pause_named_trigger(struct event_trigger_data *data)
898{
899 __pause_named_trigger(data, true);
900}
901
902/**
903 * unpause_named_trigger - Un-pause all named triggers with the same name
904 * @data: The trigger data of a named trigger to unpause
905 *
906 * Un-pauses a named trigger along with all other triggers having the
907 * same name. Because named triggers share a common set of data,
908 * unpausing only one is meaningless, so unpausing one named trigger
909 * needs to unpause all triggers with the same name.
910 */
911void unpause_named_trigger(struct event_trigger_data *data)
912{
913 __pause_named_trigger(data, false);
914}
915
916/**
917 * set_named_trigger_data - Associate common named trigger data
918 * @data: The trigger data of a named trigger to unpause
919 *
920 * Named triggers are sets of triggers that share a common set of
921 * trigger data. The first named trigger registered with a given name
922 * owns the common trigger data that the others subsequently
923 * registered with the same name will reference. This function
924 * associates the common trigger data from the first trigger with the
925 * given trigger.
926 */
927void set_named_trigger_data(struct event_trigger_data *data,
928 struct event_trigger_data *named_data)
929{
930 data->named_data = named_data;
931}
932
933struct event_trigger_data *
934get_named_trigger_data(struct event_trigger_data *data)
935{
936 return data->named_data;
937}
938
939static void
940traceon_trigger(struct event_trigger_data *data, void *rec,
941 struct ring_buffer_event *event)
942{
Olivier Deprez157378f2022-04-04 15:47:50 +0200943 struct trace_event_file *file = data->private_data;
944
945 if (file) {
946 if (tracer_tracing_is_on(file->tr))
947 return;
948
949 tracer_tracing_on(file->tr);
950 return;
951 }
952
Andrew Scullb4b6d4a2019-01-02 15:54:55 +0000953 if (tracing_is_on())
954 return;
955
956 tracing_on();
957}
958
959static void
960traceon_count_trigger(struct event_trigger_data *data, void *rec,
961 struct ring_buffer_event *event)
962{
Olivier Deprez157378f2022-04-04 15:47:50 +0200963 struct trace_event_file *file = data->private_data;
964
965 if (file) {
966 if (tracer_tracing_is_on(file->tr))
967 return;
968 } else {
969 if (tracing_is_on())
970 return;
971 }
Andrew Scullb4b6d4a2019-01-02 15:54:55 +0000972
973 if (!data->count)
974 return;
975
976 if (data->count != -1)
977 (data->count)--;
978
Olivier Deprez157378f2022-04-04 15:47:50 +0200979 if (file)
980 tracer_tracing_on(file->tr);
981 else
982 tracing_on();
Andrew Scullb4b6d4a2019-01-02 15:54:55 +0000983}
984
985static void
986traceoff_trigger(struct event_trigger_data *data, void *rec,
987 struct ring_buffer_event *event)
988{
Olivier Deprez157378f2022-04-04 15:47:50 +0200989 struct trace_event_file *file = data->private_data;
990
991 if (file) {
992 if (!tracer_tracing_is_on(file->tr))
993 return;
994
995 tracer_tracing_off(file->tr);
996 return;
997 }
998
Andrew Scullb4b6d4a2019-01-02 15:54:55 +0000999 if (!tracing_is_on())
1000 return;
1001
1002 tracing_off();
1003}
1004
1005static void
1006traceoff_count_trigger(struct event_trigger_data *data, void *rec,
1007 struct ring_buffer_event *event)
1008{
Olivier Deprez157378f2022-04-04 15:47:50 +02001009 struct trace_event_file *file = data->private_data;
1010
1011 if (file) {
1012 if (!tracer_tracing_is_on(file->tr))
1013 return;
1014 } else {
1015 if (!tracing_is_on())
1016 return;
1017 }
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00001018
1019 if (!data->count)
1020 return;
1021
1022 if (data->count != -1)
1023 (data->count)--;
1024
Olivier Deprez157378f2022-04-04 15:47:50 +02001025 if (file)
1026 tracer_tracing_off(file->tr);
1027 else
1028 tracing_off();
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00001029}
1030
1031static int
1032traceon_trigger_print(struct seq_file *m, struct event_trigger_ops *ops,
1033 struct event_trigger_data *data)
1034{
1035 return event_trigger_print("traceon", m, (void *)data->count,
1036 data->filter_str);
1037}
1038
1039static int
1040traceoff_trigger_print(struct seq_file *m, struct event_trigger_ops *ops,
1041 struct event_trigger_data *data)
1042{
1043 return event_trigger_print("traceoff", m, (void *)data->count,
1044 data->filter_str);
1045}
1046
1047static struct event_trigger_ops traceon_trigger_ops = {
1048 .func = traceon_trigger,
1049 .print = traceon_trigger_print,
1050 .init = event_trigger_init,
1051 .free = event_trigger_free,
1052};
1053
1054static struct event_trigger_ops traceon_count_trigger_ops = {
1055 .func = traceon_count_trigger,
1056 .print = traceon_trigger_print,
1057 .init = event_trigger_init,
1058 .free = event_trigger_free,
1059};
1060
1061static struct event_trigger_ops traceoff_trigger_ops = {
1062 .func = traceoff_trigger,
1063 .print = traceoff_trigger_print,
1064 .init = event_trigger_init,
1065 .free = event_trigger_free,
1066};
1067
1068static struct event_trigger_ops traceoff_count_trigger_ops = {
1069 .func = traceoff_count_trigger,
1070 .print = traceoff_trigger_print,
1071 .init = event_trigger_init,
1072 .free = event_trigger_free,
1073};
1074
1075static struct event_trigger_ops *
1076onoff_get_trigger_ops(char *cmd, char *param)
1077{
1078 struct event_trigger_ops *ops;
1079
1080 /* we register both traceon and traceoff to this callback */
1081 if (strcmp(cmd, "traceon") == 0)
1082 ops = param ? &traceon_count_trigger_ops :
1083 &traceon_trigger_ops;
1084 else
1085 ops = param ? &traceoff_count_trigger_ops :
1086 &traceoff_trigger_ops;
1087
1088 return ops;
1089}
1090
1091static struct event_command trigger_traceon_cmd = {
1092 .name = "traceon",
1093 .trigger_type = ETT_TRACE_ONOFF,
1094 .func = event_trigger_callback,
1095 .reg = register_trigger,
1096 .unreg = unregister_trigger,
1097 .get_trigger_ops = onoff_get_trigger_ops,
1098 .set_filter = set_trigger_filter,
1099};
1100
1101static struct event_command trigger_traceoff_cmd = {
1102 .name = "traceoff",
1103 .trigger_type = ETT_TRACE_ONOFF,
1104 .flags = EVENT_CMD_FL_POST_TRIGGER,
1105 .func = event_trigger_callback,
1106 .reg = register_trigger,
1107 .unreg = unregister_trigger,
1108 .get_trigger_ops = onoff_get_trigger_ops,
1109 .set_filter = set_trigger_filter,
1110};
1111
1112#ifdef CONFIG_TRACER_SNAPSHOT
1113static void
1114snapshot_trigger(struct event_trigger_data *data, void *rec,
1115 struct ring_buffer_event *event)
1116{
1117 struct trace_event_file *file = data->private_data;
1118
1119 if (file)
1120 tracing_snapshot_instance(file->tr);
1121 else
1122 tracing_snapshot();
1123}
1124
1125static void
1126snapshot_count_trigger(struct event_trigger_data *data, void *rec,
1127 struct ring_buffer_event *event)
1128{
1129 if (!data->count)
1130 return;
1131
1132 if (data->count != -1)
1133 (data->count)--;
1134
1135 snapshot_trigger(data, rec, event);
1136}
1137
1138static int
1139register_snapshot_trigger(char *glob, struct event_trigger_ops *ops,
1140 struct event_trigger_data *data,
1141 struct trace_event_file *file)
1142{
Olivier Deprez0e641232021-09-23 10:07:05 +02001143 if (tracing_alloc_snapshot_instance(file->tr) != 0)
1144 return 0;
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00001145
Olivier Deprez0e641232021-09-23 10:07:05 +02001146 return register_trigger(glob, ops, data, file);
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00001147}
1148
1149static int
1150snapshot_trigger_print(struct seq_file *m, struct event_trigger_ops *ops,
1151 struct event_trigger_data *data)
1152{
1153 return event_trigger_print("snapshot", m, (void *)data->count,
1154 data->filter_str);
1155}
1156
1157static struct event_trigger_ops snapshot_trigger_ops = {
1158 .func = snapshot_trigger,
1159 .print = snapshot_trigger_print,
1160 .init = event_trigger_init,
1161 .free = event_trigger_free,
1162};
1163
1164static struct event_trigger_ops snapshot_count_trigger_ops = {
1165 .func = snapshot_count_trigger,
1166 .print = snapshot_trigger_print,
1167 .init = event_trigger_init,
1168 .free = event_trigger_free,
1169};
1170
1171static struct event_trigger_ops *
1172snapshot_get_trigger_ops(char *cmd, char *param)
1173{
1174 return param ? &snapshot_count_trigger_ops : &snapshot_trigger_ops;
1175}
1176
1177static struct event_command trigger_snapshot_cmd = {
1178 .name = "snapshot",
1179 .trigger_type = ETT_SNAPSHOT,
1180 .func = event_trigger_callback,
1181 .reg = register_snapshot_trigger,
1182 .unreg = unregister_trigger,
1183 .get_trigger_ops = snapshot_get_trigger_ops,
1184 .set_filter = set_trigger_filter,
1185};
1186
1187static __init int register_trigger_snapshot_cmd(void)
1188{
1189 int ret;
1190
1191 ret = register_event_command(&trigger_snapshot_cmd);
1192 WARN_ON(ret < 0);
1193
1194 return ret;
1195}
1196#else
1197static __init int register_trigger_snapshot_cmd(void) { return 0; }
1198#endif /* CONFIG_TRACER_SNAPSHOT */
1199
1200#ifdef CONFIG_STACKTRACE
1201#ifdef CONFIG_UNWINDER_ORC
1202/* Skip 2:
1203 * event_triggers_post_call()
1204 * trace_event_raw_event_xxx()
1205 */
1206# define STACK_SKIP 2
1207#else
1208/*
1209 * Skip 4:
1210 * stacktrace_trigger()
1211 * event_triggers_post_call()
1212 * trace_event_buffer_commit()
1213 * trace_event_raw_event_xxx()
1214 */
1215#define STACK_SKIP 4
1216#endif
1217
1218static void
1219stacktrace_trigger(struct event_trigger_data *data, void *rec,
1220 struct ring_buffer_event *event)
1221{
1222 trace_dump_stack(STACK_SKIP);
1223}
1224
1225static void
1226stacktrace_count_trigger(struct event_trigger_data *data, void *rec,
1227 struct ring_buffer_event *event)
1228{
1229 if (!data->count)
1230 return;
1231
1232 if (data->count != -1)
1233 (data->count)--;
1234
1235 stacktrace_trigger(data, rec, event);
1236}
1237
1238static int
1239stacktrace_trigger_print(struct seq_file *m, struct event_trigger_ops *ops,
1240 struct event_trigger_data *data)
1241{
1242 return event_trigger_print("stacktrace", m, (void *)data->count,
1243 data->filter_str);
1244}
1245
1246static struct event_trigger_ops stacktrace_trigger_ops = {
1247 .func = stacktrace_trigger,
1248 .print = stacktrace_trigger_print,
1249 .init = event_trigger_init,
1250 .free = event_trigger_free,
1251};
1252
1253static struct event_trigger_ops stacktrace_count_trigger_ops = {
1254 .func = stacktrace_count_trigger,
1255 .print = stacktrace_trigger_print,
1256 .init = event_trigger_init,
1257 .free = event_trigger_free,
1258};
1259
1260static struct event_trigger_ops *
1261stacktrace_get_trigger_ops(char *cmd, char *param)
1262{
1263 return param ? &stacktrace_count_trigger_ops : &stacktrace_trigger_ops;
1264}
1265
1266static struct event_command trigger_stacktrace_cmd = {
1267 .name = "stacktrace",
1268 .trigger_type = ETT_STACKTRACE,
1269 .flags = EVENT_CMD_FL_POST_TRIGGER,
1270 .func = event_trigger_callback,
1271 .reg = register_trigger,
1272 .unreg = unregister_trigger,
1273 .get_trigger_ops = stacktrace_get_trigger_ops,
1274 .set_filter = set_trigger_filter,
1275};
1276
1277static __init int register_trigger_stacktrace_cmd(void)
1278{
1279 int ret;
1280
1281 ret = register_event_command(&trigger_stacktrace_cmd);
1282 WARN_ON(ret < 0);
1283
1284 return ret;
1285}
1286#else
1287static __init int register_trigger_stacktrace_cmd(void) { return 0; }
1288#endif /* CONFIG_STACKTRACE */
1289
1290static __init void unregister_trigger_traceon_traceoff_cmds(void)
1291{
1292 unregister_event_command(&trigger_traceon_cmd);
1293 unregister_event_command(&trigger_traceoff_cmd);
1294}
1295
1296static void
1297event_enable_trigger(struct event_trigger_data *data, void *rec,
1298 struct ring_buffer_event *event)
1299{
1300 struct enable_trigger_data *enable_data = data->private_data;
1301
1302 if (enable_data->enable)
1303 clear_bit(EVENT_FILE_FL_SOFT_DISABLED_BIT, &enable_data->file->flags);
1304 else
1305 set_bit(EVENT_FILE_FL_SOFT_DISABLED_BIT, &enable_data->file->flags);
1306}
1307
1308static void
1309event_enable_count_trigger(struct event_trigger_data *data, void *rec,
1310 struct ring_buffer_event *event)
1311{
1312 struct enable_trigger_data *enable_data = data->private_data;
1313
1314 if (!data->count)
1315 return;
1316
1317 /* Skip if the event is in a state we want to switch to */
1318 if (enable_data->enable == !(enable_data->file->flags & EVENT_FILE_FL_SOFT_DISABLED))
1319 return;
1320
1321 if (data->count != -1)
1322 (data->count)--;
1323
1324 event_enable_trigger(data, rec, event);
1325}
1326
1327int event_enable_trigger_print(struct seq_file *m,
1328 struct event_trigger_ops *ops,
1329 struct event_trigger_data *data)
1330{
1331 struct enable_trigger_data *enable_data = data->private_data;
1332
1333 seq_printf(m, "%s:%s:%s",
1334 enable_data->hist ?
1335 (enable_data->enable ? ENABLE_HIST_STR : DISABLE_HIST_STR) :
1336 (enable_data->enable ? ENABLE_EVENT_STR : DISABLE_EVENT_STR),
1337 enable_data->file->event_call->class->system,
1338 trace_event_name(enable_data->file->event_call));
1339
1340 if (data->count == -1)
1341 seq_puts(m, ":unlimited");
1342 else
1343 seq_printf(m, ":count=%ld", data->count);
1344
1345 if (data->filter_str)
1346 seq_printf(m, " if %s\n", data->filter_str);
1347 else
1348 seq_putc(m, '\n');
1349
1350 return 0;
1351}
1352
1353void event_enable_trigger_free(struct event_trigger_ops *ops,
1354 struct event_trigger_data *data)
1355{
1356 struct enable_trigger_data *enable_data = data->private_data;
1357
1358 if (WARN_ON_ONCE(data->ref <= 0))
1359 return;
1360
1361 data->ref--;
1362 if (!data->ref) {
1363 /* Remove the SOFT_MODE flag */
1364 trace_event_enable_disable(enable_data->file, 0, 1);
1365 module_put(enable_data->file->event_call->mod);
1366 trigger_data_free(data);
1367 kfree(enable_data);
1368 }
1369}
1370
1371static struct event_trigger_ops event_enable_trigger_ops = {
1372 .func = event_enable_trigger,
1373 .print = event_enable_trigger_print,
1374 .init = event_trigger_init,
1375 .free = event_enable_trigger_free,
1376};
1377
1378static struct event_trigger_ops event_enable_count_trigger_ops = {
1379 .func = event_enable_count_trigger,
1380 .print = event_enable_trigger_print,
1381 .init = event_trigger_init,
1382 .free = event_enable_trigger_free,
1383};
1384
1385static struct event_trigger_ops event_disable_trigger_ops = {
1386 .func = event_enable_trigger,
1387 .print = event_enable_trigger_print,
1388 .init = event_trigger_init,
1389 .free = event_enable_trigger_free,
1390};
1391
1392static struct event_trigger_ops event_disable_count_trigger_ops = {
1393 .func = event_enable_count_trigger,
1394 .print = event_enable_trigger_print,
1395 .init = event_trigger_init,
1396 .free = event_enable_trigger_free,
1397};
1398
1399int event_enable_trigger_func(struct event_command *cmd_ops,
1400 struct trace_event_file *file,
1401 char *glob, char *cmd, char *param)
1402{
1403 struct trace_event_file *event_enable_file;
1404 struct enable_trigger_data *enable_data;
1405 struct event_trigger_data *trigger_data;
1406 struct event_trigger_ops *trigger_ops;
1407 struct trace_array *tr = file->tr;
1408 const char *system;
1409 const char *event;
1410 bool hist = false;
1411 char *trigger;
1412 char *number;
1413 bool enable;
1414 int ret;
1415
1416 if (!param)
1417 return -EINVAL;
1418
1419 /* separate the trigger from the filter (s:e:n [if filter]) */
1420 trigger = strsep(&param, " \t");
1421 if (!trigger)
1422 return -EINVAL;
Olivier Deprez0e641232021-09-23 10:07:05 +02001423 if (param) {
1424 param = skip_spaces(param);
1425 if (!*param)
1426 param = NULL;
1427 }
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00001428
1429 system = strsep(&trigger, ":");
1430 if (!trigger)
1431 return -EINVAL;
1432
1433 event = strsep(&trigger, ":");
1434
1435 ret = -EINVAL;
1436 event_enable_file = find_event_file(tr, system, event);
1437 if (!event_enable_file)
1438 goto out;
1439
1440#ifdef CONFIG_HIST_TRIGGERS
1441 hist = ((strcmp(cmd, ENABLE_HIST_STR) == 0) ||
1442 (strcmp(cmd, DISABLE_HIST_STR) == 0));
1443
1444 enable = ((strcmp(cmd, ENABLE_EVENT_STR) == 0) ||
1445 (strcmp(cmd, ENABLE_HIST_STR) == 0));
1446#else
1447 enable = strcmp(cmd, ENABLE_EVENT_STR) == 0;
1448#endif
1449 trigger_ops = cmd_ops->get_trigger_ops(cmd, trigger);
1450
1451 ret = -ENOMEM;
1452 trigger_data = kzalloc(sizeof(*trigger_data), GFP_KERNEL);
1453 if (!trigger_data)
1454 goto out;
1455
1456 enable_data = kzalloc(sizeof(*enable_data), GFP_KERNEL);
1457 if (!enable_data) {
1458 kfree(trigger_data);
1459 goto out;
1460 }
1461
1462 trigger_data->count = -1;
1463 trigger_data->ops = trigger_ops;
1464 trigger_data->cmd_ops = cmd_ops;
1465 INIT_LIST_HEAD(&trigger_data->list);
1466 RCU_INIT_POINTER(trigger_data->filter, NULL);
1467
1468 enable_data->hist = hist;
1469 enable_data->enable = enable;
1470 enable_data->file = event_enable_file;
1471 trigger_data->private_data = enable_data;
1472
1473 if (glob[0] == '!') {
1474 cmd_ops->unreg(glob+1, trigger_ops, trigger_data, file);
1475 kfree(trigger_data);
1476 kfree(enable_data);
1477 ret = 0;
1478 goto out;
1479 }
1480
1481 /* Up the trigger_data count to make sure nothing frees it on failure */
1482 event_trigger_init(trigger_ops, trigger_data);
1483
1484 if (trigger) {
1485 number = strsep(&trigger, ":");
1486
1487 ret = -EINVAL;
1488 if (!strlen(number))
1489 goto out_free;
1490
1491 /*
1492 * We use the callback data field (which is a pointer)
1493 * as our counter.
1494 */
1495 ret = kstrtoul(number, 0, &trigger_data->count);
1496 if (ret)
1497 goto out_free;
1498 }
1499
1500 if (!param) /* if param is non-empty, it's supposed to be a filter */
1501 goto out_reg;
1502
1503 if (!cmd_ops->set_filter)
1504 goto out_reg;
1505
1506 ret = cmd_ops->set_filter(param, trigger_data, file);
1507 if (ret < 0)
1508 goto out_free;
1509
1510 out_reg:
1511 /* Don't let event modules unload while probe registered */
1512 ret = try_module_get(event_enable_file->event_call->mod);
1513 if (!ret) {
1514 ret = -EBUSY;
1515 goto out_free;
1516 }
1517
1518 ret = trace_event_enable_disable(event_enable_file, 1, 1);
1519 if (ret < 0)
1520 goto out_put;
1521 ret = cmd_ops->reg(glob, trigger_ops, trigger_data, file);
1522 /*
1523 * The above returns on success the # of functions enabled,
1524 * but if it didn't find any functions it returns zero.
1525 * Consider no functions a failure too.
1526 */
1527 if (!ret) {
1528 ret = -ENOENT;
1529 goto out_disable;
1530 } else if (ret < 0)
1531 goto out_disable;
1532 /* Just return zero, not the number of enabled functions */
1533 ret = 0;
1534 event_trigger_free(trigger_ops, trigger_data);
1535 out:
1536 return ret;
1537
1538 out_disable:
1539 trace_event_enable_disable(event_enable_file, 0, 1);
1540 out_put:
1541 module_put(event_enable_file->event_call->mod);
1542 out_free:
1543 if (cmd_ops->set_filter)
1544 cmd_ops->set_filter(NULL, trigger_data, NULL);
1545 event_trigger_free(trigger_ops, trigger_data);
1546 kfree(enable_data);
1547 goto out;
1548}
1549
1550int event_enable_register_trigger(char *glob,
1551 struct event_trigger_ops *ops,
1552 struct event_trigger_data *data,
1553 struct trace_event_file *file)
1554{
1555 struct enable_trigger_data *enable_data = data->private_data;
1556 struct enable_trigger_data *test_enable_data;
1557 struct event_trigger_data *test;
1558 int ret = 0;
1559
Olivier Deprez0e641232021-09-23 10:07:05 +02001560 lockdep_assert_held(&event_mutex);
1561
1562 list_for_each_entry(test, &file->triggers, list) {
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00001563 test_enable_data = test->private_data;
1564 if (test_enable_data &&
1565 (test->cmd_ops->trigger_type ==
1566 data->cmd_ops->trigger_type) &&
1567 (test_enable_data->file == enable_data->file)) {
1568 ret = -EEXIST;
1569 goto out;
1570 }
1571 }
1572
1573 if (data->ops->init) {
1574 ret = data->ops->init(data->ops, data);
1575 if (ret < 0)
1576 goto out;
1577 }
1578
1579 list_add_rcu(&data->list, &file->triggers);
1580 ret++;
1581
1582 update_cond_flag(file);
1583 if (trace_event_trigger_enable_disable(file, 1) < 0) {
1584 list_del_rcu(&data->list);
1585 update_cond_flag(file);
1586 ret--;
1587 }
1588out:
1589 return ret;
1590}
1591
1592void event_enable_unregister_trigger(char *glob,
1593 struct event_trigger_ops *ops,
1594 struct event_trigger_data *test,
1595 struct trace_event_file *file)
1596{
1597 struct enable_trigger_data *test_enable_data = test->private_data;
1598 struct enable_trigger_data *enable_data;
1599 struct event_trigger_data *data;
1600 bool unregistered = false;
1601
Olivier Deprez0e641232021-09-23 10:07:05 +02001602 lockdep_assert_held(&event_mutex);
1603
1604 list_for_each_entry(data, &file->triggers, list) {
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00001605 enable_data = data->private_data;
1606 if (enable_data &&
1607 (data->cmd_ops->trigger_type ==
1608 test->cmd_ops->trigger_type) &&
1609 (enable_data->file == test_enable_data->file)) {
1610 unregistered = true;
1611 list_del_rcu(&data->list);
1612 trace_event_trigger_enable_disable(file, 0);
1613 update_cond_flag(file);
1614 break;
1615 }
1616 }
1617
1618 if (unregistered && data->ops->free)
1619 data->ops->free(data->ops, data);
1620}
1621
1622static struct event_trigger_ops *
1623event_enable_get_trigger_ops(char *cmd, char *param)
1624{
1625 struct event_trigger_ops *ops;
1626 bool enable;
1627
1628#ifdef CONFIG_HIST_TRIGGERS
1629 enable = ((strcmp(cmd, ENABLE_EVENT_STR) == 0) ||
1630 (strcmp(cmd, ENABLE_HIST_STR) == 0));
1631#else
1632 enable = strcmp(cmd, ENABLE_EVENT_STR) == 0;
1633#endif
1634 if (enable)
1635 ops = param ? &event_enable_count_trigger_ops :
1636 &event_enable_trigger_ops;
1637 else
1638 ops = param ? &event_disable_count_trigger_ops :
1639 &event_disable_trigger_ops;
1640
1641 return ops;
1642}
1643
1644static struct event_command trigger_enable_cmd = {
1645 .name = ENABLE_EVENT_STR,
1646 .trigger_type = ETT_EVENT_ENABLE,
1647 .func = event_enable_trigger_func,
1648 .reg = event_enable_register_trigger,
1649 .unreg = event_enable_unregister_trigger,
1650 .get_trigger_ops = event_enable_get_trigger_ops,
1651 .set_filter = set_trigger_filter,
1652};
1653
1654static struct event_command trigger_disable_cmd = {
1655 .name = DISABLE_EVENT_STR,
1656 .trigger_type = ETT_EVENT_ENABLE,
1657 .func = event_enable_trigger_func,
1658 .reg = event_enable_register_trigger,
1659 .unreg = event_enable_unregister_trigger,
1660 .get_trigger_ops = event_enable_get_trigger_ops,
1661 .set_filter = set_trigger_filter,
1662};
1663
1664static __init void unregister_trigger_enable_disable_cmds(void)
1665{
1666 unregister_event_command(&trigger_enable_cmd);
1667 unregister_event_command(&trigger_disable_cmd);
1668}
1669
1670static __init int register_trigger_enable_disable_cmds(void)
1671{
1672 int ret;
1673
1674 ret = register_event_command(&trigger_enable_cmd);
1675 if (WARN_ON(ret < 0))
1676 return ret;
1677 ret = register_event_command(&trigger_disable_cmd);
1678 if (WARN_ON(ret < 0))
1679 unregister_trigger_enable_disable_cmds();
1680
1681 return ret;
1682}
1683
1684static __init int register_trigger_traceon_traceoff_cmds(void)
1685{
1686 int ret;
1687
1688 ret = register_event_command(&trigger_traceon_cmd);
1689 if (WARN_ON(ret < 0))
1690 return ret;
1691 ret = register_event_command(&trigger_traceoff_cmd);
1692 if (WARN_ON(ret < 0))
1693 unregister_trigger_traceon_traceoff_cmds();
1694
1695 return ret;
1696}
1697
1698__init int register_trigger_cmds(void)
1699{
1700 register_trigger_traceon_traceoff_cmds();
1701 register_trigger_snapshot_cmd();
1702 register_trigger_stacktrace_cmd();
1703 register_trigger_enable_disable_cmds();
1704 register_trigger_hist_enable_disable_cmds();
1705 register_trigger_hist_cmd();
1706
1707 return 0;
1708}