Lines Matching defs:op
368 struct optimized_kprobe *op;
370 op = container_of(p, struct optimized_kprobe, kp);
371 arch_remove_optimized_kprobe(op);
373 kfree(op);
379 struct optimized_kprobe *op;
382 op = container_of(p, struct optimized_kprobe, kp);
383 return arch_prepared_optinsn(&op->optinsn);
392 struct optimized_kprobe *op;
398 op = container_of(p, struct optimized_kprobe, kp);
400 return kprobe_disabled(p) && list_empty(&op->list);
406 struct optimized_kprobe *op;
409 op = container_of(p, struct optimized_kprobe, kp);
410 if (!list_empty(&op->list))
424 struct optimized_kprobe *op;
431 op = container_of(p, struct optimized_kprobe, kp);
432 if (arch_within_optimized_kprobe(op, addr))
482 struct optimized_kprobe *op, *tmp;
493 list_for_each_entry_safe(op, tmp, &freeing_list, list) {
495 if (kprobe_disabled(&op->kp))
496 arch_disarm_kprobe(&op->kp);
497 if (kprobe_unused(&op->kp)) {
503 hlist_del_rcu(&op->kp.hlist);
505 list_del_init(&op->list);
514 struct optimized_kprobe *op, *tmp;
516 list_for_each_entry_safe(op, tmp, &freeing_list, list) {
517 BUG_ON(!kprobe_unused(&op->kp));
518 list_del_init(&op->list);
519 free_aggr_kprobe(&op->kp);
587 struct optimized_kprobe *op;
598 op = container_of(p, struct optimized_kprobe, kp);
601 if (arch_check_optimized_kprobe(op) < 0)
605 if (op->kp.flags & KPROBE_FLAG_OPTIMIZED)
607 op->kp.flags |= KPROBE_FLAG_OPTIMIZED;
609 if (!list_empty(&op->list))
611 list_del_init(&op->list);
613 list_add(&op->list, &optimizing_list);
619 static void force_unoptimize_kprobe(struct optimized_kprobe *op)
622 arch_unoptimize_kprobe(op);
624 if (kprobe_disabled(&op->kp))
625 arch_disarm_kprobe(&op->kp);
631 struct optimized_kprobe *op;
636 op = container_of(p, struct optimized_kprobe, kp);
639 if (force && !list_empty(&op->list)) {
645 list_del_init(&op->list);
646 force_unoptimize_kprobe(op);
651 op->kp.flags &= ~KPROBE_FLAG_OPTIMIZED;
652 if (!list_empty(&op->list)) {
654 list_del_init(&op->list);
660 force_unoptimize_kprobe(op);
662 list_add(&op->list, &unoptimizing_list);
670 struct optimized_kprobe *op;
677 op = container_of(ap, struct optimized_kprobe, kp);
678 if (unlikely(list_empty(&op->list)))
683 /* Optimize it again (remove from op->list) */
691 struct optimized_kprobe *op;
693 op = container_of(p, struct optimized_kprobe, kp);
694 if (!list_empty(&op->list))
696 list_del_init(&op->list);
697 op->kp.flags &= ~KPROBE_FLAG_OPTIMIZED;
701 list_add(&op->list, &freeing_list);
707 hlist_del_rcu(&op->kp.hlist);
711 arch_remove_optimized_kprobe(op);
717 struct optimized_kprobe *op;
719 op = container_of(p, struct optimized_kprobe, kp);
720 arch_prepare_optimized_kprobe(op, p);
726 struct optimized_kprobe *op;
728 op = kzalloc(sizeof(struct optimized_kprobe), GFP_KERNEL);
729 if (!op)
732 INIT_LIST_HEAD(&op->list);
733 op->kp.addr = p->addr;
734 arch_prepare_optimized_kprobe(op, p);
736 return &op->kp;
748 struct optimized_kprobe *op;
762 op = container_of(ap, struct optimized_kprobe, kp);
763 if (!arch_prepared_optinsn(&op->optinsn)) {
765 arch_remove_optimized_kprobe(op);
766 kfree(op);