1
2
3
4
5
6
7
8
9
10
11
12#include <linux/kernel.h>
13#include <linux/sched/signal.h>
14#include <linux/sched/task_stack.h>
15#include <linux/mm.h>
16#include <linux/elf.h>
17#include <linux/smp.h>
18#include <linux/ptrace.h>
19#include <linux/user.h>
20#include <linux/security.h>
21#include <linux/init.h>
22#include <linux/signal.h>
23#include <linux/uaccess.h>
24#include <linux/perf_event.h>
25#include <linux/hw_breakpoint.h>
26#include <linux/regset.h>
27#include <linux/audit.h>
28#include <linux/tracehook.h>
29#include <linux/unistd.h>
30
31#include <asm/pgtable.h>
32#include <asm/traps.h>
33
34#define CREATE_TRACE_POINTS
35#include <trace/events/syscalls.h>
36
37#define REG_PC 15
38#define REG_PSR 16
39
40
41
42
43
44#if 0
45
46
47
48#define BREAKINST_ARM 0xef9f0001
49#define BREAKINST_THUMB 0xdf00
50#else
51
52
53
54
55
56
57
58
59#define BREAKINST_ARM 0xe7f001f0
60#define BREAKINST_THUMB 0xde01
61#endif
62
63struct pt_regs_offset {
64 const char *name;
65 int offset;
66};
67
68#define REG_OFFSET_NAME(r) \
69 {.name = #r, .offset = offsetof(struct pt_regs, ARM_##r)}
70#define REG_OFFSET_END {.name = NULL, .offset = 0}
71
72static const struct pt_regs_offset regoffset_table[] = {
73 REG_OFFSET_NAME(r0),
74 REG_OFFSET_NAME(r1),
75 REG_OFFSET_NAME(r2),
76 REG_OFFSET_NAME(r3),
77 REG_OFFSET_NAME(r4),
78 REG_OFFSET_NAME(r5),
79 REG_OFFSET_NAME(r6),
80 REG_OFFSET_NAME(r7),
81 REG_OFFSET_NAME(r8),
82 REG_OFFSET_NAME(r9),
83 REG_OFFSET_NAME(r10),
84 REG_OFFSET_NAME(fp),
85 REG_OFFSET_NAME(ip),
86 REG_OFFSET_NAME(sp),
87 REG_OFFSET_NAME(lr),
88 REG_OFFSET_NAME(pc),
89 REG_OFFSET_NAME(cpsr),
90 REG_OFFSET_NAME(ORIG_r0),
91 REG_OFFSET_END,
92};
93
94
95
96
97
98
99
100
101int regs_query_register_offset(const char *name)
102{
103 const struct pt_regs_offset *roff;
104 for (roff = regoffset_table; roff->name != NULL; roff++)
105 if (!strcmp(roff->name, name))
106 return roff->offset;
107 return -EINVAL;
108}
109
110
111
112
113
114
115
116
117const char *regs_query_register_name(unsigned int offset)
118{
119 const struct pt_regs_offset *roff;
120 for (roff = regoffset_table; roff->name != NULL; roff++)
121 if (roff->offset == offset)
122 return roff->name;
123 return NULL;
124}
125
126
127
128
129
130
131
132
133
134bool regs_within_kernel_stack(struct pt_regs *regs, unsigned long addr)
135{
136 return ((addr & ~(THREAD_SIZE - 1)) ==
137 (kernel_stack_pointer(regs) & ~(THREAD_SIZE - 1)));
138}
139
140
141
142
143
144
145
146
147
148
149unsigned long regs_get_kernel_stack_nth(struct pt_regs *regs, unsigned int n)
150{
151 unsigned long *addr = (unsigned long *)kernel_stack_pointer(regs);
152 addr += n;
153 if (regs_within_kernel_stack(regs, (unsigned long)addr))
154 return *addr;
155 else
156 return 0;
157}
158
159
160
161
162
163
164
165static inline long get_user_reg(struct task_struct *task, int offset)
166{
167 return task_pt_regs(task)->uregs[offset];
168}
169
170
171
172
173
174
175
176static inline int
177put_user_reg(struct task_struct *task, int offset, long data)
178{
179 struct pt_regs newregs, *regs = task_pt_regs(task);
180 int ret = -EINVAL;
181
182 newregs = *regs;
183 newregs.uregs[offset] = data;
184
185 if (valid_user_regs(&newregs)) {
186 regs->uregs[offset] = data;
187 ret = 0;
188 }
189
190 return ret;
191}
192
193
194
195
196void ptrace_disable(struct task_struct *child)
197{
198
199}
200
201
202
203
204void ptrace_break(struct task_struct *tsk, struct pt_regs *regs)
205{
206 force_sig_fault(SIGTRAP, TRAP_BRKPT,
207 (void __user *)instruction_pointer(regs), tsk);
208}
209
210static int break_trap(struct pt_regs *regs, unsigned int instr)
211{
212 ptrace_break(current, regs);
213 return 0;
214}
215
216static struct undef_hook arm_break_hook = {
217 .instr_mask = 0x0fffffff,
218 .instr_val = 0x07f001f0,
219 .cpsr_mask = PSR_T_BIT,
220 .cpsr_val = 0,
221 .fn = break_trap,
222};
223
224static struct undef_hook thumb_break_hook = {
225 .instr_mask = 0xffff,
226 .instr_val = 0xde01,
227 .cpsr_mask = PSR_T_BIT,
228 .cpsr_val = PSR_T_BIT,
229 .fn = break_trap,
230};
231
232static struct undef_hook thumb2_break_hook = {
233 .instr_mask = 0xffffffff,
234 .instr_val = 0xf7f0a000,
235 .cpsr_mask = PSR_T_BIT,
236 .cpsr_val = PSR_T_BIT,
237 .fn = break_trap,
238};
239
240static int __init ptrace_break_init(void)
241{
242 register_undef_hook(&arm_break_hook);
243 register_undef_hook(&thumb_break_hook);
244 register_undef_hook(&thumb2_break_hook);
245 return 0;
246}
247
248core_initcall(ptrace_break_init);
249
250
251
252
253
254static int ptrace_read_user(struct task_struct *tsk, unsigned long off,
255 unsigned long __user *ret)
256{
257 unsigned long tmp;
258
259 if (off & 3)
260 return -EIO;
261
262 tmp = 0;
263 if (off == PT_TEXT_ADDR)
264 tmp = tsk->mm->start_code;
265 else if (off == PT_DATA_ADDR)
266 tmp = tsk->mm->start_data;
267 else if (off == PT_TEXT_END_ADDR)
268 tmp = tsk->mm->end_code;
269 else if (off < sizeof(struct pt_regs))
270 tmp = get_user_reg(tsk, off >> 2);
271 else if (off >= sizeof(struct user))
272 return -EIO;
273
274 return put_user(tmp, ret);
275}
276
277
278
279
280
281static int ptrace_write_user(struct task_struct *tsk, unsigned long off,
282 unsigned long val)
283{
284 if (off & 3 || off >= sizeof(struct user))
285 return -EIO;
286
287 if (off >= sizeof(struct pt_regs))
288 return 0;
289
290 return put_user_reg(tsk, off >> 2, val);
291}
292
293#ifdef CONFIG_IWMMXT
294
295
296
297
298static int ptrace_getwmmxregs(struct task_struct *tsk, void __user *ufp)
299{
300 struct thread_info *thread = task_thread_info(tsk);
301
302 if (!test_ti_thread_flag(thread, TIF_USING_IWMMXT))
303 return -ENODATA;
304 iwmmxt_task_disable(thread);
305 return copy_to_user(ufp, &thread->fpstate.iwmmxt, IWMMXT_SIZE)
306 ? -EFAULT : 0;
307}
308
309
310
311
312static int ptrace_setwmmxregs(struct task_struct *tsk, void __user *ufp)
313{
314 struct thread_info *thread = task_thread_info(tsk);
315
316 if (!test_ti_thread_flag(thread, TIF_USING_IWMMXT))
317 return -EACCES;
318 iwmmxt_task_release(thread);
319 return copy_from_user(&thread->fpstate.iwmmxt, ufp, IWMMXT_SIZE)
320 ? -EFAULT : 0;
321}
322
323#endif
324
325#ifdef CONFIG_CRUNCH
326
327
328
329static int ptrace_getcrunchregs(struct task_struct *tsk, void __user *ufp)
330{
331 struct thread_info *thread = task_thread_info(tsk);
332
333 crunch_task_disable(thread);
334 return copy_to_user(ufp, &thread->crunchstate, CRUNCH_SIZE)
335 ? -EFAULT : 0;
336}
337
338
339
340
341static int ptrace_setcrunchregs(struct task_struct *tsk, void __user *ufp)
342{
343 struct thread_info *thread = task_thread_info(tsk);
344
345 crunch_task_release(thread);
346 return copy_from_user(&thread->crunchstate, ufp, CRUNCH_SIZE)
347 ? -EFAULT : 0;
348}
349#endif
350
351#ifdef CONFIG_HAVE_HW_BREAKPOINT
352
353
354
355
356
357
358
359static int ptrace_hbp_num_to_idx(long num)
360{
361 if (num < 0)
362 num = (ARM_MAX_BRP << 1) - num;
363 return (num - 1) >> 1;
364}
365
366
367
368
369
370static long ptrace_hbp_idx_to_num(int idx)
371{
372 long mid = ARM_MAX_BRP << 1;
373 long num = (idx << 1) + 1;
374 return num > mid ? mid - num : num;
375}
376
377
378
379
380static void ptrace_hbptriggered(struct perf_event *bp,
381 struct perf_sample_data *data,
382 struct pt_regs *regs)
383{
384 struct arch_hw_breakpoint *bkpt = counter_arch_bp(bp);
385 long num;
386 int i;
387
388 for (i = 0; i < ARM_MAX_HBP_SLOTS; ++i)
389 if (current->thread.debug.hbp[i] == bp)
390 break;
391
392 num = (i == ARM_MAX_HBP_SLOTS) ? 0 : ptrace_hbp_idx_to_num(i);
393
394 force_sig_ptrace_errno_trap((int)num, (void __user *)(bkpt->trigger));
395}
396
397
398
399
400
401
402void clear_ptrace_hw_breakpoint(struct task_struct *tsk)
403{
404 memset(tsk->thread.debug.hbp, 0, sizeof(tsk->thread.debug.hbp));
405}
406
407
408
409
410
411void flush_ptrace_hw_breakpoint(struct task_struct *tsk)
412{
413 int i;
414 struct thread_struct *t = &tsk->thread;
415
416 for (i = 0; i < ARM_MAX_HBP_SLOTS; i++) {
417 if (t->debug.hbp[i]) {
418 unregister_hw_breakpoint(t->debug.hbp[i]);
419 t->debug.hbp[i] = NULL;
420 }
421 }
422}
423
424static u32 ptrace_get_hbp_resource_info(void)
425{
426 u8 num_brps, num_wrps, debug_arch, wp_len;
427 u32 reg = 0;
428
429 num_brps = hw_breakpoint_slots(TYPE_INST);
430 num_wrps = hw_breakpoint_slots(TYPE_DATA);
431 debug_arch = arch_get_debug_arch();
432 wp_len = arch_get_max_wp_len();
433
434 reg |= debug_arch;
435 reg <<= 8;
436 reg |= wp_len;
437 reg <<= 8;
438 reg |= num_wrps;
439 reg <<= 8;
440 reg |= num_brps;
441
442 return reg;
443}
444
445static struct perf_event *ptrace_hbp_create(struct task_struct *tsk, int type)
446{
447 struct perf_event_attr attr;
448
449 ptrace_breakpoint_init(&attr);
450
451
452 attr.bp_addr = 0;
453 attr.bp_len = HW_BREAKPOINT_LEN_4;
454 attr.bp_type = type;
455 attr.disabled = 1;
456
457 return register_user_hw_breakpoint(&attr, ptrace_hbptriggered, NULL,
458 tsk);
459}
460
461static int ptrace_gethbpregs(struct task_struct *tsk, long num,
462 unsigned long __user *data)
463{
464 u32 reg;
465 int idx, ret = 0;
466 struct perf_event *bp;
467 struct arch_hw_breakpoint_ctrl arch_ctrl;
468
469 if (num == 0) {
470 reg = ptrace_get_hbp_resource_info();
471 } else {
472 idx = ptrace_hbp_num_to_idx(num);
473 if (idx < 0 || idx >= ARM_MAX_HBP_SLOTS) {
474 ret = -EINVAL;
475 goto out;
476 }
477
478 bp = tsk->thread.debug.hbp[idx];
479 if (!bp) {
480 reg = 0;
481 goto put;
482 }
483
484 arch_ctrl = counter_arch_bp(bp)->ctrl;
485
486
487
488
489
490 while (!(arch_ctrl.len & 0x1))
491 arch_ctrl.len >>= 1;
492
493 if (num & 0x1)
494 reg = bp->attr.bp_addr;
495 else
496 reg = encode_ctrl_reg(arch_ctrl);
497 }
498
499put:
500 if (put_user(reg, data))
501 ret = -EFAULT;
502
503out:
504 return ret;
505}
506
507static int ptrace_sethbpregs(struct task_struct *tsk, long num,
508 unsigned long __user *data)
509{
510 int idx, gen_len, gen_type, implied_type, ret = 0;
511 u32 user_val;
512 struct perf_event *bp;
513 struct arch_hw_breakpoint_ctrl ctrl;
514 struct perf_event_attr attr;
515
516 if (num == 0)
517 goto out;
518 else if (num < 0)
519 implied_type = HW_BREAKPOINT_RW;
520 else
521 implied_type = HW_BREAKPOINT_X;
522
523 idx = ptrace_hbp_num_to_idx(num);
524 if (idx < 0 || idx >= ARM_MAX_HBP_SLOTS) {
525 ret = -EINVAL;
526 goto out;
527 }
528
529 if (get_user(user_val, data)) {
530 ret = -EFAULT;
531 goto out;
532 }
533
534 bp = tsk->thread.debug.hbp[idx];
535 if (!bp) {
536 bp = ptrace_hbp_create(tsk, implied_type);
537 if (IS_ERR(bp)) {
538 ret = PTR_ERR(bp);
539 goto out;
540 }
541 tsk->thread.debug.hbp[idx] = bp;
542 }
543
544 attr = bp->attr;
545
546 if (num & 0x1) {
547
548 attr.bp_addr = user_val;
549 } else {
550
551 decode_ctrl_reg(user_val, &ctrl);
552 ret = arch_bp_generic_fields(ctrl, &gen_len, &gen_type);
553 if (ret)
554 goto out;
555
556 if ((gen_type & implied_type) != gen_type) {
557 ret = -EINVAL;
558 goto out;
559 }
560
561 attr.bp_len = gen_len;
562 attr.bp_type = gen_type;
563 attr.disabled = !ctrl.enabled;
564 }
565
566 ret = modify_user_hw_breakpoint(bp, &attr);
567out:
568 return ret;
569}
570#endif
571
572
573
574static int gpr_get(struct task_struct *target,
575 const struct user_regset *regset,
576 unsigned int pos, unsigned int count,
577 void *kbuf, void __user *ubuf)
578{
579 struct pt_regs *regs = task_pt_regs(target);
580
581 return user_regset_copyout(&pos, &count, &kbuf, &ubuf,
582 regs,
583 0, sizeof(*regs));
584}
585
586static int gpr_set(struct task_struct *target,
587 const struct user_regset *regset,
588 unsigned int pos, unsigned int count,
589 const void *kbuf, const void __user *ubuf)
590{
591 int ret;
592 struct pt_regs newregs = *task_pt_regs(target);
593
594 ret = user_regset_copyin(&pos, &count, &kbuf, &ubuf,
595 &newregs,
596 0, sizeof(newregs));
597 if (ret)
598 return ret;
599
600 if (!valid_user_regs(&newregs))
601 return -EINVAL;
602
603 *task_pt_regs(target) = newregs;
604 return 0;
605}
606
607static int fpa_get(struct task_struct *target,
608 const struct user_regset *regset,
609 unsigned int pos, unsigned int count,
610 void *kbuf, void __user *ubuf)
611{
612 return user_regset_copyout(&pos, &count, &kbuf, &ubuf,
613 &task_thread_info(target)->fpstate,
614 0, sizeof(struct user_fp));
615}
616
617static int fpa_set(struct task_struct *target,
618 const struct user_regset *regset,
619 unsigned int pos, unsigned int count,
620 const void *kbuf, const void __user *ubuf)
621{
622 struct thread_info *thread = task_thread_info(target);
623
624 thread->used_cp[1] = thread->used_cp[2] = 1;
625
626 return user_regset_copyin(&pos, &count, &kbuf, &ubuf,
627 &thread->fpstate,
628 0, sizeof(struct user_fp));
629}
630
631#ifdef CONFIG_VFP
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654static int vfp_get(struct task_struct *target,
655 const struct user_regset *regset,
656 unsigned int pos, unsigned int count,
657 void *kbuf, void __user *ubuf)
658{
659 int ret;
660 struct thread_info *thread = task_thread_info(target);
661 struct vfp_hard_struct const *vfp = &thread->vfpstate.hard;
662 const size_t user_fpregs_offset = offsetof(struct user_vfp, fpregs);
663 const size_t user_fpscr_offset = offsetof(struct user_vfp, fpscr);
664
665 vfp_sync_hwstate(thread);
666
667 ret = user_regset_copyout(&pos, &count, &kbuf, &ubuf,
668 &vfp->fpregs,
669 user_fpregs_offset,
670 user_fpregs_offset + sizeof(vfp->fpregs));
671 if (ret)
672 return ret;
673
674 ret = user_regset_copyout_zero(&pos, &count, &kbuf, &ubuf,
675 user_fpregs_offset + sizeof(vfp->fpregs),
676 user_fpscr_offset);
677 if (ret)
678 return ret;
679
680 return user_regset_copyout(&pos, &count, &kbuf, &ubuf,
681 &vfp->fpscr,
682 user_fpscr_offset,
683 user_fpscr_offset + sizeof(vfp->fpscr));
684}
685
686
687
688
689
690
691static int vfp_set(struct task_struct *target,
692 const struct user_regset *regset,
693 unsigned int pos, unsigned int count,
694 const void *kbuf, const void __user *ubuf)
695{
696 int ret;
697 struct thread_info *thread = task_thread_info(target);
698 struct vfp_hard_struct new_vfp;
699 const size_t user_fpregs_offset = offsetof(struct user_vfp, fpregs);
700 const size_t user_fpscr_offset = offsetof(struct user_vfp, fpscr);
701
702 vfp_sync_hwstate(thread);
703 new_vfp = thread->vfpstate.hard;
704
705 ret = user_regset_copyin(&pos, &count, &kbuf, &ubuf,
706 &new_vfp.fpregs,
707 user_fpregs_offset,
708 user_fpregs_offset + sizeof(new_vfp.fpregs));
709 if (ret)
710 return ret;
711
712 ret = user_regset_copyin_ignore(&pos, &count, &kbuf, &ubuf,
713 user_fpregs_offset + sizeof(new_vfp.fpregs),
714 user_fpscr_offset);
715 if (ret)
716 return ret;
717
718 ret = user_regset_copyin(&pos, &count, &kbuf, &ubuf,
719 &new_vfp.fpscr,
720 user_fpscr_offset,
721 user_fpscr_offset + sizeof(new_vfp.fpscr));
722 if (ret)
723 return ret;
724
725 thread->vfpstate.hard = new_vfp;
726 vfp_flush_hwstate(thread);
727
728 return 0;
729}
730#endif
731
732enum arm_regset {
733 REGSET_GPR,
734 REGSET_FPR,
735#ifdef CONFIG_VFP
736 REGSET_VFP,
737#endif
738};
739
740static const struct user_regset arm_regsets[] = {
741 [REGSET_GPR] = {
742 .core_note_type = NT_PRSTATUS,
743 .n = ELF_NGREG,
744 .size = sizeof(u32),
745 .align = sizeof(u32),
746 .get = gpr_get,
747 .set = gpr_set
748 },
749 [REGSET_FPR] = {
750
751
752
753
754 .core_note_type = NT_PRFPREG,
755 .n = sizeof(struct user_fp) / sizeof(u32),
756 .size = sizeof(u32),
757 .align = sizeof(u32),
758 .get = fpa_get,
759 .set = fpa_set
760 },
761#ifdef CONFIG_VFP
762 [REGSET_VFP] = {
763
764
765
766
767 .core_note_type = NT_ARM_VFP,
768 .n = ARM_VFPREGS_SIZE / sizeof(u32),
769 .size = sizeof(u32),
770 .align = sizeof(u32),
771 .get = vfp_get,
772 .set = vfp_set
773 },
774#endif
775};
776
777static const struct user_regset_view user_arm_view = {
778 .name = "arm", .e_machine = ELF_ARCH, .ei_osabi = ELF_OSABI,
779 .regsets = arm_regsets, .n = ARRAY_SIZE(arm_regsets)
780};
781
782const struct user_regset_view *task_user_regset_view(struct task_struct *task)
783{
784 return &user_arm_view;
785}
786
787long arch_ptrace(struct task_struct *child, long request,
788 unsigned long addr, unsigned long data)
789{
790 int ret;
791 unsigned long __user *datap = (unsigned long __user *) data;
792
793 switch (request) {
794 case PTRACE_PEEKUSR:
795 ret = ptrace_read_user(child, addr, datap);
796 break;
797
798 case PTRACE_POKEUSR:
799 ret = ptrace_write_user(child, addr, data);
800 break;
801
802 case PTRACE_GETREGS:
803 ret = copy_regset_to_user(child,
804 &user_arm_view, REGSET_GPR,
805 0, sizeof(struct pt_regs),
806 datap);
807 break;
808
809 case PTRACE_SETREGS:
810 ret = copy_regset_from_user(child,
811 &user_arm_view, REGSET_GPR,
812 0, sizeof(struct pt_regs),
813 datap);
814 break;
815
816 case PTRACE_GETFPREGS:
817 ret = copy_regset_to_user(child,
818 &user_arm_view, REGSET_FPR,
819 0, sizeof(union fp_state),
820 datap);
821 break;
822
823 case PTRACE_SETFPREGS:
824 ret = copy_regset_from_user(child,
825 &user_arm_view, REGSET_FPR,
826 0, sizeof(union fp_state),
827 datap);
828 break;
829
830#ifdef CONFIG_IWMMXT
831 case PTRACE_GETWMMXREGS:
832 ret = ptrace_getwmmxregs(child, datap);
833 break;
834
835 case PTRACE_SETWMMXREGS:
836 ret = ptrace_setwmmxregs(child, datap);
837 break;
838#endif
839
840 case PTRACE_GET_THREAD_AREA:
841 ret = put_user(task_thread_info(child)->tp_value[0],
842 datap);
843 break;
844
845 case PTRACE_SET_SYSCALL:
846 task_thread_info(child)->syscall = data;
847 ret = 0;
848 break;
849
850#ifdef CONFIG_CRUNCH
851 case PTRACE_GETCRUNCHREGS:
852 ret = ptrace_getcrunchregs(child, datap);
853 break;
854
855 case PTRACE_SETCRUNCHREGS:
856 ret = ptrace_setcrunchregs(child, datap);
857 break;
858#endif
859
860#ifdef CONFIG_VFP
861 case PTRACE_GETVFPREGS:
862 ret = copy_regset_to_user(child,
863 &user_arm_view, REGSET_VFP,
864 0, ARM_VFPREGS_SIZE,
865 datap);
866 break;
867
868 case PTRACE_SETVFPREGS:
869 ret = copy_regset_from_user(child,
870 &user_arm_view, REGSET_VFP,
871 0, ARM_VFPREGS_SIZE,
872 datap);
873 break;
874#endif
875
876#ifdef CONFIG_HAVE_HW_BREAKPOINT
877 case PTRACE_GETHBPREGS:
878 ret = ptrace_gethbpregs(child, addr,
879 (unsigned long __user *)data);
880 break;
881 case PTRACE_SETHBPREGS:
882 ret = ptrace_sethbpregs(child, addr,
883 (unsigned long __user *)data);
884 break;
885#endif
886
887 default:
888 ret = ptrace_request(child, request, addr, data);
889 break;
890 }
891
892 return ret;
893}
894
895enum ptrace_syscall_dir {
896 PTRACE_SYSCALL_ENTER = 0,
897 PTRACE_SYSCALL_EXIT,
898};
899
900static void tracehook_report_syscall(struct pt_regs *regs,
901 enum ptrace_syscall_dir dir)
902{
903 unsigned long ip;
904
905
906
907
908
909 ip = regs->ARM_ip;
910 regs->ARM_ip = dir;
911
912 if (dir == PTRACE_SYSCALL_EXIT)
913 tracehook_report_syscall_exit(regs, 0);
914 else if (tracehook_report_syscall_entry(regs))
915 current_thread_info()->syscall = -1;
916
917 regs->ARM_ip = ip;
918}
919
920asmlinkage int syscall_trace_enter(struct pt_regs *regs, int scno)
921{
922 current_thread_info()->syscall = scno;
923
924 if (test_thread_flag(TIF_SYSCALL_TRACE))
925 tracehook_report_syscall(regs, PTRACE_SYSCALL_ENTER);
926
927
928#ifdef CONFIG_HAVE_ARCH_SECCOMP_FILTER
929 if (secure_computing(NULL) == -1)
930 return -1;
931#else
932
933 secure_computing_strict(current_thread_info()->syscall);
934#endif
935
936
937 scno = current_thread_info()->syscall;
938
939 if (test_thread_flag(TIF_SYSCALL_TRACEPOINT))
940 trace_sys_enter(regs, scno);
941
942 audit_syscall_entry(scno, regs->ARM_r0, regs->ARM_r1, regs->ARM_r2,
943 regs->ARM_r3);
944
945 return scno;
946}
947
948asmlinkage void syscall_trace_exit(struct pt_regs *regs)
949{
950
951
952
953
954 audit_syscall_exit(regs);
955
956
957
958
959
960
961
962 if (test_thread_flag(TIF_SYSCALL_TRACEPOINT))
963 trace_sys_exit(regs, regs_return_value(regs));
964
965 if (test_thread_flag(TIF_SYSCALL_TRACE))
966 tracehook_report_syscall(regs, PTRACE_SYSCALL_EXIT);
967}
968