2005-04-16 15:20:36 -07:00
|
|
|
/* By Ross Biro 1/23/92 */
|
|
|
|
/*
|
|
|
|
* Pentium III FXSR, SSE support
|
|
|
|
* Gareth Hughes <gareth@valinux.com>, May 2000
|
2008-01-30 05:31:09 -07:00
|
|
|
*
|
|
|
|
* BTS tracing
|
|
|
|
* Markus Metzger <markus.t.metzger@intel.com>, Dec 2007
|
2005-04-16 15:20:36 -07:00
|
|
|
*/
|
|
|
|
|
|
|
|
#include <linux/kernel.h>
|
|
|
|
#include <linux/sched.h>
|
|
|
|
#include <linux/mm.h>
|
|
|
|
#include <linux/smp.h>
|
|
|
|
#include <linux/errno.h>
|
|
|
|
#include <linux/ptrace.h>
|
2008-01-30 05:31:52 -07:00
|
|
|
#include <linux/regset.h>
|
2005-04-16 15:20:36 -07:00
|
|
|
#include <linux/user.h>
|
2008-01-30 05:31:53 -07:00
|
|
|
#include <linux/elf.h>
|
2005-04-16 15:20:36 -07:00
|
|
|
#include <linux/security.h>
|
|
|
|
#include <linux/audit.h>
|
|
|
|
#include <linux/seccomp.h>
|
2005-05-01 08:59:14 -07:00
|
|
|
#include <linux/signal.h>
|
2005-04-16 15:20:36 -07:00
|
|
|
|
|
|
|
#include <asm/uaccess.h>
|
|
|
|
#include <asm/pgtable.h>
|
|
|
|
#include <asm/system.h>
|
|
|
|
#include <asm/processor.h>
|
|
|
|
#include <asm/i387.h>
|
|
|
|
#include <asm/debugreg.h>
|
|
|
|
#include <asm/ldt.h>
|
|
|
|
#include <asm/desc.h>
|
2008-01-30 05:31:01 -07:00
|
|
|
#include <asm/prctl.h>
|
|
|
|
#include <asm/proto.h>
|
2008-01-30 05:31:09 -07:00
|
|
|
#include <asm/ds.h>
|
|
|
|
|
2008-01-30 05:31:53 -07:00
|
|
|
#include "tls.h"
|
|
|
|
|
|
|
|
enum x86_regset {
|
|
|
|
REGSET_GENERAL,
|
|
|
|
REGSET_FP,
|
|
|
|
REGSET_XFP,
|
|
|
|
REGSET_TLS,
|
|
|
|
};
|
2008-01-30 05:31:09 -07:00
|
|
|
|
2005-04-16 15:20:36 -07:00
|
|
|
/*
|
|
|
|
* does not yet catch signals sent when the child dies.
|
|
|
|
* in exit.c or in signal.c.
|
|
|
|
*/
|
|
|
|
|
2006-01-05 21:11:29 -07:00
|
|
|
/*
|
|
|
|
* Determines which flags the user has access to [1 = access, 0 = no access].
|
|
|
|
*/
|
2008-01-30 05:31:01 -07:00
|
|
|
#define FLAG_MASK_32 ((unsigned long) \
|
|
|
|
(X86_EFLAGS_CF | X86_EFLAGS_PF | \
|
|
|
|
X86_EFLAGS_AF | X86_EFLAGS_ZF | \
|
|
|
|
X86_EFLAGS_SF | X86_EFLAGS_TF | \
|
|
|
|
X86_EFLAGS_DF | X86_EFLAGS_OF | \
|
|
|
|
X86_EFLAGS_RF | X86_EFLAGS_AC))
|
|
|
|
|
2008-01-30 05:31:01 -07:00
|
|
|
/*
|
|
|
|
* Determines whether a value may be installed in a segment register.
|
|
|
|
*/
|
|
|
|
static inline bool invalid_selector(u16 value)
|
|
|
|
{
|
|
|
|
return unlikely(value != 0 && (value & SEGMENT_RPL_MASK) != USER_RPL);
|
|
|
|
}
|
|
|
|
|
|
|
|
#ifdef CONFIG_X86_32
|
|
|
|
|
2008-01-30 05:31:01 -07:00
|
|
|
#define FLAG_MASK FLAG_MASK_32
|
2005-04-16 15:20:36 -07:00
|
|
|
|
2008-01-30 05:30:52 -07:00
|
|
|
static long *pt_regs_access(struct pt_regs *regs, unsigned long regno)
|
2005-04-16 15:20:36 -07:00
|
|
|
{
|
2008-01-30 05:30:56 -07:00
|
|
|
BUILD_BUG_ON(offsetof(struct pt_regs, bx) != 0);
|
2008-01-30 05:31:01 -07:00
|
|
|
regno >>= 2;
|
2008-01-30 05:30:52 -07:00
|
|
|
if (regno > FS)
|
|
|
|
--regno;
|
2008-01-30 05:30:56 -07:00
|
|
|
return ®s->bx + regno;
|
2005-04-16 15:20:36 -07:00
|
|
|
}
|
|
|
|
|
2008-01-30 05:31:01 -07:00
|
|
|
static u16 get_segment_reg(struct task_struct *task, unsigned long offset)
|
2005-04-16 15:20:36 -07:00
|
|
|
{
|
2008-01-30 05:31:01 -07:00
|
|
|
/*
|
|
|
|
* Returning the value truncates it to 16 bits.
|
|
|
|
*/
|
|
|
|
unsigned int retval;
|
|
|
|
if (offset != offsetof(struct user_regs_struct, gs))
|
|
|
|
retval = *pt_regs_access(task_pt_regs(task), offset);
|
|
|
|
else {
|
|
|
|
retval = task->thread.gs;
|
|
|
|
if (task == current)
|
|
|
|
savesegment(gs, retval);
|
|
|
|
}
|
|
|
|
return retval;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int set_segment_reg(struct task_struct *task,
|
|
|
|
unsigned long offset, u16 value)
|
|
|
|
{
|
|
|
|
/*
|
|
|
|
* The value argument was already truncated to 16 bits.
|
|
|
|
*/
|
2008-01-30 05:31:01 -07:00
|
|
|
if (invalid_selector(value))
|
2008-01-30 05:31:01 -07:00
|
|
|
return -EIO;
|
|
|
|
|
2008-02-06 14:39:44 -07:00
|
|
|
/*
|
|
|
|
* For %cs and %ss we cannot permit a null selector.
|
|
|
|
* We can permit a bogus selector as long as it has USER_RPL.
|
|
|
|
* Null selectors are fine for other segment registers, but
|
|
|
|
* we will never get back to user mode with invalid %cs or %ss
|
|
|
|
* and will take the trap in iret instead. Much code relies
|
|
|
|
* on user_mode() to distinguish a user trap frame (which can
|
|
|
|
* safely use invalid selectors) from a kernel trap frame.
|
|
|
|
*/
|
|
|
|
switch (offset) {
|
|
|
|
case offsetof(struct user_regs_struct, cs):
|
|
|
|
case offsetof(struct user_regs_struct, ss):
|
|
|
|
if (unlikely(value == 0))
|
|
|
|
return -EIO;
|
|
|
|
|
|
|
|
default:
|
2008-01-30 05:31:01 -07:00
|
|
|
*pt_regs_access(task_pt_regs(task), offset) = value;
|
2008-02-06 14:39:44 -07:00
|
|
|
break;
|
|
|
|
|
|
|
|
case offsetof(struct user_regs_struct, gs):
|
2008-01-30 05:31:01 -07:00
|
|
|
task->thread.gs = value;
|
|
|
|
if (task == current)
|
2008-01-30 05:30:58 -07:00
|
|
|
/*
|
|
|
|
* The user-mode %gs is not affected by
|
|
|
|
* kernel entry, so we must update the CPU.
|
|
|
|
*/
|
|
|
|
loadsegment(gs, value);
|
2005-04-16 15:20:36 -07:00
|
|
|
}
|
2008-01-30 05:31:01 -07:00
|
|
|
|
2005-04-16 15:20:36 -07:00
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2008-01-30 05:31:01 -07:00
|
|
|
static unsigned long debugreg_addr_limit(struct task_struct *task)
|
|
|
|
{
|
|
|
|
return TASK_SIZE - 3;
|
|
|
|
}
|
|
|
|
|
|
|
|
#else /* CONFIG_X86_64 */
|
|
|
|
|
|
|
|
#define FLAG_MASK (FLAG_MASK_32 | X86_EFLAGS_NT)
|
|
|
|
|
|
|
|
static unsigned long *pt_regs_access(struct pt_regs *regs, unsigned long offset)
|
|
|
|
{
|
|
|
|
BUILD_BUG_ON(offsetof(struct pt_regs, r15) != 0);
|
|
|
|
return ®s->r15 + (offset / sizeof(regs->r15));
|
|
|
|
}
|
|
|
|
|
|
|
|
static u16 get_segment_reg(struct task_struct *task, unsigned long offset)
|
|
|
|
{
|
|
|
|
/*
|
|
|
|
* Returning the value truncates it to 16 bits.
|
|
|
|
*/
|
|
|
|
unsigned int seg;
|
|
|
|
|
|
|
|
switch (offset) {
|
|
|
|
case offsetof(struct user_regs_struct, fs):
|
|
|
|
if (task == current) {
|
|
|
|
/* Older gas can't assemble movq %?s,%r?? */
|
|
|
|
asm("movl %%fs,%0" : "=r" (seg));
|
|
|
|
return seg;
|
|
|
|
}
|
|
|
|
return task->thread.fsindex;
|
|
|
|
case offsetof(struct user_regs_struct, gs):
|
|
|
|
if (task == current) {
|
|
|
|
asm("movl %%gs,%0" : "=r" (seg));
|
|
|
|
return seg;
|
|
|
|
}
|
|
|
|
return task->thread.gsindex;
|
|
|
|
case offsetof(struct user_regs_struct, ds):
|
|
|
|
if (task == current) {
|
|
|
|
asm("movl %%ds,%0" : "=r" (seg));
|
|
|
|
return seg;
|
|
|
|
}
|
|
|
|
return task->thread.ds;
|
|
|
|
case offsetof(struct user_regs_struct, es):
|
|
|
|
if (task == current) {
|
|
|
|
asm("movl %%es,%0" : "=r" (seg));
|
|
|
|
return seg;
|
|
|
|
}
|
|
|
|
return task->thread.es;
|
|
|
|
|
|
|
|
case offsetof(struct user_regs_struct, cs):
|
|
|
|
case offsetof(struct user_regs_struct, ss):
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
return *pt_regs_access(task_pt_regs(task), offset);
|
|
|
|
}
|
|
|
|
|
|
|
|
static int set_segment_reg(struct task_struct *task,
|
|
|
|
unsigned long offset, u16 value)
|
|
|
|
{
|
|
|
|
/*
|
|
|
|
* The value argument was already truncated to 16 bits.
|
|
|
|
*/
|
|
|
|
if (invalid_selector(value))
|
|
|
|
return -EIO;
|
|
|
|
|
|
|
|
switch (offset) {
|
|
|
|
case offsetof(struct user_regs_struct,fs):
|
|
|
|
/*
|
|
|
|
* If this is setting fs as for normal 64-bit use but
|
|
|
|
* setting fs_base has implicitly changed it, leave it.
|
|
|
|
*/
|
|
|
|
if ((value == FS_TLS_SEL && task->thread.fsindex == 0 &&
|
|
|
|
task->thread.fs != 0) ||
|
|
|
|
(value == 0 && task->thread.fsindex == FS_TLS_SEL &&
|
|
|
|
task->thread.fs == 0))
|
|
|
|
break;
|
|
|
|
task->thread.fsindex = value;
|
|
|
|
if (task == current)
|
|
|
|
loadsegment(fs, task->thread.fsindex);
|
|
|
|
break;
|
|
|
|
case offsetof(struct user_regs_struct,gs):
|
|
|
|
/*
|
|
|
|
* If this is setting gs as for normal 64-bit use but
|
|
|
|
* setting gs_base has implicitly changed it, leave it.
|
|
|
|
*/
|
|
|
|
if ((value == GS_TLS_SEL && task->thread.gsindex == 0 &&
|
|
|
|
task->thread.gs != 0) ||
|
|
|
|
(value == 0 && task->thread.gsindex == GS_TLS_SEL &&
|
|
|
|
task->thread.gs == 0))
|
|
|
|
break;
|
|
|
|
task->thread.gsindex = value;
|
|
|
|
if (task == current)
|
|
|
|
load_gs_index(task->thread.gsindex);
|
|
|
|
break;
|
|
|
|
case offsetof(struct user_regs_struct,ds):
|
|
|
|
task->thread.ds = value;
|
|
|
|
if (task == current)
|
|
|
|
loadsegment(ds, task->thread.ds);
|
|
|
|
break;
|
|
|
|
case offsetof(struct user_regs_struct,es):
|
|
|
|
task->thread.es = value;
|
|
|
|
if (task == current)
|
|
|
|
loadsegment(es, task->thread.es);
|
|
|
|
break;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Can't actually change these in 64-bit mode.
|
|
|
|
*/
|
|
|
|
case offsetof(struct user_regs_struct,cs):
|
2008-02-06 14:39:44 -07:00
|
|
|
if (unlikely(value == 0))
|
|
|
|
return -EIO;
|
2008-01-30 05:31:01 -07:00
|
|
|
#ifdef CONFIG_IA32_EMULATION
|
|
|
|
if (test_tsk_thread_flag(task, TIF_IA32))
|
|
|
|
task_pt_regs(task)->cs = value;
|
|
|
|
#endif
|
2008-01-30 05:31:01 -07:00
|
|
|
break;
|
2008-01-30 05:31:01 -07:00
|
|
|
case offsetof(struct user_regs_struct,ss):
|
2008-02-06 14:39:44 -07:00
|
|
|
if (unlikely(value == 0))
|
|
|
|
return -EIO;
|
2008-01-30 05:31:01 -07:00
|
|
|
#ifdef CONFIG_IA32_EMULATION
|
|
|
|
if (test_tsk_thread_flag(task, TIF_IA32))
|
|
|
|
task_pt_regs(task)->ss = value;
|
|
|
|
#endif
|
2008-01-30 05:31:01 -07:00
|
|
|
break;
|
2008-01-30 05:31:01 -07:00
|
|
|
}
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static unsigned long debugreg_addr_limit(struct task_struct *task)
|
|
|
|
{
|
|
|
|
#ifdef CONFIG_IA32_EMULATION
|
|
|
|
if (test_tsk_thread_flag(task, TIF_IA32))
|
|
|
|
return IA32_PAGE_OFFSET - 3;
|
|
|
|
#endif
|
|
|
|
return TASK_SIZE64 - 7;
|
|
|
|
}
|
|
|
|
|
|
|
|
#endif /* CONFIG_X86_32 */
|
|
|
|
|
2008-01-30 05:31:01 -07:00
|
|
|
static unsigned long get_flags(struct task_struct *task)
|
2005-04-16 15:20:36 -07:00
|
|
|
{
|
2008-01-30 05:31:01 -07:00
|
|
|
unsigned long retval = task_pt_regs(task)->flags;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* If the debugger set TF, hide it from the readout.
|
|
|
|
*/
|
|
|
|
if (test_tsk_thread_flag(task, TIF_FORCED_TF))
|
|
|
|
retval &= ~X86_EFLAGS_TF;
|
2005-04-16 15:20:36 -07:00
|
|
|
|
|
|
|
return retval;
|
|
|
|
}
|
|
|
|
|
2008-01-30 05:31:01 -07:00
|
|
|
static int set_flags(struct task_struct *task, unsigned long value)
|
|
|
|
{
|
|
|
|
struct pt_regs *regs = task_pt_regs(task);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* If the user value contains TF, mark that
|
|
|
|
* it was not "us" (the debugger) that set it.
|
|
|
|
* If not, make sure it stays set if we had.
|
|
|
|
*/
|
|
|
|
if (value & X86_EFLAGS_TF)
|
|
|
|
clear_tsk_thread_flag(task, TIF_FORCED_TF);
|
|
|
|
else if (test_tsk_thread_flag(task, TIF_FORCED_TF))
|
|
|
|
value |= X86_EFLAGS_TF;
|
|
|
|
|
|
|
|
regs->flags = (regs->flags & ~FLAG_MASK) | (value & FLAG_MASK);
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int putreg(struct task_struct *child,
|
|
|
|
unsigned long offset, unsigned long value)
|
|
|
|
{
|
|
|
|
switch (offset) {
|
|
|
|
case offsetof(struct user_regs_struct, cs):
|
|
|
|
case offsetof(struct user_regs_struct, ds):
|
|
|
|
case offsetof(struct user_regs_struct, es):
|
|
|
|
case offsetof(struct user_regs_struct, fs):
|
|
|
|
case offsetof(struct user_regs_struct, gs):
|
|
|
|
case offsetof(struct user_regs_struct, ss):
|
|
|
|
return set_segment_reg(child, offset, value);
|
|
|
|
|
|
|
|
case offsetof(struct user_regs_struct, flags):
|
|
|
|
return set_flags(child, value);
|
2008-01-30 05:31:01 -07:00
|
|
|
|
|
|
|
#ifdef CONFIG_X86_64
|
2008-03-07 15:56:02 -07:00
|
|
|
/*
|
|
|
|
* Orig_ax is really just a flag with small positive and
|
|
|
|
* negative values, so make sure to always sign-extend it
|
|
|
|
* from 32 bits so that it works correctly regardless of
|
|
|
|
* whether we come from a 32-bit environment or not.
|
|
|
|
*/
|
|
|
|
case offsetof(struct user_regs_struct, orig_ax):
|
|
|
|
value = (long) (s32) value;
|
|
|
|
break;
|
|
|
|
|
2008-01-30 05:31:01 -07:00
|
|
|
case offsetof(struct user_regs_struct,fs_base):
|
|
|
|
if (value >= TASK_SIZE_OF(child))
|
|
|
|
return -EIO;
|
|
|
|
/*
|
|
|
|
* When changing the segment base, use do_arch_prctl
|
|
|
|
* to set either thread.fs or thread.fsindex and the
|
|
|
|
* corresponding GDT slot.
|
|
|
|
*/
|
|
|
|
if (child->thread.fs != value)
|
|
|
|
return do_arch_prctl(child, ARCH_SET_FS, value);
|
|
|
|
return 0;
|
|
|
|
case offsetof(struct user_regs_struct,gs_base):
|
|
|
|
/*
|
|
|
|
* Exactly the same here as the %fs handling above.
|
|
|
|
*/
|
|
|
|
if (value >= TASK_SIZE_OF(child))
|
|
|
|
return -EIO;
|
|
|
|
if (child->thread.gs != value)
|
|
|
|
return do_arch_prctl(child, ARCH_SET_GS, value);
|
|
|
|
return 0;
|
|
|
|
#endif
|
2008-01-30 05:31:01 -07:00
|
|
|
}
|
|
|
|
|
|
|
|
*pt_regs_access(task_pt_regs(child), offset) = value;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static unsigned long getreg(struct task_struct *task, unsigned long offset)
|
|
|
|
{
|
|
|
|
switch (offset) {
|
|
|
|
case offsetof(struct user_regs_struct, cs):
|
|
|
|
case offsetof(struct user_regs_struct, ds):
|
|
|
|
case offsetof(struct user_regs_struct, es):
|
|
|
|
case offsetof(struct user_regs_struct, fs):
|
|
|
|
case offsetof(struct user_regs_struct, gs):
|
|
|
|
case offsetof(struct user_regs_struct, ss):
|
|
|
|
return get_segment_reg(task, offset);
|
|
|
|
|
|
|
|
case offsetof(struct user_regs_struct, flags):
|
|
|
|
return get_flags(task);
|
2008-01-30 05:31:01 -07:00
|
|
|
|
|
|
|
#ifdef CONFIG_X86_64
|
|
|
|
case offsetof(struct user_regs_struct, fs_base): {
|
|
|
|
/*
|
|
|
|
* do_arch_prctl may have used a GDT slot instead of
|
|
|
|
* the MSR. To userland, it appears the same either
|
|
|
|
* way, except the %fs segment selector might not be 0.
|
|
|
|
*/
|
|
|
|
unsigned int seg = task->thread.fsindex;
|
|
|
|
if (task->thread.fs != 0)
|
|
|
|
return task->thread.fs;
|
|
|
|
if (task == current)
|
|
|
|
asm("movl %%fs,%0" : "=r" (seg));
|
|
|
|
if (seg != FS_TLS_SEL)
|
|
|
|
return 0;
|
|
|
|
return get_desc_base(&task->thread.tls_array[FS_TLS]);
|
|
|
|
}
|
|
|
|
case offsetof(struct user_regs_struct, gs_base): {
|
|
|
|
/*
|
|
|
|
* Exactly the same here as the %fs handling above.
|
|
|
|
*/
|
|
|
|
unsigned int seg = task->thread.gsindex;
|
|
|
|
if (task->thread.gs != 0)
|
|
|
|
return task->thread.gs;
|
|
|
|
if (task == current)
|
|
|
|
asm("movl %%gs,%0" : "=r" (seg));
|
|
|
|
if (seg != GS_TLS_SEL)
|
|
|
|
return 0;
|
|
|
|
return get_desc_base(&task->thread.tls_array[GS_TLS]);
|
|
|
|
}
|
|
|
|
#endif
|
2008-01-30 05:31:01 -07:00
|
|
|
}
|
|
|
|
|
|
|
|
return *pt_regs_access(task_pt_regs(task), offset);
|
|
|
|
}
|
|
|
|
|
2008-01-30 05:31:52 -07:00
|
|
|
static int genregs_get(struct task_struct *target,
|
|
|
|
const struct user_regset *regset,
|
|
|
|
unsigned int pos, unsigned int count,
|
|
|
|
void *kbuf, void __user *ubuf)
|
|
|
|
{
|
|
|
|
if (kbuf) {
|
|
|
|
unsigned long *k = kbuf;
|
|
|
|
while (count > 0) {
|
|
|
|
*k++ = getreg(target, pos);
|
|
|
|
count -= sizeof(*k);
|
|
|
|
pos += sizeof(*k);
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
unsigned long __user *u = ubuf;
|
|
|
|
while (count > 0) {
|
|
|
|
if (__put_user(getreg(target, pos), u++))
|
|
|
|
return -EFAULT;
|
|
|
|
count -= sizeof(*u);
|
|
|
|
pos += sizeof(*u);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int genregs_set(struct task_struct *target,
|
|
|
|
const struct user_regset *regset,
|
|
|
|
unsigned int pos, unsigned int count,
|
|
|
|
const void *kbuf, const void __user *ubuf)
|
|
|
|
{
|
|
|
|
int ret = 0;
|
|
|
|
if (kbuf) {
|
|
|
|
const unsigned long *k = kbuf;
|
|
|
|
while (count > 0 && !ret) {
|
|
|
|
ret = putreg(target, pos, *k++);
|
|
|
|
count -= sizeof(*k);
|
|
|
|
pos += sizeof(*k);
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
const unsigned long __user *u = ubuf;
|
|
|
|
while (count > 0 && !ret) {
|
|
|
|
unsigned long word;
|
|
|
|
ret = __get_user(word, u++);
|
|
|
|
if (ret)
|
|
|
|
break;
|
|
|
|
ret = putreg(target, pos, word);
|
|
|
|
count -= sizeof(*u);
|
|
|
|
pos += sizeof(*u);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
2008-01-30 05:30:52 -07:00
|
|
|
/*
|
|
|
|
* This function is trivial and will be inlined by the compiler.
|
|
|
|
* Having it separates the implementation details of debug
|
|
|
|
* registers from the interface details of ptrace.
|
|
|
|
*/
|
|
|
|
static unsigned long ptrace_get_debugreg(struct task_struct *child, int n)
|
|
|
|
{
|
2008-01-30 05:30:59 -07:00
|
|
|
switch (n) {
|
|
|
|
case 0: return child->thread.debugreg0;
|
|
|
|
case 1: return child->thread.debugreg1;
|
|
|
|
case 2: return child->thread.debugreg2;
|
|
|
|
case 3: return child->thread.debugreg3;
|
|
|
|
case 6: return child->thread.debugreg6;
|
|
|
|
case 7: return child->thread.debugreg7;
|
|
|
|
}
|
|
|
|
return 0;
|
2008-01-30 05:30:52 -07:00
|
|
|
}
|
|
|
|
|
|
|
|
static int ptrace_set_debugreg(struct task_struct *child,
|
|
|
|
int n, unsigned long data)
|
|
|
|
{
|
2008-01-30 05:30:59 -07:00
|
|
|
int i;
|
|
|
|
|
2008-01-30 05:30:52 -07:00
|
|
|
if (unlikely(n == 4 || n == 5))
|
|
|
|
return -EIO;
|
|
|
|
|
2008-01-30 05:31:01 -07:00
|
|
|
if (n < 4 && unlikely(data >= debugreg_addr_limit(child)))
|
2008-01-30 05:30:52 -07:00
|
|
|
return -EIO;
|
|
|
|
|
2008-01-30 05:30:59 -07:00
|
|
|
switch (n) {
|
|
|
|
case 0: child->thread.debugreg0 = data; break;
|
|
|
|
case 1: child->thread.debugreg1 = data; break;
|
|
|
|
case 2: child->thread.debugreg2 = data; break;
|
|
|
|
case 3: child->thread.debugreg3 = data; break;
|
|
|
|
|
|
|
|
case 6:
|
2008-01-30 05:31:01 -07:00
|
|
|
if ((data & ~0xffffffffUL) != 0)
|
|
|
|
return -EIO;
|
2008-01-30 05:30:59 -07:00
|
|
|
child->thread.debugreg6 = data;
|
|
|
|
break;
|
|
|
|
|
|
|
|
case 7:
|
2008-01-30 05:30:52 -07:00
|
|
|
/*
|
|
|
|
* Sanity-check data. Take one half-byte at once with
|
|
|
|
* check = (val >> (16 + 4*i)) & 0xf. It contains the
|
|
|
|
* R/Wi and LENi bits; bits 0 and 1 are R/Wi, and bits
|
|
|
|
* 2 and 3 are LENi. Given a list of invalid values,
|
|
|
|
* we do mask |= 1 << invalid_value, so that
|
|
|
|
* (mask >> check) & 1 is a correct test for invalid
|
|
|
|
* values.
|
|
|
|
*
|
|
|
|
* R/Wi contains the type of the breakpoint /
|
|
|
|
* watchpoint, LENi contains the length of the watched
|
|
|
|
* data in the watchpoint case.
|
|
|
|
*
|
|
|
|
* The invalid values are:
|
2008-01-30 05:31:01 -07:00
|
|
|
* - LENi == 0x10 (undefined), so mask |= 0x0f00. [32-bit]
|
2008-01-30 05:30:52 -07:00
|
|
|
* - R/Wi == 0x10 (break on I/O reads or writes), so
|
|
|
|
* mask |= 0x4444.
|
|
|
|
* - R/Wi == 0x00 && LENi != 0x00, so we have mask |=
|
|
|
|
* 0x1110.
|
|
|
|
*
|
|
|
|
* Finally, mask = 0x0f00 | 0x4444 | 0x1110 == 0x5f54.
|
|
|
|
*
|
|
|
|
* See the Intel Manual "System Programming Guide",
|
|
|
|
* 15.2.4
|
|
|
|
*
|
|
|
|
* Note that LENi == 0x10 is defined on x86_64 in long
|
|
|
|
* mode (i.e. even for 32-bit userspace software, but
|
|
|
|
* 64-bit kernel), so the x86_64 mask value is 0x5454.
|
|
|
|
* See the AMD manual no. 24593 (AMD64 System Programming)
|
|
|
|
*/
|
2008-01-30 05:31:01 -07:00
|
|
|
#ifdef CONFIG_X86_32
|
|
|
|
#define DR7_MASK 0x5f54
|
|
|
|
#else
|
|
|
|
#define DR7_MASK 0x5554
|
|
|
|
#endif
|
2008-01-30 05:30:52 -07:00
|
|
|
data &= ~DR_CONTROL_RESERVED;
|
|
|
|
for (i = 0; i < 4; i++)
|
2008-01-30 05:31:01 -07:00
|
|
|
if ((DR7_MASK >> ((data >> (16 + 4*i)) & 0xf)) & 1)
|
2008-01-30 05:30:52 -07:00
|
|
|
return -EIO;
|
2008-01-30 05:30:59 -07:00
|
|
|
child->thread.debugreg7 = data;
|
2008-01-30 05:30:52 -07:00
|
|
|
if (data)
|
|
|
|
set_tsk_thread_flag(child, TIF_DEBUG);
|
|
|
|
else
|
|
|
|
clear_tsk_thread_flag(child, TIF_DEBUG);
|
2008-01-30 05:30:59 -07:00
|
|
|
break;
|
2008-01-30 05:30:52 -07:00
|
|
|
}
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2008-02-26 01:40:27 -07:00
|
|
|
#ifdef X86_BTS
|
|
|
|
|
2008-01-30 05:31:20 -07:00
|
|
|
static int ptrace_bts_get_size(struct task_struct *child)
|
2008-01-30 05:31:09 -07:00
|
|
|
{
|
|
|
|
if (!child->thread.ds_area_msr)
|
|
|
|
return -ENXIO;
|
|
|
|
|
2008-01-30 05:31:20 -07:00
|
|
|
return ds_get_bts_index((void *)child->thread.ds_area_msr);
|
2008-01-30 05:31:09 -07:00
|
|
|
}
|
|
|
|
|
|
|
|
static int ptrace_bts_read_record(struct task_struct *child,
|
|
|
|
long index,
|
|
|
|
struct bts_struct __user *out)
|
|
|
|
{
|
|
|
|
struct bts_struct ret;
|
|
|
|
int retval;
|
2008-01-30 05:31:20 -07:00
|
|
|
int bts_end;
|
2008-01-30 05:31:20 -07:00
|
|
|
int bts_index;
|
2008-01-30 05:31:09 -07:00
|
|
|
|
|
|
|
if (!child->thread.ds_area_msr)
|
|
|
|
return -ENXIO;
|
|
|
|
|
2008-01-30 05:31:20 -07:00
|
|
|
if (index < 0)
|
|
|
|
return -EINVAL;
|
|
|
|
|
2008-01-30 05:31:20 -07:00
|
|
|
bts_end = ds_get_bts_end((void *)child->thread.ds_area_msr);
|
|
|
|
if (bts_end <= index)
|
2008-01-30 05:31:20 -07:00
|
|
|
return -EINVAL;
|
|
|
|
|
|
|
|
/* translate the ptrace bts index into the ds bts index */
|
|
|
|
bts_index = ds_get_bts_index((void *)child->thread.ds_area_msr);
|
|
|
|
bts_index -= (index + 1);
|
|
|
|
if (bts_index < 0)
|
2008-01-30 05:31:20 -07:00
|
|
|
bts_index += bts_end;
|
2008-01-30 05:31:20 -07:00
|
|
|
|
2008-01-30 05:31:09 -07:00
|
|
|
retval = ds_read_bts((void *)child->thread.ds_area_msr,
|
2008-01-30 05:31:20 -07:00
|
|
|
bts_index, &ret);
|
2008-01-30 05:32:03 -07:00
|
|
|
if (retval < 0)
|
2008-01-30 05:31:09 -07:00
|
|
|
return retval;
|
|
|
|
|
|
|
|
if (copy_to_user(out, &ret, sizeof(ret)))
|
|
|
|
return -EFAULT;
|
|
|
|
|
|
|
|
return sizeof(ret);
|
|
|
|
}
|
|
|
|
|
2008-01-30 05:31:20 -07:00
|
|
|
static int ptrace_bts_clear(struct task_struct *child)
|
2008-01-30 05:31:09 -07:00
|
|
|
{
|
2008-01-30 05:31:20 -07:00
|
|
|
if (!child->thread.ds_area_msr)
|
|
|
|
return -ENXIO;
|
2008-01-30 05:31:09 -07:00
|
|
|
|
2008-01-30 05:31:20 -07:00
|
|
|
return ds_clear((void *)child->thread.ds_area_msr);
|
|
|
|
}
|
|
|
|
|
|
|
|
static int ptrace_bts_drain(struct task_struct *child,
|
2008-01-30 05:32:03 -07:00
|
|
|
long size,
|
2008-01-30 05:31:20 -07:00
|
|
|
struct bts_struct __user *out)
|
|
|
|
{
|
|
|
|
int end, i;
|
|
|
|
void *ds = (void *)child->thread.ds_area_msr;
|
|
|
|
|
|
|
|
if (!ds)
|
2008-01-30 05:31:09 -07:00
|
|
|
return -ENXIO;
|
|
|
|
|
2008-01-30 05:31:20 -07:00
|
|
|
end = ds_get_bts_index(ds);
|
|
|
|
if (end <= 0)
|
|
|
|
return end;
|
|
|
|
|
2008-01-30 05:32:03 -07:00
|
|
|
if (size < (end * sizeof(struct bts_struct)))
|
|
|
|
return -EIO;
|
|
|
|
|
2008-01-30 05:31:20 -07:00
|
|
|
for (i = 0; i < end; i++, out++) {
|
|
|
|
struct bts_struct ret;
|
|
|
|
int retval;
|
|
|
|
|
|
|
|
retval = ds_read_bts(ds, i, &ret);
|
|
|
|
if (retval < 0)
|
|
|
|
return retval;
|
|
|
|
|
|
|
|
if (copy_to_user(out, &ret, sizeof(ret)))
|
|
|
|
return -EFAULT;
|
|
|
|
}
|
|
|
|
|
|
|
|
ds_clear(ds);
|
|
|
|
|
2008-01-30 05:32:03 -07:00
|
|
|
return end;
|
2008-01-30 05:31:20 -07:00
|
|
|
}
|
|
|
|
|
|
|
|
static int ptrace_bts_config(struct task_struct *child,
|
2008-01-30 05:32:03 -07:00
|
|
|
long cfg_size,
|
2008-01-30 05:31:20 -07:00
|
|
|
const struct ptrace_bts_config __user *ucfg)
|
|
|
|
{
|
|
|
|
struct ptrace_bts_config cfg;
|
2008-01-30 05:32:03 -07:00
|
|
|
int bts_size, ret = 0;
|
2008-01-30 05:31:20 -07:00
|
|
|
void *ds;
|
|
|
|
|
2008-01-30 05:32:03 -07:00
|
|
|
if (cfg_size < sizeof(cfg))
|
|
|
|
return -EIO;
|
|
|
|
|
2008-01-30 05:31:20 -07:00
|
|
|
if (copy_from_user(&cfg, ucfg, sizeof(cfg)))
|
|
|
|
return -EFAULT;
|
|
|
|
|
2008-01-30 05:32:03 -07:00
|
|
|
if ((int)cfg.size < 0)
|
|
|
|
return -EINVAL;
|
|
|
|
|
2008-01-30 05:31:20 -07:00
|
|
|
bts_size = 0;
|
|
|
|
ds = (void *)child->thread.ds_area_msr;
|
|
|
|
if (ds) {
|
|
|
|
bts_size = ds_get_bts_size(ds);
|
|
|
|
if (bts_size < 0)
|
|
|
|
return bts_size;
|
|
|
|
}
|
2008-01-30 05:32:03 -07:00
|
|
|
cfg.size = PAGE_ALIGN(cfg.size);
|
2008-01-30 05:31:20 -07:00
|
|
|
|
|
|
|
if (bts_size != cfg.size) {
|
2008-01-30 05:32:03 -07:00
|
|
|
ret = ptrace_bts_realloc(child, cfg.size,
|
|
|
|
cfg.flags & PTRACE_BTS_O_CUT_SIZE);
|
2008-01-30 05:31:20 -07:00
|
|
|
if (ret < 0)
|
2008-01-30 05:32:03 -07:00
|
|
|
goto errout;
|
2008-01-30 05:31:20 -07:00
|
|
|
|
|
|
|
ds = (void *)child->thread.ds_area_msr;
|
|
|
|
}
|
|
|
|
|
2008-01-30 05:32:03 -07:00
|
|
|
if (cfg.flags & PTRACE_BTS_O_SIGNAL)
|
|
|
|
ret = ds_set_overflow(ds, DS_O_SIGNAL);
|
|
|
|
else
|
|
|
|
ret = ds_set_overflow(ds, DS_O_WRAP);
|
|
|
|
if (ret < 0)
|
|
|
|
goto errout;
|
2008-01-30 05:31:09 -07:00
|
|
|
|
2008-01-30 05:32:03 -07:00
|
|
|
if (cfg.flags & PTRACE_BTS_O_TRACE)
|
|
|
|
child->thread.debugctlmsr |= ds_debugctl_mask();
|
|
|
|
else
|
|
|
|
child->thread.debugctlmsr &= ~ds_debugctl_mask();
|
2008-01-30 05:31:09 -07:00
|
|
|
|
2008-01-30 05:32:03 -07:00
|
|
|
if (cfg.flags & PTRACE_BTS_O_SCHED)
|
2008-01-30 05:31:09 -07:00
|
|
|
set_tsk_thread_flag(child, TIF_BTS_TRACE_TS);
|
|
|
|
else
|
|
|
|
clear_tsk_thread_flag(child, TIF_BTS_TRACE_TS);
|
|
|
|
|
2008-01-30 05:32:03 -07:00
|
|
|
ret = sizeof(cfg);
|
|
|
|
|
2008-01-30 05:32:03 -07:00
|
|
|
out:
|
|
|
|
if (child->thread.debugctlmsr)
|
|
|
|
set_tsk_thread_flag(child, TIF_DEBUGCTLMSR);
|
|
|
|
else
|
|
|
|
clear_tsk_thread_flag(child, TIF_DEBUGCTLMSR);
|
|
|
|
|
|
|
|
return ret;
|
|
|
|
|
|
|
|
errout:
|
|
|
|
child->thread.debugctlmsr &= ~ds_debugctl_mask();
|
|
|
|
clear_tsk_thread_flag(child, TIF_BTS_TRACE_TS);
|
|
|
|
goto out;
|
2008-01-30 05:31:09 -07:00
|
|
|
}
|
|
|
|
|
2008-01-30 05:31:20 -07:00
|
|
|
static int ptrace_bts_status(struct task_struct *child,
|
2008-01-30 05:32:03 -07:00
|
|
|
long cfg_size,
|
2008-01-30 05:31:20 -07:00
|
|
|
struct ptrace_bts_config __user *ucfg)
|
2008-01-30 05:31:09 -07:00
|
|
|
{
|
2008-01-30 05:31:20 -07:00
|
|
|
void *ds = (void *)child->thread.ds_area_msr;
|
|
|
|
struct ptrace_bts_config cfg;
|
2008-01-30 05:31:09 -07:00
|
|
|
|
2008-01-30 05:32:03 -07:00
|
|
|
if (cfg_size < sizeof(cfg))
|
|
|
|
return -EIO;
|
|
|
|
|
2008-01-30 05:31:20 -07:00
|
|
|
memset(&cfg, 0, sizeof(cfg));
|
2008-01-30 05:31:09 -07:00
|
|
|
|
2008-01-30 05:31:20 -07:00
|
|
|
if (ds) {
|
|
|
|
cfg.size = ds_get_bts_size(ds);
|
2008-01-30 05:31:09 -07:00
|
|
|
|
2008-01-30 05:31:20 -07:00
|
|
|
if (ds_get_overflow(ds) == DS_O_SIGNAL)
|
|
|
|
cfg.flags |= PTRACE_BTS_O_SIGNAL;
|
2008-01-30 05:31:09 -07:00
|
|
|
|
2008-01-30 05:31:20 -07:00
|
|
|
if (test_tsk_thread_flag(child, TIF_DEBUGCTLMSR) &&
|
|
|
|
child->thread.debugctlmsr & ds_debugctl_mask())
|
|
|
|
cfg.flags |= PTRACE_BTS_O_TRACE;
|
2008-01-30 05:31:09 -07:00
|
|
|
|
2008-01-30 05:31:20 -07:00
|
|
|
if (test_tsk_thread_flag(child, TIF_BTS_TRACE_TS))
|
|
|
|
cfg.flags |= PTRACE_BTS_O_SCHED;
|
2008-01-30 05:31:09 -07:00
|
|
|
}
|
|
|
|
|
2008-01-30 05:32:54 -07:00
|
|
|
cfg.bts_size = sizeof(struct bts_struct);
|
|
|
|
|
2008-01-30 05:31:20 -07:00
|
|
|
if (copy_to_user(ucfg, &cfg, sizeof(cfg)))
|
|
|
|
return -EFAULT;
|
2008-01-30 05:31:09 -07:00
|
|
|
|
2008-01-30 05:31:20 -07:00
|
|
|
return sizeof(cfg);
|
2008-01-30 05:31:09 -07:00
|
|
|
}
|
|
|
|
|
2008-03-04 16:05:39 -07:00
|
|
|
|
|
|
|
static int ptrace_bts_write_record(struct task_struct *child,
|
|
|
|
const struct bts_struct *in)
|
|
|
|
{
|
|
|
|
int retval;
|
|
|
|
|
|
|
|
if (!child->thread.ds_area_msr)
|
|
|
|
return -ENXIO;
|
|
|
|
|
|
|
|
retval = ds_write_bts((void *)child->thread.ds_area_msr, in);
|
|
|
|
if (retval)
|
|
|
|
return retval;
|
|
|
|
|
|
|
|
return sizeof(*in);
|
|
|
|
}
|
|
|
|
|
|
|
|
static int ptrace_bts_realloc(struct task_struct *child,
|
|
|
|
int size, int reduce_size)
|
|
|
|
{
|
|
|
|
unsigned long rlim, vm;
|
|
|
|
int ret, old_size;
|
|
|
|
|
|
|
|
if (size < 0)
|
|
|
|
return -EINVAL;
|
|
|
|
|
|
|
|
old_size = ds_get_bts_size((void *)child->thread.ds_area_msr);
|
|
|
|
if (old_size < 0)
|
|
|
|
return old_size;
|
|
|
|
|
|
|
|
ret = ds_free((void **)&child->thread.ds_area_msr);
|
|
|
|
if (ret < 0)
|
|
|
|
goto out;
|
|
|
|
|
|
|
|
size >>= PAGE_SHIFT;
|
|
|
|
old_size >>= PAGE_SHIFT;
|
|
|
|
|
|
|
|
current->mm->total_vm -= old_size;
|
|
|
|
current->mm->locked_vm -= old_size;
|
|
|
|
|
|
|
|
if (size == 0)
|
|
|
|
goto out;
|
|
|
|
|
|
|
|
rlim = current->signal->rlim[RLIMIT_AS].rlim_cur >> PAGE_SHIFT;
|
|
|
|
vm = current->mm->total_vm + size;
|
|
|
|
if (rlim < vm) {
|
|
|
|
ret = -ENOMEM;
|
|
|
|
|
|
|
|
if (!reduce_size)
|
|
|
|
goto out;
|
|
|
|
|
|
|
|
size = rlim - current->mm->total_vm;
|
|
|
|
if (size <= 0)
|
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
|
|
|
|
rlim = current->signal->rlim[RLIMIT_MEMLOCK].rlim_cur >> PAGE_SHIFT;
|
|
|
|
vm = current->mm->locked_vm + size;
|
|
|
|
if (rlim < vm) {
|
|
|
|
ret = -ENOMEM;
|
|
|
|
|
|
|
|
if (!reduce_size)
|
|
|
|
goto out;
|
|
|
|
|
|
|
|
size = rlim - current->mm->locked_vm;
|
|
|
|
if (size <= 0)
|
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
|
|
|
|
ret = ds_allocate((void **)&child->thread.ds_area_msr,
|
|
|
|
size << PAGE_SHIFT);
|
|
|
|
if (ret < 0)
|
|
|
|
goto out;
|
|
|
|
|
|
|
|
current->mm->total_vm += size;
|
|
|
|
current->mm->locked_vm += size;
|
|
|
|
|
|
|
|
out:
|
|
|
|
if (child->thread.ds_area_msr)
|
|
|
|
set_tsk_thread_flag(child, TIF_DS_AREA_MSR);
|
|
|
|
else
|
|
|
|
clear_tsk_thread_flag(child, TIF_DS_AREA_MSR);
|
|
|
|
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
2008-01-30 05:31:09 -07:00
|
|
|
void ptrace_bts_take_timestamp(struct task_struct *tsk,
|
|
|
|
enum bts_qualifier qualifier)
|
|
|
|
{
|
|
|
|
struct bts_struct rec = {
|
|
|
|
.qualifier = qualifier,
|
2008-01-30 05:32:03 -07:00
|
|
|
.variant.jiffies = jiffies_64
|
2008-01-30 05:31:09 -07:00
|
|
|
};
|
|
|
|
|
|
|
|
ptrace_bts_write_record(tsk, &rec);
|
|
|
|
}
|
2008-02-26 01:40:27 -07:00
|
|
|
#endif /* X86_BTS */
|
2008-01-30 05:31:09 -07:00
|
|
|
|
2005-04-16 15:20:36 -07:00
|
|
|
/*
|
|
|
|
* Called by kernel/ptrace.c when detaching..
|
|
|
|
*
|
|
|
|
* Make sure the single step bit is not set.
|
|
|
|
*/
|
|
|
|
void ptrace_disable(struct task_struct *child)
|
2008-01-30 05:30:58 -07:00
|
|
|
{
|
2008-01-30 05:30:48 -07:00
|
|
|
user_disable_single_step(child);
|
2008-01-30 05:31:01 -07:00
|
|
|
#ifdef TIF_SYSCALL_EMU
|
2005-09-03 15:57:21 -07:00
|
|
|
clear_tsk_thread_flag(child, TIF_SYSCALL_EMU);
|
2008-01-30 05:31:01 -07:00
|
|
|
#endif
|
2008-01-30 05:31:09 -07:00
|
|
|
if (child->thread.ds_area_msr) {
|
2008-02-26 01:40:27 -07:00
|
|
|
#ifdef X86_BTS
|
2008-01-30 05:32:03 -07:00
|
|
|
ptrace_bts_realloc(child, 0, 0);
|
2008-02-26 01:40:27 -07:00
|
|
|
#endif
|
2008-01-30 05:32:03 -07:00
|
|
|
child->thread.debugctlmsr &= ~ds_debugctl_mask();
|
|
|
|
if (!child->thread.debugctlmsr)
|
|
|
|
clear_tsk_thread_flag(child, TIF_DEBUGCTLMSR);
|
|
|
|
clear_tsk_thread_flag(child, TIF_BTS_TRACE_TS);
|
2008-01-30 05:31:09 -07:00
|
|
|
}
|
2005-04-16 15:20:36 -07:00
|
|
|
}
|
|
|
|
|
2008-01-30 05:31:54 -07:00
|
|
|
#if defined CONFIG_X86_32 || defined CONFIG_IA32_EMULATION
|
|
|
|
static const struct user_regset_view user_x86_32_view; /* Initialized below. */
|
|
|
|
#endif
|
|
|
|
|
2005-11-07 01:59:47 -07:00
|
|
|
long arch_ptrace(struct task_struct *child, long request, long addr, long data)
|
2005-04-16 15:20:36 -07:00
|
|
|
{
|
2008-01-30 05:31:54 -07:00
|
|
|
int ret;
|
2005-04-16 15:20:36 -07:00
|
|
|
unsigned long __user *datap = (unsigned long __user *)data;
|
|
|
|
|
|
|
|
switch (request) {
|
|
|
|
/* read the word at location addr in the USER area. */
|
|
|
|
case PTRACE_PEEKUSR: {
|
|
|
|
unsigned long tmp;
|
|
|
|
|
|
|
|
ret = -EIO;
|
2008-01-30 05:31:01 -07:00
|
|
|
if ((addr & (sizeof(data) - 1)) || addr < 0 ||
|
|
|
|
addr >= sizeof(struct user))
|
2005-04-16 15:20:36 -07:00
|
|
|
break;
|
|
|
|
|
|
|
|
tmp = 0; /* Default return condition */
|
2008-01-30 05:31:01 -07:00
|
|
|
if (addr < sizeof(struct user_regs_struct))
|
2005-04-16 15:20:36 -07:00
|
|
|
tmp = getreg(child, addr);
|
2008-01-30 05:31:01 -07:00
|
|
|
else if (addr >= offsetof(struct user, u_debugreg[0]) &&
|
|
|
|
addr <= offsetof(struct user, u_debugreg[7])) {
|
|
|
|
addr -= offsetof(struct user, u_debugreg[0]);
|
|
|
|
tmp = ptrace_get_debugreg(child, addr / sizeof(data));
|
2005-04-16 15:20:36 -07:00
|
|
|
}
|
|
|
|
ret = put_user(tmp, datap);
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
case PTRACE_POKEUSR: /* write the word at location addr in the USER area */
|
|
|
|
ret = -EIO;
|
2008-01-30 05:31:01 -07:00
|
|
|
if ((addr & (sizeof(data) - 1)) || addr < 0 ||
|
|
|
|
addr >= sizeof(struct user))
|
2005-04-16 15:20:36 -07:00
|
|
|
break;
|
|
|
|
|
2008-01-30 05:31:01 -07:00
|
|
|
if (addr < sizeof(struct user_regs_struct))
|
2005-04-16 15:20:36 -07:00
|
|
|
ret = putreg(child, addr, data);
|
2008-01-30 05:31:01 -07:00
|
|
|
else if (addr >= offsetof(struct user, u_debugreg[0]) &&
|
|
|
|
addr <= offsetof(struct user, u_debugreg[7])) {
|
|
|
|
addr -= offsetof(struct user, u_debugreg[0]);
|
|
|
|
ret = ptrace_set_debugreg(child,
|
|
|
|
addr / sizeof(data), data);
|
2005-04-16 15:20:36 -07:00
|
|
|
}
|
2008-01-30 05:31:01 -07:00
|
|
|
break;
|
2005-04-16 15:20:36 -07:00
|
|
|
|
2008-01-30 05:31:54 -07:00
|
|
|
case PTRACE_GETREGS: /* Get all gp regs from the child. */
|
|
|
|
return copy_regset_to_user(child,
|
|
|
|
task_user_regset_view(current),
|
|
|
|
REGSET_GENERAL,
|
|
|
|
0, sizeof(struct user_regs_struct),
|
|
|
|
datap);
|
|
|
|
|
|
|
|
case PTRACE_SETREGS: /* Set all gp regs in the child. */
|
|
|
|
return copy_regset_from_user(child,
|
|
|
|
task_user_regset_view(current),
|
|
|
|
REGSET_GENERAL,
|
|
|
|
0, sizeof(struct user_regs_struct),
|
|
|
|
datap);
|
|
|
|
|
|
|
|
case PTRACE_GETFPREGS: /* Get the child FPU state. */
|
|
|
|
return copy_regset_to_user(child,
|
|
|
|
task_user_regset_view(current),
|
|
|
|
REGSET_FP,
|
|
|
|
0, sizeof(struct user_i387_struct),
|
|
|
|
datap);
|
|
|
|
|
|
|
|
case PTRACE_SETFPREGS: /* Set the child FPU state. */
|
|
|
|
return copy_regset_from_user(child,
|
|
|
|
task_user_regset_view(current),
|
|
|
|
REGSET_FP,
|
|
|
|
0, sizeof(struct user_i387_struct),
|
|
|
|
datap);
|
2005-04-16 15:20:36 -07:00
|
|
|
|
2008-01-30 05:31:01 -07:00
|
|
|
#ifdef CONFIG_X86_32
|
2008-01-30 05:31:54 -07:00
|
|
|
case PTRACE_GETFPXREGS: /* Get the child extended FPU state. */
|
|
|
|
return copy_regset_to_user(child, &user_x86_32_view,
|
|
|
|
REGSET_XFP,
|
|
|
|
0, sizeof(struct user_fxsr_struct),
|
|
|
|
datap);
|
|
|
|
|
|
|
|
case PTRACE_SETFPXREGS: /* Set the child extended FPU state. */
|
|
|
|
return copy_regset_from_user(child, &user_x86_32_view,
|
|
|
|
REGSET_XFP,
|
|
|
|
0, sizeof(struct user_fxsr_struct),
|
|
|
|
datap);
|
2008-01-30 05:31:01 -07:00
|
|
|
#endif
|
2005-04-16 15:20:36 -07:00
|
|
|
|
2008-01-30 05:31:01 -07:00
|
|
|
#if defined CONFIG_X86_32 || defined CONFIG_IA32_EMULATION
|
2005-04-16 15:20:36 -07:00
|
|
|
case PTRACE_GET_THREAD_AREA:
|
2008-01-30 05:30:46 -07:00
|
|
|
if (addr < 0)
|
|
|
|
return -EIO;
|
|
|
|
ret = do_get_thread_area(child, addr,
|
|
|
|
(struct user_desc __user *) data);
|
2005-04-16 15:20:36 -07:00
|
|
|
break;
|
|
|
|
|
|
|
|
case PTRACE_SET_THREAD_AREA:
|
2008-01-30 05:30:46 -07:00
|
|
|
if (addr < 0)
|
|
|
|
return -EIO;
|
|
|
|
ret = do_set_thread_area(child, addr,
|
|
|
|
(struct user_desc __user *) data, 0);
|
2005-04-16 15:20:36 -07:00
|
|
|
break;
|
2008-01-30 05:31:01 -07:00
|
|
|
#endif
|
|
|
|
|
|
|
|
#ifdef CONFIG_X86_64
|
|
|
|
/* normal 64bit interface to access TLS data.
|
|
|
|
Works just like arch_prctl, except that the arguments
|
|
|
|
are reversed. */
|
|
|
|
case PTRACE_ARCH_PRCTL:
|
|
|
|
ret = do_arch_prctl(child, data, addr);
|
|
|
|
break;
|
|
|
|
#endif
|
2005-04-16 15:20:36 -07:00
|
|
|
|
2008-02-26 01:40:27 -07:00
|
|
|
/*
|
|
|
|
* These bits need more cooking - not enabled yet:
|
|
|
|
*/
|
|
|
|
#ifdef X86_BTS
|
2008-01-30 05:31:20 -07:00
|
|
|
case PTRACE_BTS_CONFIG:
|
|
|
|
ret = ptrace_bts_config
|
2008-01-30 05:32:03 -07:00
|
|
|
(child, data, (struct ptrace_bts_config __user *)addr);
|
2008-01-30 05:31:09 -07:00
|
|
|
break;
|
|
|
|
|
2008-01-30 05:31:20 -07:00
|
|
|
case PTRACE_BTS_STATUS:
|
|
|
|
ret = ptrace_bts_status
|
2008-01-30 05:32:03 -07:00
|
|
|
(child, data, (struct ptrace_bts_config __user *)addr);
|
2008-01-30 05:31:09 -07:00
|
|
|
break;
|
|
|
|
|
2008-01-30 05:31:20 -07:00
|
|
|
case PTRACE_BTS_SIZE:
|
|
|
|
ret = ptrace_bts_get_size(child);
|
2008-01-30 05:31:09 -07:00
|
|
|
break;
|
|
|
|
|
2008-01-30 05:31:20 -07:00
|
|
|
case PTRACE_BTS_GET:
|
2008-01-30 05:31:09 -07:00
|
|
|
ret = ptrace_bts_read_record
|
2008-01-30 05:31:20 -07:00
|
|
|
(child, data, (struct bts_struct __user *) addr);
|
2008-01-30 05:31:09 -07:00
|
|
|
break;
|
|
|
|
|
2008-01-30 05:31:20 -07:00
|
|
|
case PTRACE_BTS_CLEAR:
|
|
|
|
ret = ptrace_bts_clear(child);
|
2008-01-30 05:31:09 -07:00
|
|
|
break;
|
|
|
|
|
2008-01-30 05:31:20 -07:00
|
|
|
case PTRACE_BTS_DRAIN:
|
|
|
|
ret = ptrace_bts_drain
|
2008-01-30 05:32:03 -07:00
|
|
|
(child, data, (struct bts_struct __user *) addr);
|
2008-01-30 05:31:09 -07:00
|
|
|
break;
|
2008-02-26 01:40:27 -07:00
|
|
|
#endif
|
2008-01-30 05:31:09 -07:00
|
|
|
|
2005-04-16 15:20:36 -07:00
|
|
|
default:
|
|
|
|
ret = ptrace_request(child, request, addr, data);
|
|
|
|
break;
|
|
|
|
}
|
2008-01-30 05:30:52 -07:00
|
|
|
|
2005-04-16 15:20:36 -07:00
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
2008-01-30 05:31:01 -07:00
|
|
|
#ifdef CONFIG_IA32_EMULATION
|
|
|
|
|
2008-01-30 05:31:01 -07:00
|
|
|
#include <linux/compat.h>
|
|
|
|
#include <linux/syscalls.h>
|
|
|
|
#include <asm/ia32.h>
|
2008-01-30 05:31:01 -07:00
|
|
|
#include <asm/user32.h>
|
|
|
|
|
|
|
|
#define R32(l,q) \
|
|
|
|
case offsetof(struct user32, regs.l): \
|
|
|
|
regs->q = value; break
|
|
|
|
|
|
|
|
#define SEG32(rs) \
|
|
|
|
case offsetof(struct user32, regs.rs): \
|
|
|
|
return set_segment_reg(child, \
|
|
|
|
offsetof(struct user_regs_struct, rs), \
|
|
|
|
value); \
|
|
|
|
break
|
|
|
|
|
|
|
|
static int putreg32(struct task_struct *child, unsigned regno, u32 value)
|
|
|
|
{
|
|
|
|
struct pt_regs *regs = task_pt_regs(child);
|
|
|
|
|
|
|
|
switch (regno) {
|
|
|
|
|
|
|
|
SEG32(cs);
|
|
|
|
SEG32(ds);
|
|
|
|
SEG32(es);
|
|
|
|
SEG32(fs);
|
|
|
|
SEG32(gs);
|
|
|
|
SEG32(ss);
|
|
|
|
|
|
|
|
R32(ebx, bx);
|
|
|
|
R32(ecx, cx);
|
|
|
|
R32(edx, dx);
|
|
|
|
R32(edi, di);
|
|
|
|
R32(esi, si);
|
|
|
|
R32(ebp, bp);
|
|
|
|
R32(eax, ax);
|
|
|
|
R32(eip, ip);
|
|
|
|
R32(esp, sp);
|
|
|
|
|
2008-02-28 20:57:07 -07:00
|
|
|
case offsetof(struct user32, regs.orig_eax):
|
|
|
|
/*
|
|
|
|
* Sign-extend the value so that orig_eax = -1
|
|
|
|
* causes (long)orig_ax < 0 tests to fire correctly.
|
|
|
|
*/
|
|
|
|
regs->orig_ax = (long) (s32) value;
|
|
|
|
break;
|
|
|
|
|
2008-01-30 05:31:01 -07:00
|
|
|
case offsetof(struct user32, regs.eflags):
|
|
|
|
return set_flags(child, value);
|
|
|
|
|
|
|
|
case offsetof(struct user32, u_debugreg[0]) ...
|
|
|
|
offsetof(struct user32, u_debugreg[7]):
|
|
|
|
regno -= offsetof(struct user32, u_debugreg[0]);
|
|
|
|
return ptrace_set_debugreg(child, regno / 4, value);
|
|
|
|
|
|
|
|
default:
|
|
|
|
if (regno > sizeof(struct user32) || (regno & 3))
|
|
|
|
return -EIO;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Other dummy fields in the virtual user structure
|
|
|
|
* are ignored
|
|
|
|
*/
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
#undef R32
|
|
|
|
#undef SEG32
|
|
|
|
|
|
|
|
#define R32(l,q) \
|
|
|
|
case offsetof(struct user32, regs.l): \
|
|
|
|
*val = regs->q; break
|
|
|
|
|
|
|
|
#define SEG32(rs) \
|
|
|
|
case offsetof(struct user32, regs.rs): \
|
|
|
|
*val = get_segment_reg(child, \
|
|
|
|
offsetof(struct user_regs_struct, rs)); \
|
|
|
|
break
|
|
|
|
|
|
|
|
static int getreg32(struct task_struct *child, unsigned regno, u32 *val)
|
|
|
|
{
|
|
|
|
struct pt_regs *regs = task_pt_regs(child);
|
|
|
|
|
|
|
|
switch (regno) {
|
|
|
|
|
|
|
|
SEG32(ds);
|
|
|
|
SEG32(es);
|
|
|
|
SEG32(fs);
|
|
|
|
SEG32(gs);
|
|
|
|
|
|
|
|
R32(cs, cs);
|
|
|
|
R32(ss, ss);
|
|
|
|
R32(ebx, bx);
|
|
|
|
R32(ecx, cx);
|
|
|
|
R32(edx, dx);
|
|
|
|
R32(edi, di);
|
|
|
|
R32(esi, si);
|
|
|
|
R32(ebp, bp);
|
|
|
|
R32(eax, ax);
|
|
|
|
R32(orig_eax, orig_ax);
|
|
|
|
R32(eip, ip);
|
|
|
|
R32(esp, sp);
|
|
|
|
|
|
|
|
case offsetof(struct user32, regs.eflags):
|
|
|
|
*val = get_flags(child);
|
|
|
|
break;
|
|
|
|
|
|
|
|
case offsetof(struct user32, u_debugreg[0]) ...
|
|
|
|
offsetof(struct user32, u_debugreg[7]):
|
|
|
|
regno -= offsetof(struct user32, u_debugreg[0]);
|
|
|
|
*val = ptrace_get_debugreg(child, regno / 4);
|
|
|
|
break;
|
|
|
|
|
|
|
|
default:
|
|
|
|
if (regno > sizeof(struct user32) || (regno & 3))
|
|
|
|
return -EIO;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Other dummy fields in the virtual user structure
|
|
|
|
* are ignored
|
|
|
|
*/
|
|
|
|
*val = 0;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
#undef R32
|
|
|
|
#undef SEG32
|
|
|
|
|
2008-01-30 05:31:52 -07:00
|
|
|
static int genregs32_get(struct task_struct *target,
|
|
|
|
const struct user_regset *regset,
|
|
|
|
unsigned int pos, unsigned int count,
|
|
|
|
void *kbuf, void __user *ubuf)
|
|
|
|
{
|
|
|
|
if (kbuf) {
|
|
|
|
compat_ulong_t *k = kbuf;
|
|
|
|
while (count > 0) {
|
|
|
|
getreg32(target, pos, k++);
|
|
|
|
count -= sizeof(*k);
|
|
|
|
pos += sizeof(*k);
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
compat_ulong_t __user *u = ubuf;
|
|
|
|
while (count > 0) {
|
|
|
|
compat_ulong_t word;
|
|
|
|
getreg32(target, pos, &word);
|
|
|
|
if (__put_user(word, u++))
|
|
|
|
return -EFAULT;
|
|
|
|
count -= sizeof(*u);
|
|
|
|
pos += sizeof(*u);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int genregs32_set(struct task_struct *target,
|
|
|
|
const struct user_regset *regset,
|
|
|
|
unsigned int pos, unsigned int count,
|
|
|
|
const void *kbuf, const void __user *ubuf)
|
|
|
|
{
|
|
|
|
int ret = 0;
|
|
|
|
if (kbuf) {
|
|
|
|
const compat_ulong_t *k = kbuf;
|
|
|
|
while (count > 0 && !ret) {
|
2008-02-21 21:37:24 -07:00
|
|
|
ret = putreg32(target, pos, *k++);
|
2008-01-30 05:31:52 -07:00
|
|
|
count -= sizeof(*k);
|
|
|
|
pos += sizeof(*k);
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
const compat_ulong_t __user *u = ubuf;
|
|
|
|
while (count > 0 && !ret) {
|
|
|
|
compat_ulong_t word;
|
|
|
|
ret = __get_user(word, u++);
|
|
|
|
if (ret)
|
|
|
|
break;
|
2008-02-21 21:37:24 -07:00
|
|
|
ret = putreg32(target, pos, word);
|
2008-01-30 05:31:52 -07:00
|
|
|
count -= sizeof(*u);
|
|
|
|
pos += sizeof(*u);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
2008-01-30 05:31:01 -07:00
|
|
|
static long ptrace32_siginfo(unsigned request, u32 pid, u32 addr, u32 data)
|
|
|
|
{
|
|
|
|
siginfo_t __user *si = compat_alloc_user_space(sizeof(siginfo_t));
|
|
|
|
compat_siginfo_t __user *si32 = compat_ptr(data);
|
|
|
|
siginfo_t ssi;
|
|
|
|
int ret;
|
|
|
|
|
|
|
|
if (request == PTRACE_SETSIGINFO) {
|
|
|
|
memset(&ssi, 0, sizeof(siginfo_t));
|
|
|
|
ret = copy_siginfo_from_user32(&ssi, si32);
|
|
|
|
if (ret)
|
|
|
|
return ret;
|
|
|
|
if (copy_to_user(si, &ssi, sizeof(siginfo_t)))
|
|
|
|
return -EFAULT;
|
|
|
|
}
|
|
|
|
ret = sys_ptrace(request, pid, addr, (unsigned long)si);
|
|
|
|
if (ret)
|
|
|
|
return ret;
|
|
|
|
if (request == PTRACE_GETSIGINFO) {
|
|
|
|
if (copy_from_user(&ssi, si, sizeof(siginfo_t)))
|
|
|
|
return -EFAULT;
|
|
|
|
ret = copy_siginfo_to_user32(si32, &ssi);
|
|
|
|
}
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
|
|
|
asmlinkage long sys32_ptrace(long request, u32 pid, u32 addr, u32 data)
|
|
|
|
{
|
|
|
|
struct task_struct *child;
|
|
|
|
struct pt_regs *childregs;
|
|
|
|
void __user *datap = compat_ptr(data);
|
|
|
|
int ret;
|
|
|
|
__u32 val;
|
|
|
|
|
|
|
|
switch (request) {
|
|
|
|
case PTRACE_TRACEME:
|
|
|
|
case PTRACE_ATTACH:
|
|
|
|
case PTRACE_KILL:
|
|
|
|
case PTRACE_CONT:
|
|
|
|
case PTRACE_SINGLESTEP:
|
|
|
|
case PTRACE_SINGLEBLOCK:
|
|
|
|
case PTRACE_DETACH:
|
|
|
|
case PTRACE_SYSCALL:
|
|
|
|
case PTRACE_OLDSETOPTIONS:
|
|
|
|
case PTRACE_SETOPTIONS:
|
|
|
|
case PTRACE_SET_THREAD_AREA:
|
|
|
|
case PTRACE_GET_THREAD_AREA:
|
2008-02-26 01:40:27 -07:00
|
|
|
#ifdef X86_BTS
|
2008-01-30 05:31:09 -07:00
|
|
|
case PTRACE_BTS_CONFIG:
|
|
|
|
case PTRACE_BTS_STATUS:
|
2008-01-30 05:31:20 -07:00
|
|
|
case PTRACE_BTS_SIZE:
|
|
|
|
case PTRACE_BTS_GET:
|
|
|
|
case PTRACE_BTS_CLEAR:
|
|
|
|
case PTRACE_BTS_DRAIN:
|
2008-02-26 01:40:27 -07:00
|
|
|
#endif
|
2008-01-30 05:31:01 -07:00
|
|
|
return sys_ptrace(request, pid, addr, data);
|
|
|
|
|
|
|
|
default:
|
|
|
|
return -EINVAL;
|
|
|
|
|
|
|
|
case PTRACE_PEEKTEXT:
|
|
|
|
case PTRACE_PEEKDATA:
|
|
|
|
case PTRACE_POKEDATA:
|
|
|
|
case PTRACE_POKETEXT:
|
|
|
|
case PTRACE_POKEUSR:
|
|
|
|
case PTRACE_PEEKUSR:
|
|
|
|
case PTRACE_GETREGS:
|
|
|
|
case PTRACE_SETREGS:
|
|
|
|
case PTRACE_SETFPREGS:
|
|
|
|
case PTRACE_GETFPREGS:
|
|
|
|
case PTRACE_SETFPXREGS:
|
|
|
|
case PTRACE_GETFPXREGS:
|
|
|
|
case PTRACE_GETEVENTMSG:
|
|
|
|
break;
|
|
|
|
|
|
|
|
case PTRACE_SETSIGINFO:
|
|
|
|
case PTRACE_GETSIGINFO:
|
|
|
|
return ptrace32_siginfo(request, pid, addr, data);
|
|
|
|
}
|
|
|
|
|
|
|
|
child = ptrace_get_task_struct(pid);
|
|
|
|
if (IS_ERR(child))
|
|
|
|
return PTR_ERR(child);
|
|
|
|
|
|
|
|
ret = ptrace_check_attach(child, request == PTRACE_KILL);
|
|
|
|
if (ret < 0)
|
|
|
|
goto out;
|
|
|
|
|
|
|
|
childregs = task_pt_regs(child);
|
|
|
|
|
|
|
|
switch (request) {
|
|
|
|
case PTRACE_PEEKUSR:
|
|
|
|
ret = getreg32(child, addr, &val);
|
|
|
|
if (ret == 0)
|
|
|
|
ret = put_user(val, (__u32 __user *)datap);
|
|
|
|
break;
|
|
|
|
|
|
|
|
case PTRACE_POKEUSR:
|
|
|
|
ret = putreg32(child, addr, data);
|
|
|
|
break;
|
|
|
|
|
2008-01-30 05:31:54 -07:00
|
|
|
case PTRACE_GETREGS: /* Get all gp regs from the child. */
|
|
|
|
return copy_regset_to_user(child, &user_x86_32_view,
|
|
|
|
REGSET_GENERAL,
|
|
|
|
0, sizeof(struct user_regs_struct32),
|
|
|
|
datap);
|
|
|
|
|
|
|
|
case PTRACE_SETREGS: /* Set all gp regs in the child. */
|
|
|
|
return copy_regset_from_user(child, &user_x86_32_view,
|
|
|
|
REGSET_GENERAL, 0,
|
|
|
|
sizeof(struct user_regs_struct32),
|
|
|
|
datap);
|
|
|
|
|
|
|
|
case PTRACE_GETFPREGS: /* Get the child FPU state. */
|
|
|
|
return copy_regset_to_user(child, &user_x86_32_view,
|
|
|
|
REGSET_FP, 0,
|
|
|
|
sizeof(struct user_i387_ia32_struct),
|
|
|
|
datap);
|
|
|
|
|
|
|
|
case PTRACE_SETFPREGS: /* Set the child FPU state. */
|
|
|
|
return copy_regset_from_user(
|
|
|
|
child, &user_x86_32_view, REGSET_FP,
|
|
|
|
0, sizeof(struct user_i387_ia32_struct), datap);
|
|
|
|
|
|
|
|
case PTRACE_GETFPXREGS: /* Get the child extended FPU state. */
|
|
|
|
return copy_regset_to_user(child, &user_x86_32_view,
|
|
|
|
REGSET_XFP, 0,
|
|
|
|
sizeof(struct user32_fxsr_struct),
|
|
|
|
datap);
|
|
|
|
|
|
|
|
case PTRACE_SETFPXREGS: /* Set the child extended FPU state. */
|
|
|
|
return copy_regset_from_user(child, &user_x86_32_view,
|
|
|
|
REGSET_XFP, 0,
|
|
|
|
sizeof(struct user32_fxsr_struct),
|
|
|
|
datap);
|
2008-01-30 05:31:01 -07:00
|
|
|
|
|
|
|
default:
|
2008-01-30 05:31:56 -07:00
|
|
|
return compat_ptrace_request(child, request, addr, data);
|
2008-01-30 05:31:01 -07:00
|
|
|
}
|
|
|
|
|
|
|
|
out:
|
|
|
|
put_task_struct(child);
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
2008-01-30 05:31:01 -07:00
|
|
|
#endif /* CONFIG_IA32_EMULATION */
|
|
|
|
|
2008-01-30 05:31:53 -07:00
|
|
|
#ifdef CONFIG_X86_64
|
|
|
|
|
|
|
|
static const struct user_regset x86_64_regsets[] = {
|
|
|
|
[REGSET_GENERAL] = {
|
|
|
|
.core_note_type = NT_PRSTATUS,
|
|
|
|
.n = sizeof(struct user_regs_struct) / sizeof(long),
|
|
|
|
.size = sizeof(long), .align = sizeof(long),
|
|
|
|
.get = genregs_get, .set = genregs_set
|
|
|
|
},
|
|
|
|
[REGSET_FP] = {
|
|
|
|
.core_note_type = NT_PRFPREG,
|
|
|
|
.n = sizeof(struct user_i387_struct) / sizeof(long),
|
|
|
|
.size = sizeof(long), .align = sizeof(long),
|
|
|
|
.active = xfpregs_active, .get = xfpregs_get, .set = xfpregs_set
|
|
|
|
},
|
|
|
|
};
|
|
|
|
|
|
|
|
static const struct user_regset_view user_x86_64_view = {
|
|
|
|
.name = "x86_64", .e_machine = EM_X86_64,
|
|
|
|
.regsets = x86_64_regsets, .n = ARRAY_SIZE(x86_64_regsets)
|
|
|
|
};
|
|
|
|
|
|
|
|
#else /* CONFIG_X86_32 */
|
|
|
|
|
|
|
|
#define user_regs_struct32 user_regs_struct
|
|
|
|
#define genregs32_get genregs_get
|
|
|
|
#define genregs32_set genregs_set
|
|
|
|
|
|
|
|
#endif /* CONFIG_X86_64 */
|
|
|
|
|
|
|
|
#if defined CONFIG_X86_32 || defined CONFIG_IA32_EMULATION
|
|
|
|
static const struct user_regset x86_32_regsets[] = {
|
|
|
|
[REGSET_GENERAL] = {
|
|
|
|
.core_note_type = NT_PRSTATUS,
|
|
|
|
.n = sizeof(struct user_regs_struct32) / sizeof(u32),
|
|
|
|
.size = sizeof(u32), .align = sizeof(u32),
|
|
|
|
.get = genregs32_get, .set = genregs32_set
|
|
|
|
},
|
|
|
|
[REGSET_FP] = {
|
|
|
|
.core_note_type = NT_PRFPREG,
|
|
|
|
.n = sizeof(struct user_i387_struct) / sizeof(u32),
|
|
|
|
.size = sizeof(u32), .align = sizeof(u32),
|
|
|
|
.active = fpregs_active, .get = fpregs_get, .set = fpregs_set
|
|
|
|
},
|
|
|
|
[REGSET_XFP] = {
|
|
|
|
.core_note_type = NT_PRXFPREG,
|
|
|
|
.n = sizeof(struct user_i387_struct) / sizeof(u32),
|
|
|
|
.size = sizeof(u32), .align = sizeof(u32),
|
|
|
|
.active = xfpregs_active, .get = xfpregs_get, .set = xfpregs_set
|
|
|
|
},
|
|
|
|
[REGSET_TLS] = {
|
2008-01-30 05:31:56 -07:00
|
|
|
.core_note_type = NT_386_TLS,
|
2008-01-30 05:31:53 -07:00
|
|
|
.n = GDT_ENTRY_TLS_ENTRIES, .bias = GDT_ENTRY_TLS_MIN,
|
|
|
|
.size = sizeof(struct user_desc),
|
|
|
|
.align = sizeof(struct user_desc),
|
|
|
|
.active = regset_tls_active,
|
|
|
|
.get = regset_tls_get, .set = regset_tls_set
|
|
|
|
},
|
|
|
|
};
|
|
|
|
|
|
|
|
static const struct user_regset_view user_x86_32_view = {
|
|
|
|
.name = "i386", .e_machine = EM_386,
|
|
|
|
.regsets = x86_32_regsets, .n = ARRAY_SIZE(x86_32_regsets)
|
|
|
|
};
|
|
|
|
#endif
|
|
|
|
|
|
|
|
const struct user_regset_view *task_user_regset_view(struct task_struct *task)
|
|
|
|
{
|
|
|
|
#ifdef CONFIG_IA32_EMULATION
|
|
|
|
if (test_tsk_thread_flag(task, TIF_IA32))
|
|
|
|
#endif
|
|
|
|
#if defined CONFIG_X86_32 || defined CONFIG_IA32_EMULATION
|
|
|
|
return &user_x86_32_view;
|
|
|
|
#endif
|
|
|
|
#ifdef CONFIG_X86_64
|
|
|
|
return &user_x86_64_view;
|
|
|
|
#endif
|
|
|
|
}
|
|
|
|
|
2008-01-30 05:31:01 -07:00
|
|
|
#ifdef CONFIG_X86_32
|
|
|
|
|
2005-04-16 15:20:36 -07:00
|
|
|
void send_sigtrap(struct task_struct *tsk, struct pt_regs *regs, int error_code)
|
|
|
|
{
|
|
|
|
struct siginfo info;
|
|
|
|
|
|
|
|
tsk->thread.trap_no = 1;
|
|
|
|
tsk->thread.error_code = error_code;
|
|
|
|
|
|
|
|
memset(&info, 0, sizeof(info));
|
|
|
|
info.si_signo = SIGTRAP;
|
|
|
|
info.si_code = TRAP_BRKPT;
|
|
|
|
|
2008-01-30 05:30:56 -07:00
|
|
|
/* User-mode ip? */
|
|
|
|
info.si_addr = user_mode_vm(regs) ? (void __user *) regs->ip : NULL;
|
2005-04-16 15:20:36 -07:00
|
|
|
|
2007-10-19 16:13:56 -07:00
|
|
|
/* Send us the fake SIGTRAP */
|
2005-04-16 15:20:36 -07:00
|
|
|
force_sig_info(SIGTRAP, &info, tsk);
|
|
|
|
}
|
|
|
|
|
|
|
|
/* notification of system call entry/exit
|
|
|
|
* - triggered by current->work.syscall_trace
|
|
|
|
*/
|
[PATCH] UML Support - Ptrace: adds the host SYSEMU support, for UML and general usage
Jeff Dike <jdike@addtoit.com>,
Paolo 'Blaisorblade' Giarrusso <blaisorblade_spam@yahoo.it>,
Bodo Stroesser <bstroesser@fujitsu-siemens.com>
Adds a new ptrace(2) mode, called PTRACE_SYSEMU, resembling PTRACE_SYSCALL
except that the kernel does not execute the requested syscall; this is useful
to improve performance for virtual environments, like UML, which want to run
the syscall on their own.
In fact, using PTRACE_SYSCALL means stopping child execution twice, on entry
and on exit, and each time you also have two context switches; with SYSEMU you
avoid the 2nd stop and so save two context switches per syscall.
Also, some architectures don't have support in the host for changing the
syscall number via ptrace(), which is currently needed to skip syscall
execution (UML turns any syscall into getpid() to avoid it being executed on
the host). Fixing that is hard, while SYSEMU is easier to implement.
* This version of the patch includes some suggestions of Jeff Dike to avoid
adding any instructions to the syscall fast path, plus some other little
changes, by myself, to make it work even when the syscall is executed with
SYSENTER (but I'm unsure about them). It has been widely tested for quite a
lot of time.
* Various fixed were included to handle the various switches between
various states, i.e. when for instance a syscall entry is traced with one of
PT_SYSCALL / _SYSEMU / _SINGLESTEP and another one is used on exit.
Basically, this is done by remembering which one of them was used even after
the call to ptrace_notify().
* We're combining TIF_SYSCALL_EMU with TIF_SYSCALL_TRACE or TIF_SINGLESTEP
to make do_syscall_trace() notice that the current syscall was started with
SYSEMU on entry, so that no notification ought to be done in the exit path;
this is a bit of a hack, so this problem is solved in another way in next
patches.
* Also, the effects of the patch:
"Ptrace - i386: fix Syscall Audit interaction with singlestep"
are cancelled; they are restored back in the last patch of this series.
Detailed descriptions of the patches doing this kind of processing follow (but
I've already summed everything up).
* Fix behaviour when changing interception kind #1.
In do_syscall_trace(), we check the status of the TIF_SYSCALL_EMU flag
only after doing the debugger notification; but the debugger might have
changed the status of this flag because he continued execution with
PTRACE_SYSCALL, so this is wrong. This patch fixes it by saving the flag
status before calling ptrace_notify().
* Fix behaviour when changing interception kind #2:
avoid intercepting syscall on return when using SYSCALL again.
A guest process switching from using PTRACE_SYSEMU to PTRACE_SYSCALL
crashes.
The problem is in arch/i386/kernel/entry.S. The current SYSEMU patch
inhibits the syscall-handler to be called, but does not prevent
do_syscall_trace() to be called after this for syscall completion
interception.
The appended patch fixes this. It reuses the flag TIF_SYSCALL_EMU to
remember "we come from PTRACE_SYSEMU and now are in PTRACE_SYSCALL", since
the flag is unused in the depicted situation.
* Fix behaviour when changing interception kind #3:
avoid intercepting syscall on return when using SINGLESTEP.
When testing 2.6.9 and the skas3.v6 patch, with my latest patch and had
problems with singlestepping on UML in SKAS with SYSEMU. It looped
receiving SIGTRAPs without moving forward. EIP of the traced process was
the same for all SIGTRAPs.
What's missing is to handle switching from PTRACE_SYSCALL_EMU to
PTRACE_SINGLESTEP in a way very similar to what is done for the change from
PTRACE_SYSCALL_EMU to PTRACE_SYSCALL_TRACE.
I.e., after calling ptrace(PTRACE_SYSEMU), on the return path, the debugger is
notified and then wake ups the process; the syscall is executed (or skipped,
when do_syscall_trace() returns 0, i.e. when using PTRACE_SYSEMU), and
do_syscall_trace() is called again. Since we are on the return path of a
SYSEMU'd syscall, if the wake up is performed through ptrace(PTRACE_SYSCALL),
we must still avoid notifying the parent of the syscall exit. Now, this
behaviour is extended even to resuming with PTRACE_SINGLESTEP.
Signed-off-by: Paolo 'Blaisorblade' Giarrusso <blaisorblade@yahoo.it>
Cc: Jeff Dike <jdike@addtoit.com>
Signed-off-by: Andrew Morton <akpm@osdl.org>
Signed-off-by: Linus Torvalds <torvalds@osdl.org>
2005-09-03 15:57:18 -07:00
|
|
|
int do_syscall_trace(struct pt_regs *regs, int entryexit)
|
2005-04-16 15:20:36 -07:00
|
|
|
{
|
2005-09-09 13:01:51 -07:00
|
|
|
int is_sysemu = test_thread_flag(TIF_SYSCALL_EMU);
|
|
|
|
/*
|
|
|
|
* With TIF_SYSCALL_EMU set we want to ignore TIF_SINGLESTEP for syscall
|
|
|
|
* interception
|
|
|
|
*/
|
2005-09-03 15:57:20 -07:00
|
|
|
int is_singlestep = !is_sysemu && test_thread_flag(TIF_SINGLESTEP);
|
2005-09-09 13:01:51 -07:00
|
|
|
int ret = 0;
|
2005-09-03 15:57:20 -07:00
|
|
|
|
2005-04-16 15:20:36 -07:00
|
|
|
/* do the secure computing check first */
|
2005-09-09 13:01:51 -07:00
|
|
|
if (!entryexit)
|
2008-01-30 05:30:56 -07:00
|
|
|
secure_computing(regs->orig_ax);
|
2005-04-16 15:20:36 -07:00
|
|
|
|
2005-09-03 15:57:21 -07:00
|
|
|
if (unlikely(current->audit_context)) {
|
|
|
|
if (entryexit)
|
2008-01-30 05:30:56 -07:00
|
|
|
audit_syscall_exit(AUDITSC_RESULT(regs->ax),
|
|
|
|
regs->ax);
|
2005-09-03 15:57:21 -07:00
|
|
|
/* Debug traps, when using PTRACE_SINGLESTEP, must be sent only
|
|
|
|
* on the syscall exit path. Normally, when TIF_SYSCALL_AUDIT is
|
|
|
|
* not used, entry.S will call us only on syscall exit, not
|
|
|
|
* entry; so when TIF_SYSCALL_AUDIT is used we must avoid
|
|
|
|
* calling send_sigtrap() on syscall entry.
|
|
|
|
*
|
|
|
|
* Note that when PTRACE_SYSEMU_SINGLESTEP is used,
|
|
|
|
* is_singlestep is false, despite his name, so we will still do
|
|
|
|
* the correct thing.
|
|
|
|
*/
|
|
|
|
else if (is_singlestep)
|
|
|
|
goto out;
|
|
|
|
}
|
2005-04-16 15:20:36 -07:00
|
|
|
|
|
|
|
if (!(current->ptrace & PT_PTRACED))
|
2005-04-29 08:08:28 -07:00
|
|
|
goto out;
|
2005-04-16 15:20:36 -07:00
|
|
|
|
2005-09-03 15:57:20 -07:00
|
|
|
/* If a process stops on the 1st tracepoint with SYSCALL_TRACE
|
|
|
|
* and then is resumed with SYSEMU_SINGLESTEP, it will come in
|
|
|
|
* here. We have to check this and return */
|
|
|
|
if (is_sysemu && entryexit)
|
|
|
|
return 0;
|
[PATCH] UML Support - Ptrace: adds the host SYSEMU support, for UML and general usage
Jeff Dike <jdike@addtoit.com>,
Paolo 'Blaisorblade' Giarrusso <blaisorblade_spam@yahoo.it>,
Bodo Stroesser <bstroesser@fujitsu-siemens.com>
Adds a new ptrace(2) mode, called PTRACE_SYSEMU, resembling PTRACE_SYSCALL
except that the kernel does not execute the requested syscall; this is useful
to improve performance for virtual environments, like UML, which want to run
the syscall on their own.
In fact, using PTRACE_SYSCALL means stopping child execution twice, on entry
and on exit, and each time you also have two context switches; with SYSEMU you
avoid the 2nd stop and so save two context switches per syscall.
Also, some architectures don't have support in the host for changing the
syscall number via ptrace(), which is currently needed to skip syscall
execution (UML turns any syscall into getpid() to avoid it being executed on
the host). Fixing that is hard, while SYSEMU is easier to implement.
* This version of the patch includes some suggestions of Jeff Dike to avoid
adding any instructions to the syscall fast path, plus some other little
changes, by myself, to make it work even when the syscall is executed with
SYSENTER (but I'm unsure about them). It has been widely tested for quite a
lot of time.
* Various fixed were included to handle the various switches between
various states, i.e. when for instance a syscall entry is traced with one of
PT_SYSCALL / _SYSEMU / _SINGLESTEP and another one is used on exit.
Basically, this is done by remembering which one of them was used even after
the call to ptrace_notify().
* We're combining TIF_SYSCALL_EMU with TIF_SYSCALL_TRACE or TIF_SINGLESTEP
to make do_syscall_trace() notice that the current syscall was started with
SYSEMU on entry, so that no notification ought to be done in the exit path;
this is a bit of a hack, so this problem is solved in another way in next
patches.
* Also, the effects of the patch:
"Ptrace - i386: fix Syscall Audit interaction with singlestep"
are cancelled; they are restored back in the last patch of this series.
Detailed descriptions of the patches doing this kind of processing follow (but
I've already summed everything up).
* Fix behaviour when changing interception kind #1.
In do_syscall_trace(), we check the status of the TIF_SYSCALL_EMU flag
only after doing the debugger notification; but the debugger might have
changed the status of this flag because he continued execution with
PTRACE_SYSCALL, so this is wrong. This patch fixes it by saving the flag
status before calling ptrace_notify().
* Fix behaviour when changing interception kind #2:
avoid intercepting syscall on return when using SYSCALL again.
A guest process switching from using PTRACE_SYSEMU to PTRACE_SYSCALL
crashes.
The problem is in arch/i386/kernel/entry.S. The current SYSEMU patch
inhibits the syscall-handler to be called, but does not prevent
do_syscall_trace() to be called after this for syscall completion
interception.
The appended patch fixes this. It reuses the flag TIF_SYSCALL_EMU to
remember "we come from PTRACE_SYSEMU and now are in PTRACE_SYSCALL", since
the flag is unused in the depicted situation.
* Fix behaviour when changing interception kind #3:
avoid intercepting syscall on return when using SINGLESTEP.
When testing 2.6.9 and the skas3.v6 patch, with my latest patch and had
problems with singlestepping on UML in SKAS with SYSEMU. It looped
receiving SIGTRAPs without moving forward. EIP of the traced process was
the same for all SIGTRAPs.
What's missing is to handle switching from PTRACE_SYSCALL_EMU to
PTRACE_SINGLESTEP in a way very similar to what is done for the change from
PTRACE_SYSCALL_EMU to PTRACE_SYSCALL_TRACE.
I.e., after calling ptrace(PTRACE_SYSEMU), on the return path, the debugger is
notified and then wake ups the process; the syscall is executed (or skipped,
when do_syscall_trace() returns 0, i.e. when using PTRACE_SYSEMU), and
do_syscall_trace() is called again. Since we are on the return path of a
SYSEMU'd syscall, if the wake up is performed through ptrace(PTRACE_SYSCALL),
we must still avoid notifying the parent of the syscall exit. Now, this
behaviour is extended even to resuming with PTRACE_SINGLESTEP.
Signed-off-by: Paolo 'Blaisorblade' Giarrusso <blaisorblade@yahoo.it>
Cc: Jeff Dike <jdike@addtoit.com>
Signed-off-by: Andrew Morton <akpm@osdl.org>
Signed-off-by: Linus Torvalds <torvalds@osdl.org>
2005-09-03 15:57:18 -07:00
|
|
|
|
2005-04-16 15:20:36 -07:00
|
|
|
/* Fake a debug trap */
|
2005-09-03 15:57:19 -07:00
|
|
|
if (is_singlestep)
|
2005-04-16 15:20:36 -07:00
|
|
|
send_sigtrap(current, regs, 0);
|
|
|
|
|
2005-09-03 15:57:19 -07:00
|
|
|
if (!test_thread_flag(TIF_SYSCALL_TRACE) && !is_sysemu)
|
2005-04-29 08:08:28 -07:00
|
|
|
goto out;
|
2005-04-16 15:20:36 -07:00
|
|
|
|
|
|
|
/* the 0x80 provides a way for the tracing parent to distinguish
|
|
|
|
between a syscall stop and SIGTRAP delivery */
|
[PATCH] UML Support - Ptrace: adds the host SYSEMU support, for UML and general usage
Jeff Dike <jdike@addtoit.com>,
Paolo 'Blaisorblade' Giarrusso <blaisorblade_spam@yahoo.it>,
Bodo Stroesser <bstroesser@fujitsu-siemens.com>
Adds a new ptrace(2) mode, called PTRACE_SYSEMU, resembling PTRACE_SYSCALL
except that the kernel does not execute the requested syscall; this is useful
to improve performance for virtual environments, like UML, which want to run
the syscall on their own.
In fact, using PTRACE_SYSCALL means stopping child execution twice, on entry
and on exit, and each time you also have two context switches; with SYSEMU you
avoid the 2nd stop and so save two context switches per syscall.
Also, some architectures don't have support in the host for changing the
syscall number via ptrace(), which is currently needed to skip syscall
execution (UML turns any syscall into getpid() to avoid it being executed on
the host). Fixing that is hard, while SYSEMU is easier to implement.
* This version of the patch includes some suggestions of Jeff Dike to avoid
adding any instructions to the syscall fast path, plus some other little
changes, by myself, to make it work even when the syscall is executed with
SYSENTER (but I'm unsure about them). It has been widely tested for quite a
lot of time.
* Various fixed were included to handle the various switches between
various states, i.e. when for instance a syscall entry is traced with one of
PT_SYSCALL / _SYSEMU / _SINGLESTEP and another one is used on exit.
Basically, this is done by remembering which one of them was used even after
the call to ptrace_notify().
* We're combining TIF_SYSCALL_EMU with TIF_SYSCALL_TRACE or TIF_SINGLESTEP
to make do_syscall_trace() notice that the current syscall was started with
SYSEMU on entry, so that no notification ought to be done in the exit path;
this is a bit of a hack, so this problem is solved in another way in next
patches.
* Also, the effects of the patch:
"Ptrace - i386: fix Syscall Audit interaction with singlestep"
are cancelled; they are restored back in the last patch of this series.
Detailed descriptions of the patches doing this kind of processing follow (but
I've already summed everything up).
* Fix behaviour when changing interception kind #1.
In do_syscall_trace(), we check the status of the TIF_SYSCALL_EMU flag
only after doing the debugger notification; but the debugger might have
changed the status of this flag because he continued execution with
PTRACE_SYSCALL, so this is wrong. This patch fixes it by saving the flag
status before calling ptrace_notify().
* Fix behaviour when changing interception kind #2:
avoid intercepting syscall on return when using SYSCALL again.
A guest process switching from using PTRACE_SYSEMU to PTRACE_SYSCALL
crashes.
The problem is in arch/i386/kernel/entry.S. The current SYSEMU patch
inhibits the syscall-handler to be called, but does not prevent
do_syscall_trace() to be called after this for syscall completion
interception.
The appended patch fixes this. It reuses the flag TIF_SYSCALL_EMU to
remember "we come from PTRACE_SYSEMU and now are in PTRACE_SYSCALL", since
the flag is unused in the depicted situation.
* Fix behaviour when changing interception kind #3:
avoid intercepting syscall on return when using SINGLESTEP.
When testing 2.6.9 and the skas3.v6 patch, with my latest patch and had
problems with singlestepping on UML in SKAS with SYSEMU. It looped
receiving SIGTRAPs without moving forward. EIP of the traced process was
the same for all SIGTRAPs.
What's missing is to handle switching from PTRACE_SYSCALL_EMU to
PTRACE_SINGLESTEP in a way very similar to what is done for the change from
PTRACE_SYSCALL_EMU to PTRACE_SYSCALL_TRACE.
I.e., after calling ptrace(PTRACE_SYSEMU), on the return path, the debugger is
notified and then wake ups the process; the syscall is executed (or skipped,
when do_syscall_trace() returns 0, i.e. when using PTRACE_SYSEMU), and
do_syscall_trace() is called again. Since we are on the return path of a
SYSEMU'd syscall, if the wake up is performed through ptrace(PTRACE_SYSCALL),
we must still avoid notifying the parent of the syscall exit. Now, this
behaviour is extended even to resuming with PTRACE_SINGLESTEP.
Signed-off-by: Paolo 'Blaisorblade' Giarrusso <blaisorblade@yahoo.it>
Cc: Jeff Dike <jdike@addtoit.com>
Signed-off-by: Andrew Morton <akpm@osdl.org>
Signed-off-by: Linus Torvalds <torvalds@osdl.org>
2005-09-03 15:57:18 -07:00
|
|
|
/* Note that the debugger could change the result of test_thread_flag!*/
|
2005-09-09 13:01:51 -07:00
|
|
|
ptrace_notify(SIGTRAP | ((current->ptrace & PT_TRACESYSGOOD) ? 0x80:0));
|
2005-04-16 15:20:36 -07:00
|
|
|
|
|
|
|
/*
|
|
|
|
* this isn't the same as continuing with a signal, but it will do
|
|
|
|
* for normal use. strace only continues with a signal if the
|
|
|
|
* stopping signal is not SIGTRAP. -brl
|
|
|
|
*/
|
|
|
|
if (current->exit_code) {
|
|
|
|
send_sig(current->exit_code, current, 1);
|
|
|
|
current->exit_code = 0;
|
|
|
|
}
|
[PATCH] UML Support - Ptrace: adds the host SYSEMU support, for UML and general usage
Jeff Dike <jdike@addtoit.com>,
Paolo 'Blaisorblade' Giarrusso <blaisorblade_spam@yahoo.it>,
Bodo Stroesser <bstroesser@fujitsu-siemens.com>
Adds a new ptrace(2) mode, called PTRACE_SYSEMU, resembling PTRACE_SYSCALL
except that the kernel does not execute the requested syscall; this is useful
to improve performance for virtual environments, like UML, which want to run
the syscall on their own.
In fact, using PTRACE_SYSCALL means stopping child execution twice, on entry
and on exit, and each time you also have two context switches; with SYSEMU you
avoid the 2nd stop and so save two context switches per syscall.
Also, some architectures don't have support in the host for changing the
syscall number via ptrace(), which is currently needed to skip syscall
execution (UML turns any syscall into getpid() to avoid it being executed on
the host). Fixing that is hard, while SYSEMU is easier to implement.
* This version of the patch includes some suggestions of Jeff Dike to avoid
adding any instructions to the syscall fast path, plus some other little
changes, by myself, to make it work even when the syscall is executed with
SYSENTER (but I'm unsure about them). It has been widely tested for quite a
lot of time.
* Various fixed were included to handle the various switches between
various states, i.e. when for instance a syscall entry is traced with one of
PT_SYSCALL / _SYSEMU / _SINGLESTEP and another one is used on exit.
Basically, this is done by remembering which one of them was used even after
the call to ptrace_notify().
* We're combining TIF_SYSCALL_EMU with TIF_SYSCALL_TRACE or TIF_SINGLESTEP
to make do_syscall_trace() notice that the current syscall was started with
SYSEMU on entry, so that no notification ought to be done in the exit path;
this is a bit of a hack, so this problem is solved in another way in next
patches.
* Also, the effects of the patch:
"Ptrace - i386: fix Syscall Audit interaction with singlestep"
are cancelled; they are restored back in the last patch of this series.
Detailed descriptions of the patches doing this kind of processing follow (but
I've already summed everything up).
* Fix behaviour when changing interception kind #1.
In do_syscall_trace(), we check the status of the TIF_SYSCALL_EMU flag
only after doing the debugger notification; but the debugger might have
changed the status of this flag because he continued execution with
PTRACE_SYSCALL, so this is wrong. This patch fixes it by saving the flag
status before calling ptrace_notify().
* Fix behaviour when changing interception kind #2:
avoid intercepting syscall on return when using SYSCALL again.
A guest process switching from using PTRACE_SYSEMU to PTRACE_SYSCALL
crashes.
The problem is in arch/i386/kernel/entry.S. The current SYSEMU patch
inhibits the syscall-handler to be called, but does not prevent
do_syscall_trace() to be called after this for syscall completion
interception.
The appended patch fixes this. It reuses the flag TIF_SYSCALL_EMU to
remember "we come from PTRACE_SYSEMU and now are in PTRACE_SYSCALL", since
the flag is unused in the depicted situation.
* Fix behaviour when changing interception kind #3:
avoid intercepting syscall on return when using SINGLESTEP.
When testing 2.6.9 and the skas3.v6 patch, with my latest patch and had
problems with singlestepping on UML in SKAS with SYSEMU. It looped
receiving SIGTRAPs without moving forward. EIP of the traced process was
the same for all SIGTRAPs.
What's missing is to handle switching from PTRACE_SYSCALL_EMU to
PTRACE_SINGLESTEP in a way very similar to what is done for the change from
PTRACE_SYSCALL_EMU to PTRACE_SYSCALL_TRACE.
I.e., after calling ptrace(PTRACE_SYSEMU), on the return path, the debugger is
notified and then wake ups the process; the syscall is executed (or skipped,
when do_syscall_trace() returns 0, i.e. when using PTRACE_SYSEMU), and
do_syscall_trace() is called again. Since we are on the return path of a
SYSEMU'd syscall, if the wake up is performed through ptrace(PTRACE_SYSCALL),
we must still avoid notifying the parent of the syscall exit. Now, this
behaviour is extended even to resuming with PTRACE_SINGLESTEP.
Signed-off-by: Paolo 'Blaisorblade' Giarrusso <blaisorblade@yahoo.it>
Cc: Jeff Dike <jdike@addtoit.com>
Signed-off-by: Andrew Morton <akpm@osdl.org>
Signed-off-by: Linus Torvalds <torvalds@osdl.org>
2005-09-03 15:57:18 -07:00
|
|
|
ret = is_sysemu;
|
2005-09-09 13:01:51 -07:00
|
|
|
out:
|
2005-04-29 08:08:28 -07:00
|
|
|
if (unlikely(current->audit_context) && !entryexit)
|
2008-01-30 05:30:56 -07:00
|
|
|
audit_syscall_entry(AUDIT_ARCH_I386, regs->orig_ax,
|
|
|
|
regs->bx, regs->cx, regs->dx, regs->si);
|
2005-09-03 15:57:19 -07:00
|
|
|
if (ret == 0)
|
|
|
|
return 0;
|
|
|
|
|
2008-01-30 05:30:56 -07:00
|
|
|
regs->orig_ax = -1; /* force skip of syscall restarting */
|
2005-09-03 15:57:19 -07:00
|
|
|
if (unlikely(current->audit_context))
|
2008-01-30 05:30:56 -07:00
|
|
|
audit_syscall_exit(AUDITSC_RESULT(regs->ax), regs->ax);
|
2005-09-03 15:57:19 -07:00
|
|
|
return 1;
|
2005-04-16 15:20:36 -07:00
|
|
|
}
|
2008-01-30 05:31:01 -07:00
|
|
|
|
|
|
|
#else /* CONFIG_X86_64 */
|
|
|
|
|
|
|
|
static void syscall_trace(struct pt_regs *regs)
|
|
|
|
{
|
|
|
|
|
|
|
|
#if 0
|
|
|
|
printk("trace %s ip %lx sp %lx ax %d origrax %d caller %lx tiflags %x ptrace %x\n",
|
|
|
|
current->comm,
|
|
|
|
regs->ip, regs->sp, regs->ax, regs->orig_ax, __builtin_return_address(0),
|
|
|
|
current_thread_info()->flags, current->ptrace);
|
|
|
|
#endif
|
|
|
|
|
|
|
|
ptrace_notify(SIGTRAP | ((current->ptrace & PT_TRACESYSGOOD)
|
|
|
|
? 0x80 : 0));
|
|
|
|
/*
|
|
|
|
* this isn't the same as continuing with a signal, but it will do
|
|
|
|
* for normal use. strace only continues with a signal if the
|
|
|
|
* stopping signal is not SIGTRAP. -brl
|
|
|
|
*/
|
|
|
|
if (current->exit_code) {
|
|
|
|
send_sig(current->exit_code, current, 1);
|
|
|
|
current->exit_code = 0;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
asmlinkage void syscall_trace_enter(struct pt_regs *regs)
|
|
|
|
{
|
|
|
|
/* do the secure computing check first */
|
|
|
|
secure_computing(regs->orig_ax);
|
|
|
|
|
|
|
|
if (test_thread_flag(TIF_SYSCALL_TRACE)
|
|
|
|
&& (current->ptrace & PT_PTRACED))
|
|
|
|
syscall_trace(regs);
|
|
|
|
|
|
|
|
if (unlikely(current->audit_context)) {
|
|
|
|
if (test_thread_flag(TIF_IA32)) {
|
|
|
|
audit_syscall_entry(AUDIT_ARCH_I386,
|
|
|
|
regs->orig_ax,
|
|
|
|
regs->bx, regs->cx,
|
|
|
|
regs->dx, regs->si);
|
|
|
|
} else {
|
|
|
|
audit_syscall_entry(AUDIT_ARCH_X86_64,
|
|
|
|
regs->orig_ax,
|
|
|
|
regs->di, regs->si,
|
|
|
|
regs->dx, regs->r10);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
asmlinkage void syscall_trace_leave(struct pt_regs *regs)
|
|
|
|
{
|
|
|
|
if (unlikely(current->audit_context))
|
|
|
|
audit_syscall_exit(AUDITSC_RESULT(regs->ax), regs->ax);
|
|
|
|
|
|
|
|
if ((test_thread_flag(TIF_SYSCALL_TRACE)
|
|
|
|
|| test_thread_flag(TIF_SINGLESTEP))
|
|
|
|
&& (current->ptrace & PT_PTRACED))
|
|
|
|
syscall_trace(regs);
|
|
|
|
}
|
|
|
|
|
|
|
|
#endif /* CONFIG_X86_32 */
|