Skip to content

Commit

Permalink
Merge branch 'release' of git://git.kernel.org/pub/scm/linux/kernel/g…
Browse files Browse the repository at this point in the history
…it/aegl/linux-2.6

* 'release' of git://git.kernel.org/pub/scm/linux/kernel/git/aegl/linux-2.6:
  [IA64] optimize pagefaults a little
  [IA64] Fix section conflict of ia64_mlogbuf_finish
  [IA64] s/scalibility/scalability/
  [IA64] kdump on INIT needs multi-nodes sync-up (v.2)
  [IA64] wire up {signal,timer,event}fd syscalls
  [IA64] spelling fixes: arch/ia64/
  • Loading branch information
Linus Torvalds committed May 17, 2007
2 parents 52ade9b + 576fe0b commit 5b58e21
Show file tree
Hide file tree
Showing 35 changed files with 129 additions and 121 deletions.
2 changes: 1 addition & 1 deletion arch/ia64/kernel/acpi.c
Original file line number Diff line number Diff line change
Expand Up @@ -791,7 +791,7 @@ static __init int setup_additional_cpus(char *s)
early_param("additional_cpus", setup_additional_cpus);

/*
* cpu_possible_map should be static, it cannot change as cpu's
* cpu_possible_map should be static, it cannot change as CPUs
* are onlined, or offlined. The reason is per-cpu data-structures
* are allocated by some modules at init time, and dont expect to
* do this dynamically on cpu arrival/departure.
Expand Down
24 changes: 16 additions & 8 deletions arch/ia64/kernel/crash.c
Original file line number Diff line number Diff line change
Expand Up @@ -156,24 +156,30 @@ kdump_init_notifier(struct notifier_block *self, unsigned long val, void *data)
if (!kdump_on_init)
return NOTIFY_DONE;

if (val != DIE_INIT_MONARCH_ENTER &&
val != DIE_INIT_SLAVE_ENTER &&
if (val != DIE_INIT_MONARCH_LEAVE &&
val != DIE_INIT_SLAVE_LEAVE &&
val != DIE_INIT_MONARCH_PROCESS &&
val != DIE_MCA_RENDZVOUS_LEAVE &&
val != DIE_MCA_MONARCH_LEAVE)
return NOTIFY_DONE;

nd = (struct ia64_mca_notify_die *)args->err;
/* Reason code 1 means machine check rendezous*/
if ((val == DIE_INIT_MONARCH_ENTER || val == DIE_INIT_SLAVE_ENTER) &&
nd->sos->rv_rc == 1)
/* Reason code 1 means machine check rendezvous*/
if ((val == DIE_INIT_MONARCH_LEAVE || val == DIE_INIT_SLAVE_LEAVE
|| val == DIE_INIT_MONARCH_PROCESS) && nd->sos->rv_rc == 1)
return NOTIFY_DONE;

switch (val) {
case DIE_INIT_MONARCH_ENTER:
case DIE_INIT_MONARCH_PROCESS:
atomic_set(&kdump_in_progress, 1);
*(nd->monarch_cpu) = -1;
break;
case DIE_INIT_MONARCH_LEAVE:
machine_kdump_on_init();
break;
case DIE_INIT_SLAVE_ENTER:
unw_init_running(kdump_cpu_freeze, NULL);
case DIE_INIT_SLAVE_LEAVE:
if (atomic_read(&kdump_in_progress))
unw_init_running(kdump_cpu_freeze, NULL);
break;
case DIE_MCA_RENDZVOUS_LEAVE:
if (atomic_read(&kdump_in_progress))
Expand Down Expand Up @@ -215,8 +221,10 @@ static ctl_table sys_table[] = {
static int
machine_crash_setup(void)
{
/* be notified before default_monarch_init_process */
static struct notifier_block kdump_init_notifier_nb = {
.notifier_call = kdump_init_notifier,
.priority = 1,
};
int ret;
if((ret = register_die_notifier(&kdump_init_notifier_nb)) != 0)
Expand Down
3 changes: 3 additions & 0 deletions arch/ia64/kernel/entry.S
Original file line number Diff line number Diff line change
Expand Up @@ -1585,5 +1585,8 @@ sys_call_table:
data8 sys_getcpu
data8 sys_epoll_pwait // 1305
data8 sys_utimensat
data8 sys_signalfd
data8 sys_timerfd
data8 sys_eventfd

.org sys_call_table + 8*NR_syscalls // guard against failures to increase NR_syscalls
6 changes: 3 additions & 3 deletions arch/ia64/kernel/irq.c
Original file line number Diff line number Diff line change
Expand Up @@ -4,15 +4,15 @@
* Copyright (C) 1992, 1998 Linus Torvalds, Ingo Molnar
*
* This file contains the code used by various IRQ handling routines:
* asking for different IRQ's should be done through these routines
* asking for different IRQs should be done through these routines
* instead of just grabbing them. Thus setups with different IRQ numbers
* shouldn't result in any weird surprises, and installing new handlers
* should be easier.
*
* Copyright (C) Ashok Raj<ashok.raj@intel.com>, Intel Corporation 2004
*
* 4/14/2004: Added code to handle cpu migration and do safe irq
* migration without lossing interrupts for iosapic
* migration without losing interrupts for iosapic
* architecture.
*/

Expand Down Expand Up @@ -190,7 +190,7 @@ void fixup_irqs(void)
}

/*
* Phase 1: Locate irq's bound to this cpu and
* Phase 1: Locate IRQs bound to this cpu and
* relocate them for cpu removal.
*/
migrate_irqs();
Expand Down
2 changes: 1 addition & 1 deletion arch/ia64/kernel/irq_lsapic.c
Original file line number Diff line number Diff line change
Expand Up @@ -23,7 +23,7 @@ lsapic_noop_startup (unsigned int irq)
static void
lsapic_noop (unsigned int irq)
{
/* nuthing to do... */
/* nothing to do... */
}

static int lsapic_retrigger(unsigned int irq)
Expand Down
15 changes: 4 additions & 11 deletions arch/ia64/kernel/kprobes.c
Original file line number Diff line number Diff line change
Expand Up @@ -151,12 +151,12 @@ static uint __kprobes is_cmp_ctype_unc_inst(uint template, uint slot,

cmp_inst.l = kprobe_inst;
if ((cmp_inst.f.x2 == 0) || (cmp_inst.f.x2 == 1)) {
/* Integere compare - Register Register (A6 type)*/
/* Integer compare - Register Register (A6 type)*/
if ((cmp_inst.f.tb == 0) && (cmp_inst.f.ta == 0)
&&(cmp_inst.f.c == 1))
ctype_unc = 1;
} else if ((cmp_inst.f.x2 == 2)||(cmp_inst.f.x2 == 3)) {
/* Integere compare - Immediate Register (A8 type)*/
/* Integer compare - Immediate Register (A8 type)*/
if ((cmp_inst.f.ta == 0) &&(cmp_inst.f.c == 1))
ctype_unc = 1;
}
Expand Down Expand Up @@ -820,7 +820,7 @@ static int __kprobes post_kprobes_handler(struct pt_regs *regs)
return 1;
}

static int __kprobes kprobes_fault_handler(struct pt_regs *regs, int trapnr)
int __kprobes kprobes_fault_handler(struct pt_regs *regs, int trapnr)
{
struct kprobe *cur = kprobe_running();
struct kprobe_ctlblk *kcb = get_kprobe_ctlblk();
Expand Down Expand Up @@ -904,13 +904,6 @@ int __kprobes kprobe_exceptions_notify(struct notifier_block *self,
if (post_kprobes_handler(args->regs))
ret = NOTIFY_STOP;
break;
case DIE_PAGE_FAULT:
/* kprobe_running() needs smp_processor_id() */
preempt_disable();
if (kprobe_running() &&
kprobes_fault_handler(args->regs, args->trapnr))
ret = NOTIFY_STOP;
preempt_enable();
default:
break;
}
Expand Down Expand Up @@ -954,7 +947,7 @@ int __kprobes setjmp_pre_handler(struct kprobe *p, struct pt_regs *regs)
/*
* Callee owns the argument space and could overwrite it, eg
* tail call optimization. So to be absolutely safe
* we save the argument space before transfering the control
* we save the argument space before transferring the control
* to instrumented jprobe function which runs in
* the process context
*/
Expand Down
5 changes: 4 additions & 1 deletion arch/ia64/kernel/mca.c
Original file line number Diff line number Diff line change
Expand Up @@ -273,7 +273,6 @@ static void ia64_mlogbuf_finish(int wait)

mlogbuf_finished = 1;
}
EXPORT_SYMBOL(ia64_mlogbuf_finish);

/*
* Print buffered messages from INIT context.
Expand Down Expand Up @@ -1477,6 +1476,10 @@ default_monarch_init_process(struct notifier_block *self, unsigned long val, voi
struct task_struct *g, *t;
if (val != DIE_INIT_MONARCH_PROCESS)
return NOTIFY_DONE;
#ifdef CONFIG_KEXEC
if (atomic_read(&kdump_in_progress))
return NOTIFY_DONE;
#endif

/*
* FIXME: mlogbuf will brim over with INIT stack dumps.
Expand Down
4 changes: 2 additions & 2 deletions arch/ia64/kernel/mca_drv.c
Original file line number Diff line number Diff line change
Expand Up @@ -438,7 +438,7 @@ is_mca_global(peidx_table_t *peidx, pal_bus_check_info_t *pbci,
* @peidx: pointer of index of processor error section
*
* Return value:
* target address on Success / 0 on Failue
* target address on Success / 0 on Failure
*/
static u64
get_target_identifier(peidx_table_t *peidx)
Expand Down Expand Up @@ -701,7 +701,7 @@ recover_from_processor_error(int platform, slidx_table_t *slidx,
return fatal_mca("External bus check fatal status");

/*
* This is a local MCA and estimated as a recoverble error.
* This is a local MCA and estimated as a recoverable error.
*/
if (platform)
return recover_from_platform_error(slidx, peidx, pbci, sos);
Expand Down
2 changes: 1 addition & 1 deletion arch/ia64/kernel/module.c
Original file line number Diff line number Diff line change
Expand Up @@ -861,7 +861,7 @@ apply_relocate (Elf64_Shdr *sechdrs, const char *strtab, unsigned int symindex,
/*
* Modules contain a single unwind table which covers both the core and the init text
* sections but since the two are not contiguous, we need to split this table up such that
* we can register (and unregister) each "segment" seperately. Fortunately, this sounds
* we can register (and unregister) each "segment" separately. Fortunately, this sounds
* more complicated than it really is.
*/
static void
Expand Down
18 changes: 9 additions & 9 deletions arch/ia64/kernel/perfmon.c
Original file line number Diff line number Diff line change
Expand Up @@ -1318,7 +1318,7 @@ pfm_reserve_session(struct task_struct *task, int is_syswide, unsigned int cpu)
{
unsigned long flags;
/*
* validy checks on cpu_mask have been done upstream
* validity checks on cpu_mask have been done upstream
*/
LOCK_PFS(flags);

Expand Down Expand Up @@ -1384,7 +1384,7 @@ pfm_unreserve_session(pfm_context_t *ctx, int is_syswide, unsigned int cpu)
{
unsigned long flags;
/*
* validy checks on cpu_mask have been done upstream
* validity checks on cpu_mask have been done upstream
*/
LOCK_PFS(flags);

Expand Down Expand Up @@ -1835,7 +1835,7 @@ pfm_flush(struct file *filp, fl_owner_t id)
/*
* remove our file from the async queue, if we use this mode.
* This can be done without the context being protected. We come
* here when the context has become unreacheable by other tasks.
* here when the context has become unreachable by other tasks.
*
* We may still have active monitoring at this point and we may
* end up in pfm_overflow_handler(). However, fasync_helper()
Expand Down Expand Up @@ -2132,7 +2132,7 @@ pfm_close(struct inode *inode, struct file *filp)
filp->private_data = NULL;

/*
* if we free on the spot, the context is now completely unreacheable
* if we free on the spot, the context is now completely unreachable
* from the callers side. The monitored task side is also cut, so we
* can freely cut.
*
Expand Down Expand Up @@ -2562,7 +2562,7 @@ pfm_reset_pmu_state(pfm_context_t *ctx)
ctx->ctx_all_pmcs[0] = pmu_conf->impl_pmcs[0] & ~0x1;

/*
* bitmask of all PMDs that are accesible to this context
* bitmask of all PMDs that are accessible to this context
*/
ctx->ctx_all_pmds[0] = pmu_conf->impl_pmds[0];

Expand Down Expand Up @@ -3395,7 +3395,7 @@ pfm_read_pmds(pfm_context_t *ctx, void *arg, int count, struct pt_regs *regs)
if (unlikely(!PMD_IS_IMPL(cnum))) goto error;
/*
* we can only read the register that we use. That includes
* the one we explicitely initialize AND the one we want included
* the one we explicitly initialize AND the one we want included
* in the sampling buffer (smpl_regs).
*
* Having this restriction allows optimization in the ctxsw routine
Expand Down Expand Up @@ -3715,7 +3715,7 @@ pfm_restart(pfm_context_t *ctx, void *arg, int count, struct pt_regs *regs)
* if non-blocking, then we ensure that the task will go into
* pfm_handle_work() before returning to user mode.
*
* We cannot explicitely reset another task, it MUST always
* We cannot explicitly reset another task, it MUST always
* be done by the task itself. This works for system wide because
* the tool that is controlling the session is logically doing
* "self-monitoring".
Expand Down Expand Up @@ -4644,7 +4644,7 @@ pfm_exit_thread(struct task_struct *task)
switch(state) {
case PFM_CTX_UNLOADED:
/*
* only comes to thios function if pfm_context is not NULL, i.e., cannot
* only comes to this function if pfm_context is not NULL, i.e., cannot
* be in unloaded state
*/
printk(KERN_ERR "perfmon: pfm_exit_thread [%d] ctx unloaded\n", task->pid);
Expand Down Expand Up @@ -5247,7 +5247,7 @@ pfm_end_notify_user(pfm_context_t *ctx)

/*
* main overflow processing routine.
* it can be called from the interrupt path or explicitely during the context switch code
* it can be called from the interrupt path or explicitly during the context switch code
*/
static void
pfm_overflow_handler(struct task_struct *task, pfm_context_t *ctx, u64 pmc0, struct pt_regs *regs)
Expand Down
2 changes: 1 addition & 1 deletion arch/ia64/kernel/perfmon_mckinley.h
Original file line number Diff line number Diff line change
Expand Up @@ -181,7 +181,7 @@ static pmu_config_t pmu_conf_mck={
.pmc_desc = pfm_mck_pmc_desc,
.num_ibrs = 8,
.num_dbrs = 8,
.use_rr_dbregs = 1 /* debug register are use for range retrictions */
.use_rr_dbregs = 1 /* debug register are use for range restrictions */
};


2 changes: 1 addition & 1 deletion arch/ia64/kernel/sal.c
Original file line number Diff line number Diff line change
Expand Up @@ -134,7 +134,7 @@ set_smp_redirect (int flag)
* interrupt redirection. The reason is this would require that
* All interrupts be stopped and hard bind the irq to a cpu.
* Later when the interrupt is fired we need to set the redir hint
* on again in the vector. This is combersome for something that the
* on again in the vector. This is cumbersome for something that the
* user mode irq balancer will solve anyways.
*/
no_int_routing=1;
Expand Down
2 changes: 1 addition & 1 deletion arch/ia64/kernel/salinfo.c
Original file line number Diff line number Diff line change
Expand Up @@ -162,7 +162,7 @@ static DEFINE_SPINLOCK(data_saved_lock);
/** salinfo_platform_oemdata - optional callback to decode oemdata from an error
* record.
* @sect_header: pointer to the start of the section to decode.
* @oemdata: returns vmalloc area containing the decded output.
* @oemdata: returns vmalloc area containing the decoded output.
* @oemdata_size: returns length of decoded output (strlen).
*
* Description: If user space asks for oem data to be decoded by the kernel
Expand Down
6 changes: 3 additions & 3 deletions arch/ia64/kernel/setup.c
Original file line number Diff line number Diff line change
Expand Up @@ -576,7 +576,7 @@ setup_arch (char **cmdline_p)
}

/*
* Display cpu info for all cpu's.
* Display cpu info for all CPUs.
*/
static int
show_cpuinfo (struct seq_file *m, void *v)
Expand Down Expand Up @@ -761,7 +761,7 @@ identify_cpu (struct cpuinfo_ia64 *c)
c->cpu = smp_processor_id();

/* below default values will be overwritten by identify_siblings()
* for Multi-Threading/Multi-Core capable cpu's
* for Multi-Threading/Multi-Core capable CPUs
*/
c->threads_per_core = c->cores_per_socket = c->num_log = 1;
c->socket_id = -1;
Expand Down Expand Up @@ -947,7 +947,7 @@ cpu_init (void)
ia32_cpu_init();
#endif

/* Clear ITC to eliminiate sched_clock() overflows in human time. */
/* Clear ITC to eliminate sched_clock() overflows in human time. */
ia64_set_itc(0);

/* disable all local interrupt sources: */
Expand Down
12 changes: 6 additions & 6 deletions arch/ia64/kernel/smp.c
Original file line number Diff line number Diff line change
Expand Up @@ -186,7 +186,7 @@ handle_IPI (int irq, void *dev_id)
}

/*
* Called with preeemption disabled.
* Called with preemption disabled.
*/
static inline void
send_IPI_single (int dest_cpu, int op)
Expand All @@ -196,7 +196,7 @@ send_IPI_single (int dest_cpu, int op)
}

/*
* Called with preeemption disabled.
* Called with preemption disabled.
*/
static inline void
send_IPI_allbutself (int op)
Expand All @@ -210,7 +210,7 @@ send_IPI_allbutself (int op)
}

/*
* Called with preeemption disabled.
* Called with preemption disabled.
*/
static inline void
send_IPI_all (int op)
Expand All @@ -223,7 +223,7 @@ send_IPI_all (int op)
}

/*
* Called with preeemption disabled.
* Called with preemption disabled.
*/
static inline void
send_IPI_self (int op)
Expand Down Expand Up @@ -252,7 +252,7 @@ kdump_smp_send_init(void)
}
#endif
/*
* Called with preeemption disabled.
* Called with preemption disabled.
*/
void
smp_send_reschedule (int cpu)
Expand All @@ -261,7 +261,7 @@ smp_send_reschedule (int cpu)
}

/*
* Called with preeemption disabled.
* Called with preemption disabled.
*/
static void
smp_send_local_flush_tlb (int cpu)
Expand Down
Loading

0 comments on commit 5b58e21

Please sign in to comment.