| /* |
| * Copyright (c) 1991,1992,1995 Linus Torvalds |
| * Copyright (c) 1994 Alan Modra |
| * Copyright (c) 1995 Markus Kuhn |
| * Copyright (c) 1996 Ingo Molnar |
| * Copyright (c) 1998 Andrea Arcangeli |
| * Copyright (c) 2002,2006 Vojtech Pavlik |
| * Copyright (c) 2003 Andi Kleen |
| * |
| */ |
| |
| #include <linux/clockchips.h> |
| #include <linux/interrupt.h> |
| #include <linux/time.h> |
| #include <linux/mca.h> |
| |
| #include <asm/vsyscall.h> |
| #include <asm/x86_init.h> |
| #include <asm/i8259.h> |
| #include <asm/i8253.h> |
| #include <asm/timer.h> |
| #include <asm/hpet.h> |
| #include <asm/time.h> |
| #include <asm/nmi.h> |
| |
| #if defined(CONFIG_X86_32) && defined(CONFIG_X86_IO_APIC) |
| int timer_ack; |
| #endif |
| |
| #ifdef CONFIG_X86_64 |
| volatile unsigned long __jiffies __section_jiffies = INITIAL_JIFFIES; |
| #endif |
| |
| unsigned long profile_pc(struct pt_regs *regs) |
| { |
| unsigned long pc = instruction_pointer(regs); |
| |
| /* Assume the lock function has either no stack frame or a copy |
| of flags from PUSHF |
| Eflags always has bits 22 and up cleared unlike kernel addresses. */ |
| if (!user_mode_vm(regs) && in_lock_functions(pc)) { |
| #ifdef CONFIG_FRAME_POINTER |
| return *(unsigned long *)(regs->bp + sizeof(long)); |
| #else |
| unsigned long *sp = (unsigned long *)regs->sp; |
| if (sp[0] >> 22) |
| return sp[0]; |
| if (sp[1] >> 22) |
| return sp[1]; |
| #endif |
| } |
| return pc; |
| } |
| EXPORT_SYMBOL(profile_pc); |
| |
| /* |
| * Default timer interrupt handler for PIT/HPET |
| */ |
| static irqreturn_t timer_interrupt(int irq, void *dev_id) |
| { |
| /* Keep nmi watchdog up to date */ |
| inc_irq_stat(irq0_irqs); |
| |
| /* Optimized out for !IO_APIC and x86_64 */ |
| if (timer_ack) { |
| /* |
| * Subtle, when I/O APICs are used we have to ack timer IRQ |
| * manually to deassert NMI lines for the watchdog if run |
| * on an 82489DX-based system. |
| */ |
| spin_lock(&i8259A_lock); |
| outb(0x0c, PIC_MASTER_OCW3); |
| /* Ack the IRQ; AEOI will end it automatically. */ |
| inb(PIC_MASTER_POLL); |
| spin_unlock(&i8259A_lock); |
| } |
| |
| global_clock_event->event_handler(global_clock_event); |
| |
| /* MCA bus quirk: Acknowledge irq0 by setting bit 7 in port 0x61 */ |
| if (MCA_bus) |
| outb_p(inb_p(0x61)| 0x80, 0x61); |
| |
| return IRQ_HANDLED; |
| } |
| |
| /* |
| * calibrate_cpu is used on systems with fixed rate TSCs to determine |
| * processor frequency |
| */ |
| #define TICK_COUNT 100000000 |
| unsigned long __init calibrate_cpu(void) |
| { |
| int tsc_start, tsc_now; |
| int i, no_ctr_free; |
| unsigned long evntsel3 = 0, pmc3 = 0, pmc_now = 0; |
| unsigned long flags; |
| |
| for (i = 0; i < 4; i++) |
| if (avail_to_resrv_perfctr_nmi_bit(i)) |
| break; |
| no_ctr_free = (i == 4); |
| if (no_ctr_free) { |
| WARN(1, KERN_WARNING "Warning: AMD perfctrs busy ... " |
| "cpu_khz value may be incorrect.\n"); |
| i = 3; |
| rdmsrl(MSR_K7_EVNTSEL3, evntsel3); |
| wrmsrl(MSR_K7_EVNTSEL3, 0); |
| rdmsrl(MSR_K7_PERFCTR3, pmc3); |
| } else { |
| reserve_perfctr_nmi(MSR_K7_PERFCTR0 + i); |
| reserve_evntsel_nmi(MSR_K7_EVNTSEL0 + i); |
| } |
| local_irq_save(flags); |
| /* start measuring cycles, incrementing from 0 */ |
| wrmsrl(MSR_K7_PERFCTR0 + i, 0); |
| wrmsrl(MSR_K7_EVNTSEL0 + i, 1 << 22 | 3 << 16 | 0x76); |
| rdtscl(tsc_start); |
| do { |
| rdmsrl(MSR_K7_PERFCTR0 + i, pmc_now); |
| tsc_now = get_cycles(); |
| } while ((tsc_now - tsc_start) < TICK_COUNT); |
| |
| local_irq_restore(flags); |
| if (no_ctr_free) { |
| wrmsrl(MSR_K7_EVNTSEL3, 0); |
| wrmsrl(MSR_K7_PERFCTR3, pmc3); |
| wrmsrl(MSR_K7_EVNTSEL3, evntsel3); |
| } else { |
| release_perfctr_nmi(MSR_K7_PERFCTR0 + i); |
| release_evntsel_nmi(MSR_K7_EVNTSEL0 + i); |
| } |
| |
| return pmc_now * tsc_khz / (tsc_now - tsc_start); |
| } |
| |
| static struct irqaction irq0 = { |
| .handler = timer_interrupt, |
| .flags = IRQF_DISABLED | IRQF_NOBALANCING | IRQF_IRQPOLL | IRQF_TIMER, |
| .name = "timer" |
| }; |
| |
| void __init setup_default_timer_irq(void) |
| { |
| irq0.mask = cpumask_of_cpu(0); |
| setup_irq(0, &irq0); |
| } |
| |
| /* Default timer init function */ |
| void __init hpet_time_init(void) |
| { |
| if (!hpet_enable()) |
| setup_pit_timer(); |
| setup_default_timer_irq(); |
| } |
| |
| static void x86_late_time_init(void) |
| { |
| x86_init.timers.timer_init(); |
| } |
| |
| /* |
| * Initialize TSC and delay the periodic timer init to |
| * late x86_late_time_init() so ioremap works. |
| */ |
| void __init time_init(void) |
| { |
| tsc_init(); |
| late_time_init = x86_late_time_init; |
| } |