Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 1 | #include <linux/mm.h> |
| 2 | #include <linux/module.h> |
| 3 | #include <linux/sched.h> |
| 4 | #include <linux/init.h> |
| 5 | #include <linux/init_task.h> |
| 6 | #include <linux/fs.h> |
| 7 | #include <linux/mqueue.h> |
| 8 | |
| 9 | #include <asm/uaccess.h> |
| 10 | #include <asm/pgtable.h> |
| 11 | #include <asm/desc.h> |
| 12 | |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 13 | static struct signal_struct init_signals = INIT_SIGNALS(init_signals); |
| 14 | static struct sighand_struct init_sighand = INIT_SIGHAND(init_sighand); |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 15 | |
| 16 | /* |
| 17 | * Initial thread structure. |
| 18 | * |
| 19 | * We need to make sure that this is THREAD_SIZE aligned due to the |
| 20 | * way process stacks are handled. This is done by having a special |
| 21 | * "init_task" linker map entry.. |
| 22 | */ |
Joe Perches | d200c92 | 2009-09-20 18:14:13 -0400 | [diff] [blame] | 23 | union thread_union init_thread_union __init_task_data = |
| 24 | { INIT_THREAD_INFO(init_task) }; |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 25 | |
| 26 | /* |
| 27 | * Initial task structure. |
| 28 | * |
| 29 | * All other task structs will be allocated on slabs in fork.c |
| 30 | */ |
| 31 | struct task_struct init_task = INIT_TASK(init_task); |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 32 | EXPORT_SYMBOL(init_task); |
| 33 | |
| 34 | /* |
| 35 | * per-CPU TSS segments. Threads are completely 'soft' on Linux, |
Hiroshi Shimamoto | 7778887 | 2007-10-19 20:35:02 +0200 | [diff] [blame] | 36 | * no more per-task TSS's. The TSS size is kept cacheline-aligned |
Tim Abbott | 4af57b7 | 2010-02-20 01:03:34 +0100 | [diff] [blame] | 37 | * so they are allowed to end up in the .data..cacheline_aligned |
Hiroshi Shimamoto | 7778887 | 2007-10-19 20:35:02 +0200 | [diff] [blame] | 38 | * section. Since TSS's are completely CPU-local, we want them |
| 39 | * on exact cacheline boundaries, to eliminate cacheline ping-pong. |
| 40 | */ |
Fenghua Yu | f34e3b6 | 2007-07-19 01:48:13 -0700 | [diff] [blame] | 41 | DEFINE_PER_CPU_SHARED_ALIGNED(struct tss_struct, init_tss) = INIT_TSS; |
Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 42 | |