Blame view

kernel/linux-rt-4.4.41/arch/alpha/include/asm/processor.h 2.07 KB
5113f6f70   김현기   kernel add
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
  /*
   * include/asm-alpha/processor.h
   *
   * Copyright (C) 1994 Linus Torvalds
   */
  
  #ifndef __ASM_ALPHA_PROCESSOR_H
  #define __ASM_ALPHA_PROCESSOR_H
  
  #include <linux/personality.h>	/* for ADDR_LIMIT_32BIT */
  
  /*
   * Returns current instruction pointer ("program counter").
   */
  #define current_text_addr() \
    ({ void *__pc; __asm__ ("br %0,.+4" : "=r"(__pc)); __pc; })
  
  /*
   * We have a 42-bit user address space: 4TB user VM...
   */
  #define TASK_SIZE (0x40000000000UL)
  
  #define STACK_TOP \
    (current->personality & ADDR_LIMIT_32BIT ? 0x80000000 : 0x00120000000UL)
  
  #define STACK_TOP_MAX	0x00120000000UL
  
  /* This decides where the kernel will search for a free chunk of vm
   * space during mmap's.
   */
  #define TASK_UNMAPPED_BASE \
    ((current->personality & ADDR_LIMIT_32BIT) ? 0x40000000 : TASK_SIZE / 2)
  
  typedef struct {
  	unsigned long seg;
  } mm_segment_t;
  
  /* This is dead.  Everything has been moved to thread_info.  */
  struct thread_struct { };
  #define INIT_THREAD  { }
  
  /* Return saved PC of a blocked thread.  */
  struct task_struct;
  extern unsigned long thread_saved_pc(struct task_struct *);
  
  /* Do necessary setup to start up a newly executed thread.  */
  struct pt_regs;
  extern void start_thread(struct pt_regs *, unsigned long, unsigned long);
  
  /* Free all resources held by a thread. */
  extern void release_thread(struct task_struct *);
  
  unsigned long get_wchan(struct task_struct *p);
  
  #define KSTK_EIP(tsk) (task_pt_regs(tsk)->pc)
  
  #define KSTK_ESP(tsk) \
    ((tsk) == current ? rdusp() : task_thread_info(tsk)->pcb.usp)
  
  #define cpu_relax()	barrier()
  #define cpu_relax_lowlatency() cpu_relax()
  
  #define ARCH_HAS_PREFETCH
  #define ARCH_HAS_PREFETCHW
  #define ARCH_HAS_SPINLOCK_PREFETCH
  
  #ifndef CONFIG_SMP
  /* Nothing to prefetch. */
  #define spin_lock_prefetch(lock)  	do { } while (0)
  #endif
  
  extern inline void prefetch(const void *ptr)  
  { 
  	__builtin_prefetch(ptr, 0, 3);
  }
  
  extern inline void prefetchw(const void *ptr)  
  {
  	__builtin_prefetch(ptr, 1, 3);
  }
  
  #ifdef CONFIG_SMP
  extern inline void spin_lock_prefetch(const void *ptr)  
  {
  	__builtin_prefetch(ptr, 1, 3);
  }
  #endif
  
  #endif /* __ASM_ALPHA_PROCESSOR_H */