Blame view

kernel/linux-imx6_3.14.28/include/linux/slub_def.h 3.55 KB
6b13f685e   김민수   BSP 최초 추가
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
  #ifndef _LINUX_SLUB_DEF_H
  #define _LINUX_SLUB_DEF_H
  
  /*
   * SLUB : A Slab allocator without object queues.
   *
   * (C) 2007 SGI, Christoph Lameter
   */
  #include <linux/kobject.h>
  
  enum stat_item {
  	ALLOC_FASTPATH,		/* Allocation from cpu slab */
  	ALLOC_SLOWPATH,		/* Allocation by getting a new cpu slab */
  	FREE_FASTPATH,		/* Free to cpu slab */
  	FREE_SLOWPATH,		/* Freeing not to cpu slab */
  	FREE_FROZEN,		/* Freeing to frozen slab */
  	FREE_ADD_PARTIAL,	/* Freeing moves slab to partial list */
  	FREE_REMOVE_PARTIAL,	/* Freeing removes last object */
  	ALLOC_FROM_PARTIAL,	/* Cpu slab acquired from node partial list */
  	ALLOC_SLAB,		/* Cpu slab acquired from page allocator */
  	ALLOC_REFILL,		/* Refill cpu slab from slab freelist */
  	ALLOC_NODE_MISMATCH,	/* Switching cpu slab */
  	FREE_SLAB,		/* Slab freed to the page allocator */
  	CPUSLAB_FLUSH,		/* Abandoning of the cpu slab */
  	DEACTIVATE_FULL,	/* Cpu slab was full when deactivated */
  	DEACTIVATE_EMPTY,	/* Cpu slab was empty when deactivated */
  	DEACTIVATE_TO_HEAD,	/* Cpu slab was moved to the head of partials */
  	DEACTIVATE_TO_TAIL,	/* Cpu slab was moved to the tail of partials */
  	DEACTIVATE_REMOTE_FREES,/* Slab contained remotely freed objects */
  	DEACTIVATE_BYPASS,	/* Implicit deactivation */
  	ORDER_FALLBACK,		/* Number of times fallback was necessary */
  	CMPXCHG_DOUBLE_CPU_FAIL,/* Failure of this_cpu_cmpxchg_double */
  	CMPXCHG_DOUBLE_FAIL,	/* Number of times that cmpxchg double did not match */
  	CPU_PARTIAL_ALLOC,	/* Used cpu partial on alloc */
  	CPU_PARTIAL_FREE,	/* Refill cpu partial on free */
  	CPU_PARTIAL_NODE,	/* Refill cpu partial from node partial */
  	CPU_PARTIAL_DRAIN,	/* Drain cpu partial to node partial */
  	NR_SLUB_STAT_ITEMS };
  
  struct kmem_cache_cpu {
  	void **freelist;	/* Pointer to next available object */
  	unsigned long tid;	/* Globally unique transaction id */
  	struct page *page;	/* The slab from which we are allocating */
  	struct page *partial;	/* Partially allocated frozen slabs */
  #ifdef CONFIG_SLUB_STATS
  	unsigned stat[NR_SLUB_STAT_ITEMS];
  #endif
  };
  
  /*
   * Word size structure that can be atomically updated or read and that
   * contains both the order and the number of objects that a slab of the
   * given order would contain.
   */
  struct kmem_cache_order_objects {
  	unsigned long x;
  };
  
  /*
   * Slab cache management.
   */
  struct kmem_cache {
  	struct kmem_cache_cpu __percpu *cpu_slab;
  	/* Used for retriving partial slabs etc */
  	unsigned long flags;
  	unsigned long min_partial;
  	int size;		/* The size of an object including meta data */
  	int object_size;	/* The size of an object without meta data */
  	int offset;		/* Free pointer offset. */
  	int cpu_partial;	/* Number of per cpu partial objects to keep around */
  	struct kmem_cache_order_objects oo;
  
  	/* Allocation and freeing of slabs */
  	struct kmem_cache_order_objects max;
  	struct kmem_cache_order_objects min;
  	gfp_t allocflags;	/* gfp flags to use on each alloc */
  	int refcount;		/* Refcount for slab cache destroy */
  	void (*ctor)(void *);
  	int inuse;		/* Offset to metadata */
  	int align;		/* Alignment */
  	int reserved;		/* Reserved bytes at the end of slabs */
  	const char *name;	/* Name (only for display!) */
  	struct list_head list;	/* List of slab caches */
  #ifdef CONFIG_SYSFS
  	struct kobject kobj;	/* For sysfs */
  #endif
  #ifdef CONFIG_MEMCG_KMEM
  	struct memcg_cache_params *memcg_params;
  	int max_attr_size; /* for propagation, maximum size of a stored attr */
  #endif
  
  #ifdef CONFIG_NUMA
  	/*
  	 * Defragmentation by allocating from a remote node.
  	 */
  	int remote_node_defrag_ratio;
  #endif
  	struct kmem_cache_node *node[MAX_NUMNODES];
  };
  
  #endif /* _LINUX_SLUB_DEF_H */