Blame view

kernel/linux-imx6_3.14.28/include/linux/hash.h 3.06 KB
6b13f685e   김민수   BSP 최초 추가
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
  #ifndef _LINUX_HASH_H
  #define _LINUX_HASH_H
  /* Fast hashing routine for ints,  longs and pointers.
     (C) 2002 Nadia Yvette Chambers, IBM */
  
  /*
   * Knuth recommends primes in approximately golden ratio to the maximum
   * integer representable by a machine word for multiplicative hashing.
   * Chuck Lever verified the effectiveness of this technique:
   * http://www.citi.umich.edu/techreports/reports/citi-tr-00-1.pdf
   *
   * These primes are chosen to be bit-sparse, that is operations on
   * them can use shifts and additions instead of multiplications for
   * machines where multiplications are slow.
   */
  
  #include <asm/types.h>
  #include <asm/hash.h>
  #include <linux/compiler.h>
  
  /* 2^31 + 2^29 - 2^25 + 2^22 - 2^19 - 2^16 + 1 */
  #define GOLDEN_RATIO_PRIME_32 0x9e370001UL
  /*  2^63 + 2^61 - 2^57 + 2^54 - 2^51 - 2^18 + 1 */
  #define GOLDEN_RATIO_PRIME_64 0x9e37fffffffc0001UL
  
  #if BITS_PER_LONG == 32
  #define GOLDEN_RATIO_PRIME GOLDEN_RATIO_PRIME_32
  #define hash_long(val, bits) hash_32(val, bits)
  #elif BITS_PER_LONG == 64
  #define hash_long(val, bits) hash_64(val, bits)
  #define GOLDEN_RATIO_PRIME GOLDEN_RATIO_PRIME_64
  #else
  #error Wordsize not 32 or 64
  #endif
  
  static __always_inline u64 hash_64(u64 val, unsigned int bits)
  {
  	u64 hash = val;
  
  	/*  Sigh, gcc can't optimise this alone like it does for 32 bits. */
  	u64 n = hash;
  	n <<= 18;
  	hash -= n;
  	n <<= 33;
  	hash -= n;
  	n <<= 3;
  	hash += n;
  	n <<= 3;
  	hash -= n;
  	n <<= 4;
  	hash += n;
  	n <<= 2;
  	hash += n;
  
  	/* High bits are more random, so use them. */
  	return hash >> (64 - bits);
  }
  
  static inline u32 hash_32(u32 val, unsigned int bits)
  {
  	/* On some cpus multiply is faster, on others gcc will do shifts */
  	u32 hash = val * GOLDEN_RATIO_PRIME_32;
  
  	/* High bits are more random, so use them. */
  	return hash >> (32 - bits);
  }
  
  static inline unsigned long hash_ptr(const void *ptr, unsigned int bits)
  {
  	return hash_long((unsigned long)ptr, bits);
  }
  
  static inline u32 hash32_ptr(const void *ptr)
  {
  	unsigned long val = (unsigned long)ptr;
  
  #if BITS_PER_LONG == 64
  	val ^= (val >> 32);
  #endif
  	return (u32)val;
  }
  
  struct fast_hash_ops {
  	u32 (*hash)(const void *data, u32 len, u32 seed);
  	u32 (*hash2)(const u32 *data, u32 len, u32 seed);
  };
  
  /**
   *	arch_fast_hash - Caclulates a hash over a given buffer that can have
   *			 arbitrary size. This function will eventually use an
   *			 architecture-optimized hashing implementation if
   *			 available, and trades off distribution for speed.
   *
   *	@data: buffer to hash
   *	@len: length of buffer in bytes
   *	@seed: start seed
   *
   *	Returns 32bit hash.
   */
  extern u32 arch_fast_hash(const void *data, u32 len, u32 seed);
  
  /**
   *	arch_fast_hash2 - Caclulates a hash over a given buffer that has a
   *			  size that is of a multiple of 32bit words. This
   *			  function will eventually use an architecture-
   *			  optimized hashing implementation if available,
   *			  and trades off distribution for speed.
   *
   *	@data: buffer to hash (must be 32bit padded)
   *	@len: number of 32bit words
   *	@seed: start seed
   *
   *	Returns 32bit hash.
   */
  extern u32 arch_fast_hash2(const u32 *data, u32 len, u32 seed);
  
  #endif /* _LINUX_HASH_H */