Blame view

kernel/linux-rt-4.4.41/arch/powerpc/boot/div64.S 3.26 KB
5113f6f70   김현기   kernel add
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
  /*
   * Divide a 64-bit unsigned number by a 32-bit unsigned number.
   * This routine assumes that the top 32 bits of the dividend are
   * non-zero to start with.
   * On entry, r3 points to the dividend, which get overwritten with
   * the 64-bit quotient, and r4 contains the divisor.
   * On exit, r3 contains the remainder.
   *
   * Copyright (C) 2002 Paul Mackerras, IBM Corp.
   *
   * This program is free software; you can redistribute it and/or
   * modify it under the terms of the GNU General Public License
   * as published by the Free Software Foundation; either version
   * 2 of the License, or (at your option) any later version.
   */
  #include "ppc_asm.h"
  
  	.globl __div64_32
  __div64_32:
  	lwz	r5,0(r3)	# get the dividend into r5/r6
  	lwz	r6,4(r3)
  	cmplw	r5,r4
  	li	r7,0
  	li	r8,0
  	blt	1f
  	divwu	r7,r5,r4	# if dividend.hi >= divisor,
  	mullw	r0,r7,r4	# quotient.hi = dividend.hi / divisor
  	subf.	r5,r0,r5	# dividend.hi %= divisor
  	beq	3f
  1:	mr	r11,r5		# here dividend.hi != 0
  	andis.	r0,r5,0xc000
  	bne	2f
  	cntlzw	r0,r5		# we are shifting the dividend right
  	li	r10,-1		# to make it < 2^32, and shifting
  	srw	r10,r10,r0	# the divisor right the same amount,
  	addc	r9,r4,r10	# rounding up (so the estimate cannot
  	andc	r11,r6,r10	# ever be too large, only too small)
  	andc	r9,r9,r10
  	addze	r9,r9
  	or	r11,r5,r11
  	rotlw	r9,r9,r0
  	rotlw	r11,r11,r0
  	divwu	r11,r11,r9	# then we divide the shifted quantities
  2:	mullw	r10,r11,r4	# to get an estimate of the quotient,
  	mulhwu	r9,r11,r4	# multiply the estimate by the divisor,
  	subfc	r6,r10,r6	# take the product from the divisor,
  	add	r8,r8,r11	# and add the estimate to the accumulated
  	subfe.	r5,r9,r5	# quotient
  	bne	1b
  3:	cmplw	r6,r4
  	blt	4f
  	divwu	r0,r6,r4	# perform the remaining 32-bit division
  	mullw	r10,r0,r4	# and get the remainder
  	add	r8,r8,r0
  	subf	r6,r10,r6
  4:	stw	r7,0(r3)	# return the quotient in *r3
  	stw	r8,4(r3)
  	mr	r3,r6		# return the remainder in r3
  	blr
  
  /*
   * Extended precision shifts.
   *
   * Updated to be valid for shift counts from 0 to 63 inclusive.
   * -- Gabriel
   *
   * R3/R4 has 64 bit value
   * R5    has shift count
   * result in R3/R4
   *
   *  ashrdi3: arithmetic right shift (sign propagation)	
   *  lshrdi3: logical right shift
   *  ashldi3: left shift
   */
  	.globl __ashrdi3
  __ashrdi3:
  	subfic	r6,r5,32
  	srw	r4,r4,r5	# LSW = count > 31 ? 0 : LSW >> count
  	addi	r7,r5,32	# could be xori, or addi with -32
  	slw	r6,r3,r6	# t1 = count > 31 ? 0 : MSW << (32-count)
  	rlwinm	r8,r7,0,32	# t3 = (count < 32) ? 32 : 0
  	sraw	r7,r3,r7	# t2 = MSW >> (count-32)
  	or	r4,r4,r6	# LSW |= t1
  	slw	r7,r7,r8	# t2 = (count < 32) ? 0 : t2
  	sraw	r3,r3,r5	# MSW = MSW >> count
  	or	r4,r4,r7	# LSW |= t2
  	blr
  
  	.globl __ashldi3
  __ashldi3:
  	subfic	r6,r5,32
  	slw	r3,r3,r5	# MSW = count > 31 ? 0 : MSW << count
  	addi	r7,r5,32	# could be xori, or addi with -32
  	srw	r6,r4,r6	# t1 = count > 31 ? 0 : LSW >> (32-count)
  	slw	r7,r4,r7	# t2 = count < 32 ? 0 : LSW << (count-32)
  	or	r3,r3,r6	# MSW |= t1
  	slw	r4,r4,r5	# LSW = LSW << count
  	or	r3,r3,r7	# MSW |= t2
  	blr
  
  	.globl __lshrdi3
  __lshrdi3:
  	subfic	r6,r5,32
  	srw	r4,r4,r5	# LSW = count > 31 ? 0 : LSW >> count
  	addi	r7,r5,32	# could be xori, or addi with -32
  	slw	r6,r3,r6	# t1 = count > 31 ? 0 : MSW << (32-count)
  	srw	r7,r3,r7	# t2 = count < 32 ? 0 : MSW >> (count-32)
  	or	r4,r4,r6	# LSW |= t1
  	srw	r3,r3,r5	# MSW = MSW >> count
  	or	r4,r4,r7	# LSW |= t2
  	blr