+++ /dev/null
---- a/arch/arm/lib/Makefile
-+++ b/arch/arm/lib/Makefile
-@@ -26,7 +26,6 @@ include $(TOPDIR)/config.mk
- LIB = $(obj)lib$(ARCH).o
- LIBGCC = $(obj)libgcc.o
-
--ifndef CONFIG_SPL_BUILD
- GLSOBJS += _ashldi3.o
- GLSOBJS += _ashrdi3.o
- GLSOBJS += _divsi3.o
-@@ -34,9 +33,11 @@ GLSOBJS += _lshrdi3.o
- GLSOBJS += _modsi3.o
- GLSOBJS += _udivsi3.o
- GLSOBJS += _umodsi3.o
-+GLSOBJS += uldivmod.o
-
- GLCOBJS += div0.o
-
-+ifndef CONFIG_SPL_BUILD
- COBJS-y += board.o
- COBJS-y += bootm.o
- COBJS-$(CONFIG_SYS_L2_PL310) += cache-pl310.o
---- /dev/null
-+++ b/arch/arm/lib/uldivmod.S
-@@ -0,0 +1,249 @@
-+/*
-+ * Copyright 2010, Google Inc.
-+ *
-+ * Brought in from coreboot uldivmod.S
-+ *
-+ * SPDX-License-Identifier: GPL-2.0
-+ */
-+
-+#include <linux/linkage.h>
-+#include <asm/assembler.h>
-+
-+/*
-+ * A, Q = r0 + (r1 << 32)
-+ * B, R = r2 + (r3 << 32)
-+ * A / B = Q ... R
-+ */
-+
-+#define ARM(x...) x
-+#define THUMB(x...)
-+
-+A_0 .req r0
-+A_1 .req r1
-+B_0 .req r2
-+B_1 .req r3
-+C_0 .req r4
-+C_1 .req r5
-+D_0 .req r6
-+D_1 .req r7
-+
-+Q_0 .req r0
-+Q_1 .req r1
-+R_0 .req r2
-+R_1 .req r3
-+
-+THUMB(
-+TMP .req r8
-+)
-+
-+.pushsection .text.__aeabi_uldivmod, "ax"
-+ENTRY(__aeabi_uldivmod)
-+
-+ stmfd sp!, {r4, r5, r6, r7, THUMB(TMP,) lr}
-+ @ Test if B == 0
-+ orrs ip, B_0, B_1 @ Z set -> B == 0
-+ beq L_div_by_0
-+ @ Test if B is power of 2: (B & (B - 1)) == 0
-+ subs C_0, B_0, #1
-+ sbc C_1, B_1, #0
-+ tst C_0, B_0
-+ tsteq B_1, C_1
-+ beq L_pow2
-+ @ Test if A_1 == B_1 == 0
-+ orrs ip, A_1, B_1
-+ beq L_div_32_32
-+
-+L_div_64_64:
-+/* CLZ only exists in ARM architecture version 5 and above. */
-+#ifdef HAVE_CLZ
-+ mov C_0, #1
-+ mov C_1, #0
-+ @ D_0 = clz A
-+ teq A_1, #0
-+ clz D_0, A_1
-+ clzeq ip, A_0
-+ addeq D_0, D_0, ip
-+ @ D_1 = clz B
-+ teq B_1, #0
-+ clz D_1, B_1
-+ clzeq ip, B_0
-+ addeq D_1, D_1, ip
-+ @ if clz B - clz A > 0
-+ subs D_0, D_1, D_0
-+ bls L_done_shift
-+ @ B <<= (clz B - clz A)
-+ subs D_1, D_0, #32
-+ rsb ip, D_0, #32
-+ movmi B_1, B_1, lsl D_0
-+ARM( orrmi B_1, B_1, B_0, lsr ip )
-+THUMB( lsrmi TMP, B_0, ip )
-+THUMB( orrmi B_1, B_1, TMP )
-+ movpl B_1, B_0, lsl D_1
-+ mov B_0, B_0, lsl D_0
-+ @ C = 1 << (clz B - clz A)
-+ movmi C_1, C_1, lsl D_0
-+ARM( orrmi C_1, C_1, C_0, lsr ip )
-+THUMB( lsrmi TMP, C_0, ip )
-+THUMB( orrmi C_1, C_1, TMP )
-+ movpl C_1, C_0, lsl D_1
-+ mov C_0, C_0, lsl D_0
-+L_done_shift:
-+ mov D_0, #0
-+ mov D_1, #0
-+ @ C: current bit; D: result
-+#else
-+ @ C: current bit; D: result
-+ mov C_0, #1
-+ mov C_1, #0
-+ mov D_0, #0
-+ mov D_1, #0
-+L_lsl_4:
-+ cmp B_1, #0x10000000
-+ cmpcc B_1, A_1
-+ cmpeq B_0, A_0
-+ bcs L_lsl_1
-+ @ B <<= 4
-+ mov B_1, B_1, lsl #4
-+ orr B_1, B_1, B_0, lsr #28
-+ mov B_0, B_0, lsl #4
-+ @ C <<= 4
-+ mov C_1, C_1, lsl #4
-+ orr C_1, C_1, C_0, lsr #28
-+ mov C_0, C_0, lsl #4
-+ b L_lsl_4
-+L_lsl_1:
-+ cmp B_1, #0x80000000
-+ cmpcc B_1, A_1
-+ cmpeq B_0, A_0
-+ bcs L_subtract
-+ @ B <<= 1
-+ mov B_1, B_1, lsl #1
-+ orr B_1, B_1, B_0, lsr #31
-+ mov B_0, B_0, lsl #1
-+ @ C <<= 1
-+ mov C_1, C_1, lsl #1
-+ orr C_1, C_1, C_0, lsr #31
-+ mov C_0, C_0, lsl #1
-+ b L_lsl_1
-+#endif
-+L_subtract:
-+ @ if A >= B
-+ cmp A_1, B_1
-+ cmpeq A_0, B_0
-+ bcc L_update
-+ @ A -= B
-+ subs A_0, A_0, B_0
-+ sbc A_1, A_1, B_1
-+ @ D |= C
-+ orr D_0, D_0, C_0
-+ orr D_1, D_1, C_1
-+L_update:
-+ @ if A == 0: break
-+ orrs ip, A_1, A_0
-+ beq L_exit
-+ @ C >>= 1
-+ movs C_1, C_1, lsr #1
-+ movs C_0, C_0, rrx
-+ @ if C == 0: break
-+ orrs ip, C_1, C_0
-+ beq L_exit
-+ @ B >>= 1
-+ movs B_1, B_1, lsr #1
-+ mov B_0, B_0, rrx
-+ b L_subtract
-+L_exit:
-+ @ Note: A, B & Q, R are aliases
-+ mov R_0, A_0
-+ mov R_1, A_1
-+ mov Q_0, D_0
-+ mov Q_1, D_1
-+ ldmfd sp!, {r4, r5, r6, r7, THUMB(TMP,) pc}
-+
-+L_div_32_32:
-+ @ Note: A_0 & r0 are aliases
-+ @ Q_1 r1
-+ mov r1, B_0
-+ bl __aeabi_uidivmod
-+ mov R_0, r1
-+ mov R_1, #0
-+ mov Q_1, #0
-+ ldmfd sp!, {r4, r5, r6, r7, THUMB(TMP,) pc}
-+
-+L_pow2:
-+#ifdef HAVE_CLZ
-+ @ Note: A, B and Q, R are aliases
-+ @ R = A & (B - 1)
-+ and C_0, A_0, C_0
-+ and C_1, A_1, C_1
-+ @ Q = A >> log2(B)
-+ @ Note: B must not be 0 here!
-+ clz D_0, B_0
-+ add D_1, D_0, #1
-+ rsbs D_0, D_0, #31
-+ bpl L_1
-+ clz D_0, B_1
-+ rsb D_0, D_0, #31
-+ mov A_0, A_1, lsr D_0
-+ add D_0, D_0, #32
-+L_1:
-+ movpl A_0, A_0, lsr D_0
-+ARM( orrpl A_0, A_0, A_1, lsl D_1 )
-+THUMB( lslpl TMP, A_1, D_1 )
-+THUMB( orrpl A_0, A_0, TMP )
-+ mov A_1, A_1, lsr D_0
-+ @ Mov back C to R
-+ mov R_0, C_0
-+ mov R_1, C_1
-+ ldmfd sp!, {r4, r5, r6, r7, THUMB(TMP,) pc}
-+#else
-+ @ Note: A, B and Q, R are aliases
-+ @ R = A & (B - 1)
-+ and C_0, A_0, C_0
-+ and C_1, A_1, C_1
-+ @ Q = A >> log2(B)
-+ @ Note: B must not be 0 here!
-+ @ Count the leading zeroes in B.
-+ mov D_0, #0
-+ orrs B_0, B_0, B_0
-+ @ If B is greater than 1 << 31, divide A and B by 1 << 32.
-+ moveq A_0, A_1
-+ moveq A_1, #0
-+ moveq B_0, B_1
-+ @ Count the remaining leading zeroes in B.
-+ movs B_1, B_0, lsl #16
-+ addeq D_0, #16
-+ moveq B_0, B_0, lsr #16
-+ tst B_0, #0xff
-+ addeq D_0, #8
-+ moveq B_0, B_0, lsr #8
-+ tst B_0, #0xf
-+ addeq D_0, #4
-+ moveq B_0, B_0, lsr #4
-+ tst B_0, #0x3
-+ addeq D_0, #2
-+ moveq B_0, B_0, lsr #2
-+ tst B_0, #0x1
-+ addeq D_0, #1
-+ @ Shift A to the right by the appropriate amount.
-+ rsb D_1, D_0, #32
-+ mov Q_0, A_0, lsr D_0
-+ ARM( orr Q_0, Q_0, A_1, lsl D_1 )
-+ THUMB( lsl A_1, D_1 )
-+ THUMB( orr Q_0, A_1 )
-+ mov Q_1, A_1, lsr D_0
-+ @ Move C to R
-+ mov R_0, C_0
-+ mov R_1, C_1
-+ ldmfd sp!, {r4, r5, r6, r7, THUMB(TMP,) pc}
-+#endif
-+
-+L_div_by_0:
-+ bl __div0
-+ @ As wrong as it could be
-+ mov Q_0, #0
-+ mov Q_1, #0
-+ mov R_0, #0
-+ mov R_1, #0
-+ ldmfd sp!, {r4, r5, r6, r7, THUMB(TMP,) pc}
-+ENDPROC(__aeabi_uldivmod)
-+.popsection