svn commit: r306645 - in head/sys: arm/arm arm/include conf

Andrew Turner andrew at FreeBSD.org
Mon Oct 3 16:10:40 UTC 2016


Author: andrew
Date: Mon Oct  3 16:10:38 2016
New Revision: 306645
URL: https://svnweb.freebsd.org/changeset/base/306645

Log:
  Remove unused functions on armv6. Some of the cache handling code is still
  used in the elf trampoline so add a macro to handle this.
  
  Sponsored by:	ABT Systems Ltd

Modified:
  head/sys/arm/arm/cpufunc_asm_arm11.S
  head/sys/arm/arm/cpufunc_asm_arm11x6.S
  head/sys/arm/arm/cpufunc_asm_armv6.S
  head/sys/arm/arm/cpufunc_asm_armv7.S
  head/sys/arm/include/cpufunc.h
  head/sys/conf/Makefile.arm

Modified: head/sys/arm/arm/cpufunc_asm_arm11.S
==============================================================================
--- head/sys/arm/arm/cpufunc_asm_arm11.S	Mon Oct  3 15:17:22 2016	(r306644)
+++ head/sys/arm/arm/cpufunc_asm_arm11.S	Mon Oct  3 16:10:38 2016	(r306645)
@@ -38,62 +38,6 @@
 __FBSDID("$FreeBSD$");
 
 /*
- * TLB functions
- */
-ENTRY(arm11_tlb_flushID_SE)
-	mcr	p15, 0, r0, c8, c6, 1	/* flush D tlb single entry */
-	mcr	p15, 0, r0, c8, c5, 1	/* flush I tlb single entry */
-	mcr	p15, 0, r0, c7, c10, 4	/* drain write buffer */
-	RET
-END(arm11_tlb_flushID_SE)
-
-/*
- * Context switch.
- *
- * These is the CPU-specific parts of the context switcher cpu_switch()
- * These functions actually perform the TTB reload.
- *
- * NOTE: Special calling convention
- *	r1, r4-r13 must be preserved
- */
-ENTRY(arm11_context_switch)
-	/*
-	 * We can assume that the caches will only contain kernel addresses
-	 * at this point.  So no need to flush them again.
-	 */
-	mcr	p15, 0, r0, c7, c10, 4	/* drain the write buffer */
-	mcr	p15, 0, r0, c2, c0, 0	/* set the new TTB */
-	mcr	p15, 0, r0, c8, c7, 0	/* and flush the I+D tlbs */
-
-	/* Paranoia -- make sure the pipeline is empty. */
-	nop
-	nop
-	nop
-	RET
-END(arm11_context_switch)
-
-/*
- * TLB functions
- */
-ENTRY(arm11_tlb_flushID)
-	mcr	p15, 0, r0, c8, c7, 0	/* flush I+D tlb */
-	mcr	p15, 0, r0, c7, c10, 4	/* drain write buffer */
-	mov	pc, lr
-END(arm11_tlb_flushID)
-
-ENTRY(arm11_tlb_flushD)
-	mcr	p15, 0, r0, c8, c6, 0	/* flush D tlb */
-	mcr	p15, 0, r0, c7, c10, 4	/* drain write buffer */
-	mov	pc, lr
-END(arm11_tlb_flushD)
-
-ENTRY(arm11_tlb_flushD_SE)
-	mcr	p15, 0, r0, c8, c6, 1	/* flush D tlb single entry */
-	mcr	p15, 0, r0, c7, c10, 4	/* drain write buffer */
-	mov	pc, lr
-END(arm11_tlb_flushD_SE)
-
-/*
  * Other functions
  */
 ENTRY(arm11_drain_writebuf)

Modified: head/sys/arm/arm/cpufunc_asm_arm11x6.S
==============================================================================
--- head/sys/arm/arm/cpufunc_asm_arm11x6.S	Mon Oct  3 15:17:22 2016	(r306644)
+++ head/sys/arm/arm/cpufunc_asm_arm11x6.S	Mon Oct  3 16:10:38 2016	(r306645)
@@ -64,55 +64,6 @@ __FBSDID("$FreeBSD$");
 
 	.cpu arm1176jz-s
 
-#if 0
-#define Invalidate_I_cache(Rtmp1, Rtmp2) \
-	mcr	p15, 0, Rtmp1, c7, c5, 0	/* Invalidate Entire I cache */
-#else
-/*
- * Workaround for
- *
- *    Erratum 411920 in ARM1136 (fixed in r1p4)
- *    Erratum 415045 in ARM1176 (fixed in r0p5?)
- *
- *	- value of arg 'reg' Should Be Zero
- */
-#define Invalidate_I_cache(Rtmp1, Rtmp2) \
-	mov	Rtmp1, #0;		/* SBZ */			\
-	mrs	Rtmp2, cpsr;						\
-	cpsid	ifa;							\
-	mcr	p15, 0, Rtmp1, c7, c5, 0;	/* Nuke Whole Icache */	\
-	mcr	p15, 0, Rtmp1, c7, c5, 0;	/* Nuke Whole Icache */	\
-	mcr	p15, 0, Rtmp1, c7, c5, 0;	/* Nuke Whole Icache */	\
-	mcr	p15, 0, Rtmp1, c7, c5, 0;	/* Nuke Whole Icache */	\
-	msr	cpsr_cx, Rtmp2;						\
-	nop;								\
-	nop;								\
-	nop;								\
-	nop;								\
-	nop;								\
-	nop;								\
-	nop;								\
-	nop;								\
-	nop;								\
-	nop;								\
-	nop;
-#endif
-
-#if 1
-#define Flush_D_cache(reg) \
-	mov	reg, #0;		/* SBZ */					\
-	mcr	p15, 0, reg, c7, c14, 0;/* Clean and Invalidate Entire Data Cache */	\
-	mcr	p15, 0, reg, c7, c10, 4;/* Data Synchronization Barrier */
-#else
-#define Flush_D_cache(reg) \
-1:	mov	reg, #0;		/* SBZ */					\
-	mcr	p15, 0, reg, c7, c14, 0;/* Clean and Invalidate Entire Data Cache */	\
-	mrc	p15, 0, reg, C7, C10, 6;/* Read Cache Dirty Status Register */		\
-	ands	reg, reg, #01;		/* Check if it is clean */			\
-	bne	1b;			/* loop if not */				\
-	mcr	p15, 0, reg, c7, c10, 4;/* Data Synchronization Barrier */
-#endif
-
 ENTRY(arm11x6_setttb)
 	mov	r1, #0
 	mcr	p15, 0, r0, c2, c0, 0	/* load new TTB */
@@ -121,71 +72,6 @@ ENTRY(arm11x6_setttb)
 	RET
 END(arm11x6_setttb)
 
-ENTRY_NP(arm11x6_idcache_wbinv_all)
-	Flush_D_cache(r0)
-	Invalidate_I_cache(r0, r1)
-	RET
-END(arm11x6_idcache_wbinv_all)
-
-ENTRY_NP(arm11x6_dcache_wbinv_all)
-	Flush_D_cache(r0)
-	RET
-END(arm11x6_dcache_wbinv_all)
-
-ENTRY_NP(arm11x6_icache_sync_range)
-	add	r1, r1, r0
-	sub	r1, r1, #1
-	/* Erratum ARM1136 371025, workaround #2 */
-	/* Erratum ARM1176 371367 */
-	mrs	r2, cpsr		/* save the CPSR */
-	cpsid	ifa			/* disable interrupts (irq,fiq,abort) */
-	mov	r3, #0
-	mcr	p15, 0, r3, c13, c0, 0	/* write FCSE (uTLB invalidate) */
-	mcr	p15, 0, r3, c7, c5, 4	/* flush prefetch buffer */
-	add	r3, pc, #0x24
-	mcr	p15, 0, r3, c7, c13, 1	/* prefetch I-cache line */
-	mcrr	p15, 0, r1, r0, c5	/* invalidate I-cache range */
-	msr	cpsr_cx, r2		/* local_irq_restore */
-	nop
-	nop
-	nop
-	nop
-	nop
-	nop
-	nop
-
-	mcrr	p15, 0, r1, r0, c12	/* clean and invalidate D cache range */ /* XXXNH */
-	mcr	p15, 0, r0, c7, c10, 4	/* drain the write buffer */
-	RET
-END(arm11x6_icache_sync_range)
-
-ENTRY_NP(arm11x6_idcache_wbinv_range)
-	add	r1, r1, r0
-	sub	r1, r1, #1
-	/* Erratum ARM1136 371025, workaround #2 */
-	/* Erratum ARM1176 371367 */
-	mrs	r2, cpsr		/* save the CPSR */
-	cpsid	ifa			/* disable interrupts (irq,fiq,abort) */
-	mov	r3, #0
-	mcr	p15, 0, r3, c13, c0, 0	/* write FCSE (uTLB invalidate) */
-	mcr	p15, 0, r3, c7, c5, 4	/* flush prefetch buffer */
-	add	r3, pc, #0x24
-	mcr	p15, 0, r3, c7, c13, 1	/* prefetch I-cache line */
-	mcrr	p15, 0, r1, r0, c5	/* invalidate I-cache range */
-	msr	cpsr_cx, r2		/* local_irq_restore */
-	nop
-	nop
-	nop
-	nop
-	nop
-	nop
-	nop
-
-	mcrr	p15, 0, r1, r0, c14	/* clean and invalidate D cache range */
-	mcr	p15, 0, r0, c7, c10, 4	/* drain the write buffer */
-	RET
-END(arm11x6_idcache_wbinv_range)
-
 /*
  * Preload the cache before issuing the WFI by conditionally disabling the
  * mcr intstructions the first time around the loop. Ensure the function is
@@ -208,4 +94,3 @@ ENTRY_NP(arm11x6_sleep)
 	bne	1b
 	RET
 END(arm11x6_sleep)
-

Modified: head/sys/arm/arm/cpufunc_asm_armv6.S
==============================================================================
--- head/sys/arm/arm/cpufunc_asm_armv6.S	Mon Oct  3 15:17:22 2016	(r306644)
+++ head/sys/arm/arm/cpufunc_asm_armv6.S	Mon Oct  3 16:10:38 2016	(r306645)
@@ -43,56 +43,10 @@
 	.arch	armv6
 
 /*
- * Functions to set the MMU Translation Table Base register
- *
- * We need to clean and flush the cache as it uses virtual
- * addresses that are about to change.
- */
-ENTRY(armv6_setttb)
-	mcr	p15, 0, r0, c7, c10, 4	/* drain the write buffer */
-
-	mcr	p15, 0, r0, c2, c0, 0	/* load new TTB */
-
-	mcr	p15, 0, r0, c8, c7, 0	/* invalidate I+D TLBs */
-	RET
-END(armv6_setttb)
-
-/*
  * Cache operations.
  */
 
-/* LINTSTUB: void armv6_dcache_wb_range(vaddr_t, vsize_t); */
-ENTRY(armv6_dcache_wb_range)
-	add	r1, r1, r0
-	sub	r1, r1, #1
-	mcrr	p15, 0, r1, r0, c12	/* clean D cache range */
-	mcr	p15, 0, r0, c7, c10, 4	/* drain the write buffer */
-	RET
-END(armv6_dcache_wb_range)
-
-/* LINTSTUB: void armv6_dcache_wbinv_range(vaddr_t, vsize_t); */
-ENTRY(armv6_dcache_wbinv_range)
-	add	r1, r1, r0
-	sub	r1, r1, #1
-	mcrr	p15, 0, r1, r0, c14	/* clean and invaliate D cache range */
-	mcr	p15, 0, r0, c7, c10, 4	/* drain the write buffer */
-	RET
-END(armv6_dcache_wbinv_range)
-
-/*
- * Note, we must not invalidate everything.  If the range is too big we
- * must use wb-inv of the entire cache.
- *
- * LINTSTUB: void armv6_dcache_inv_range(vaddr_t, vsize_t);
- */
-ENTRY(armv6_dcache_inv_range)
-	add	r1, r1, r0
-	sub	r1, r1, #1
-	mcrr	p15, 0, r1, r0, c6	/* invaliate D cache range */
-	mcr	p15, 0, r0, c7, c10, 4	/* drain the write buffer */
-	RET
-END(armv6_dcache_inv_range)
-
+#ifdef ELF_TRAMPOLINE
 /* LINTSTUB: void armv6_idcache_wbinv_all(void); */
 ENTRY_NP(armv6_idcache_wbinv_all)
 	/*
@@ -107,10 +61,4 @@ ENTRY_NP(armv6_idcache_wbinv_all)
 	mcr	p15, 0, r0, c7, c10, 4	/* drain the write buffer */
 	RET
 END(armv6_idcache_wbinv_all)
-
-ENTRY(armv6_idcache_inv_all)
-	mov	r0, #0
-	mcr	p15, 0, r0, c7, c7, 0	/* invalidate all I+D cache */
-	RET
-END(armv6_idcache_inv_all)
-
+#endif /* ELF_TRAMPOLINE */

Modified: head/sys/arm/arm/cpufunc_asm_armv7.S
==============================================================================
--- head/sys/arm/arm/cpufunc_asm_armv7.S	Mon Oct  3 15:17:22 2016	(r306644)
+++ head/sys/arm/arm/cpufunc_asm_armv7.S	Mon Oct  3 16:10:38 2016	(r306645)
@@ -86,35 +86,7 @@ ENTRY(armv7_setttb)
 	RET
 END(armv7_setttb)
 
-ENTRY(armv7_tlb_flushID)
-	dsb
-#ifdef SMP
-	mcr	CP15_TLBIALLIS
-	mcr	CP15_BPIALLIS
-#else
-	mcr	CP15_TLBIALL
-	mcr	CP15_BPIALL
-#endif
-	dsb
-	isb
-	mov	pc, lr
-END(armv7_tlb_flushID)
-
-ENTRY(armv7_tlb_flushID_SE)
-	ldr	r1, .Lpage_mask
-	bic	r0, r0, r1
-#ifdef SMP
-	mcr	CP15_TLBIMVAAIS(r0)
-	mcr	CP15_BPIALLIS
-#else
-	mcr	CP15_TLBIMVA(r0)
-	mcr	CP15_BPIALL
-#endif
-	dsb
-	isb
-	mov	pc, lr
-END(armv7_tlb_flushID_SE)
-
+#ifdef ELF_TRAMPOLINE
 /* Based on algorithm from ARM Architecture Reference Manual */
 ENTRY(armv7_dcache_wbinv_all)
 	stmdb	sp!, {r4, r5, r6, r7, r8, r9}
@@ -181,94 +153,7 @@ ENTRY(armv7_idcache_wbinv_all)
 	ldmia	sp!, {lr}
 	RET
 END(armv7_idcache_wbinv_all)
-
-ENTRY(armv7_dcache_wb_range)
-	ldr	ip, .Larmv7_dcache_line_size
-	ldr	ip, [ip]
-	sub	r3, ip, #1
-	and	r2, r0, r3
-	add	r1, r1, r2
-	bic	r0, r0, r3
-.Larmv7_wb_next:
-	mcr	CP15_DCCMVAC(r0)
-	add	r0, r0, ip
-	subs	r1, r1, ip
-	bhi	.Larmv7_wb_next
-	dsb				/* data synchronization barrier */
-	RET
-END(armv7_dcache_wb_range)
-
-ENTRY(armv7_dcache_wbinv_range)
-	ldr     ip, .Larmv7_dcache_line_size
-	ldr     ip, [ip]
-	sub     r3, ip, #1
-	and     r2, r0, r3
-	add     r1, r1, r2
-	bic     r0, r0, r3
-.Larmv7_wbinv_next:
-	mcr	CP15_DCCIMVAC(r0)
-	add	r0, r0, ip
-	subs	r1, r1, ip
-	bhi	.Larmv7_wbinv_next
-	dsb				/* data synchronization barrier */
-	RET
-END(armv7_dcache_wbinv_range)
-
-/*
- * Note, we must not invalidate everything.  If the range is too big we
- * must use wb-inv of the entire cache.
- */
-ENTRY(armv7_dcache_inv_range)
-	ldr     ip, .Larmv7_dcache_line_size
-	ldr     ip, [ip]
-	sub     r3, ip, #1
-	and     r2, r0, r3
-	add     r1, r1, r2
-	bic     r0, r0, r3
-.Larmv7_inv_next:
-	mcr	CP15_DCIMVAC(r0)
-	add	r0, r0, ip
-	subs	r1, r1, ip
-	bhi	.Larmv7_inv_next
-	dsb				/* data synchronization barrier */
-	RET
-END(armv7_dcache_inv_range)
-
-ENTRY(armv7_idcache_wbinv_range)
-	ldr     ip, .Larmv7_idcache_line_size
-	ldr     ip, [ip]
-	sub     r3, ip, #1
-	and     r2, r0, r3
-	add     r1, r1, r2
-	bic     r0, r0, r3
-.Larmv7_id_wbinv_next:
-	mcr	CP15_ICIMVAU(r0)
-	mcr	CP15_DCCIMVAC(r0)
-	add	r0, r0, ip
-	subs	r1, r1, ip
-	bhi	.Larmv7_id_wbinv_next
-	dsb				/* data synchronization barrier */
-	isb				/* instruction synchronization barrier */
-	RET
-END(armv7_idcache_wbinv_range)
-
-
-ENTRY_NP(armv7_icache_sync_range)
-	ldr	ip, .Larmv7_icache_line_size
-	ldr	ip, [ip]
-	sub	r3, ip, #1		/* Address need not be aligned, but */
-	and	r2, r0, r3		/* round length up if op spans line */
-	add	r1, r1, r2		/* boundary: len += addr & linemask; */
-.Larmv7_sync_next:
-	mcr	CP15_DCCMVAC(r0)
-	mcr	CP15_ICIMVAU(r0)
-	add	r0, r0, ip
-	subs	r1, r1, ip
-	bhi	.Larmv7_sync_next
-	dsb				/* data synchronization barrier */
-	isb				/* instruction synchronization barrier */
-	RET
-END(armv7_icache_sync_range)
+#endif
 
 ENTRY(armv7_cpu_sleep)
 	dsb				/* data synchronization barrier */
@@ -276,22 +161,6 @@ ENTRY(armv7_cpu_sleep)
 	RET
 END(armv7_cpu_sleep)
 
-ENTRY(armv7_context_switch)
-	dsb
-	orr     r0, r0, #PT_ATTR
-
-	mcr	CP15_TTBR0(r0)
-	isb
-#ifdef SMP
-	mcr	CP15_TLBIALLIS
-#else
-	mcr	CP15_TLBIALL
-#endif
-	dsb
-	isb
-	RET
-END(armv7_context_switch)
-
 ENTRY(armv7_drain_writebuf)
 	dsb
 	RET
@@ -303,56 +172,3 @@ ENTRY(armv7_sev)
 	nop
 	RET
 END(armv7_sev)
-
-ENTRY(armv7_auxctrl)
-	mrc	CP15_ACTLR(r2)
-	bic r3, r2, r0	/* Clear bits */
-	eor r3, r3, r1  /* XOR bits */
-
-	teq r2, r3
-	mcrne	CP15_ACTLR(r3)
-	mov r0, r2
-	RET
-END(armv7_auxctrl)
-
-/*
- * Invalidate all I+D+branch cache.  Used by startup code, which counts
- * on the fact that only r0-r3,ip are modified and no stack space is used.
- */
-ENTRY(armv7_idcache_inv_all)
-	mov     r0, #0
-	mcr	CP15_CSSELR(r0)		@ set cache level to L1
-	mrc	CP15_CCSIDR(r0)
-
-	ubfx    r2, r0, #13, #15        @ get num sets - 1 from CCSIDR
-	ubfx    r3, r0, #3, #10         @ get numways - 1 from CCSIDR
-	clz     r1, r3                  @ number of bits to MSB of way
-	lsl     r3, r3, r1              @ shift into position
-	mov     ip, #1                  @
-	lsl     ip, ip, r1              @ ip now contains the way decr
-
-	ubfx    r0, r0, #0, #3          @ get linesize from CCSIDR
-	add     r0, r0, #4              @ apply bias
-	lsl     r2, r2, r0              @ shift sets by log2(linesize)
-	add     r3, r3, r2              @ merge numsets - 1 with numways - 1
-	sub     ip, ip, r2              @ subtract numsets - 1 from way decr
-	mov     r1, #1
-	lsl     r1, r1, r0              @ r1 now contains the set decr
-	mov     r2, ip                  @ r2 now contains set way decr
-
-	/* r3 = ways/sets, r2 = way decr, r1 = set decr, r0 and ip are free */
-1:	mcr	CP15_DCISW(r3)		@ invalidate line
-	movs    r0, r3                  @ get current way/set
-	beq     2f                      @ at 0 means we are done.
-	movs    r0, r0, lsl #10         @ clear way bits leaving only set bits
-	subne   r3, r3, r1              @ non-zero?, decrement set #
-	subeq   r3, r3, r2              @ zero?, decrement way # and restore set count
-	b       1b
-
-2:	dsb                             @ wait for stores to finish
-	mov     r0, #0                  @ and ...
-	mcr	CP15_ICIALLU		@ invalidate instruction+branch cache
-	isb                             @ instruction sync barrier
-	bx      lr                      @ return
-END(armv7_idcache_inv_all)
-

Modified: head/sys/arm/include/cpufunc.h
==============================================================================
--- head/sys/arm/include/cpufunc.h	Mon Oct  3 15:17:22 2016	(r306644)
+++ head/sys/arm/include/cpufunc.h	Mon Oct  3 16:10:38 2016	(r306645)
@@ -280,22 +280,11 @@ void	armv6_idcache_wbinv_all		(void);
 #endif
 #if defined(CPU_MV_PJ4B) || defined(CPU_CORTEXA) || defined(CPU_KRAIT)
 void	armv7_setttb			(u_int);
-void	armv7_tlb_flushID		(void);
-void	armv7_tlb_flushID_SE		(u_int);
-void	armv7_icache_sync_range		(vm_offset_t, vm_size_t);
-void	armv7_idcache_wbinv_range	(vm_offset_t, vm_size_t);
-void	armv7_idcache_inv_all		(void);
-void	armv7_dcache_wbinv_all		(void);
 void	armv7_idcache_wbinv_all		(void);
-void	armv7_dcache_wbinv_range	(vm_offset_t, vm_size_t);
-void	armv7_dcache_inv_range		(vm_offset_t, vm_size_t);
-void	armv7_dcache_wb_range		(vm_offset_t, vm_size_t);
 void	armv7_cpu_sleep			(int);
 void	armv7_setup			(void);
-void	armv7_context_switch		(void);
 void	armv7_drain_writebuf		(void);
 void	armv7_sev			(void);
-u_int	armv7_auxctrl			(u_int, u_int);
 
 void	armadaxp_idcache_wbinv_all	(void);
 
@@ -307,26 +296,9 @@ void	pj4bv7_setup			(void);
 #endif
 
 #if defined(CPU_ARM1176)
-void	arm11_tlb_flushID	(void);
-void	arm11_tlb_flushID_SE	(u_int);
-void	arm11_tlb_flushD	(void);
-void	arm11_tlb_flushD_SE	(u_int va);
-
-void	arm11_context_switch	(void);
-
 void	arm11_drain_writebuf	(void);
 
-void	armv6_dcache_wbinv_range	(vm_offset_t, vm_size_t);
-void	armv6_dcache_inv_range		(vm_offset_t, vm_size_t);
-void	armv6_dcache_wb_range		(vm_offset_t, vm_size_t);
-
-void	armv6_idcache_inv_all		(void);
-
 void    arm11x6_setttb                  (u_int);
-void    arm11x6_idcache_wbinv_all       (void);
-void    arm11x6_dcache_wbinv_all        (void);
-void    arm11x6_icache_sync_range       (vm_offset_t, vm_size_t);
-void    arm11x6_idcache_wbinv_range     (vm_offset_t, vm_size_t);
 void    arm11x6_setup                   (void);
 void    arm11x6_sleep                   (int);  /* no ref. for errata */
 #endif

Modified: head/sys/conf/Makefile.arm
==============================================================================
--- head/sys/conf/Makefile.arm	Mon Oct  3 15:17:22 2016	(r306644)
+++ head/sys/conf/Makefile.arm	Mon Oct  3 16:10:38 2016	(r306645)
@@ -92,6 +92,7 @@ ${KERNEL_KO}.tramp: ${KERNEL_KO} $S/$M/$
 	    echo "#define KERNSIZE $$st_size" >>opt_kernname.h
 	${CC} -O -nostdlib -I. -I$S \
 	    -Xlinker -T -Xlinker ldscript.$M.tramp \
+	    -DELF_TRAMPOLINE \
 	    tmphack.S \
 	    $S/$M/$M/elf_trampoline.c \
 	    $S/$M/$M/inckern.S \
@@ -99,6 +100,7 @@ ${KERNEL_KO}.tramp: ${KERNEL_KO} $S/$M/$
 	    -o ${KERNEL_KO}.tramp
 	${CC} -O -nostdlib -I. -I$S \
 	    -Xlinker -T -Xlinker ldscript.$M.tramp.noheader \
+	    -DELF_TRAMPOLINE \
 	    tmphack.S \
 	    $S/$M/$M/elf_trampoline.c \
 	    $S/$M/$M/inckern.S \
@@ -114,12 +116,12 @@ ${KERNEL_KO}.tramp: ${KERNEL_KO} $S/$M/$
 	eval $$(stat -s ${KERNEL_KO}.tmp.gz) && \
 	    echo "#define KERNCOMPSIZE $$st_size" >>opt_kernname.h
 	${CC} -O2 -ffreestanding -I. -I$S -c \
-	    -DKZIP \
+	    -DKZIP -DELF_TRAMPOLINE \
 	    $S/kern/inflate.c \
 	    -o inflate-tramp.o
 	${CC} -O -nostdlib -I. -I$S \
 	    -Xlinker -T -Xlinker ldscript.$M.tramp \
-	    -DKZIP \
+	    -DKZIP -DELF_TRAMPOLINE \
 	    tmphack.S \
 	    $S/$M/$M/elf_trampoline.c \
 	    inflate-tramp.o \
@@ -128,7 +130,7 @@ ${KERNEL_KO}.tramp: ${KERNEL_KO} $S/$M/$
 	    -o ${KERNEL_KO}.gz.tramp
 	${CC} -O -nostdlib -I. -I$S \
 	    -Xlinker -T -Xlinker ldscript.$M.tramp.noheader \
-	    -DKZIP \
+	    -DKZIP -DELF_TRAMPOLINE \
 	    tmphack.S \
 	    $S/$M/$M/elf_trampoline.c \
 	    inflate-tramp.o \


More information about the svn-src-head mailing list