mirror of
https://github.com/netwide-assembler/nasm.git
synced 2024-12-03 08:41:02 +08:00
f29123b936
When running in preprocess-only mode generate the equivalent of standard alignment using nops. This at the very least allows some kind of reasonable output and allows for dependency generation to proceed; the only way to *really* address this problem is to move alignment generation into the assembler proper; this would also allow the align/alignb distinction to be removed and handle padding with instructions which are more than one byte. This should resolve bug 3392319. Signed-off-by: H. Peter Anvin <hpa@linux.intel.com>
190 lines
6.9 KiB
Plaintext
190 lines
6.9 KiB
Plaintext
;; --------------------------------------------------------------------------
|
|
;;
|
|
;; Copyright 1996-2016 The NASM Authors - All Rights Reserved
|
|
;; See the file AUTHORS included with the NASM distribution for
|
|
;; the specific copyright holders.
|
|
;;
|
|
;; Redistribution and use in source and binary forms, with or without
|
|
;; modification, are permitted provided that the following
|
|
;; conditions are met:
|
|
;;
|
|
;; * Redistributions of source code must retain the above copyright
|
|
;; notice, this list of conditions and the following disclaimer.
|
|
;; * Redistributions in binary form must reproduce the above
|
|
;; copyright notice, this list of conditions and the following
|
|
;; disclaimer in the documentation and/or other materials provided
|
|
;; with the distribution.
|
|
;;
|
|
;; THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND
|
|
;; CONTRIBUTORS "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES,
|
|
;; INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF
|
|
;; MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
|
|
;; DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR
|
|
;; CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
|
|
;; SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT
|
|
;; NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
|
|
;; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
|
|
;; HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
|
|
;; CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR
|
|
;; OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE,
|
|
;; EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
|
;;
|
|
;; --------------------------------------------------------------------------
|
|
|
|
;
|
|
; Smart alignment macros
|
|
;
|
|
USE: smartalign
|
|
|
|
%imacro alignmode 1-2.nolist
|
|
%ifidni %1,nop
|
|
%define __ALIGN_JMP_THRESHOLD__ 16
|
|
|
|
%define __ALIGN_16BIT_1B__ 0x90
|
|
%define __ALIGN_16BIT_GROUP__ 1
|
|
|
|
%define __ALIGN_32BIT_1B__ 0x90
|
|
%define __ALIGN_32BIT_GROUP__ 1
|
|
|
|
%define __ALIGN_64BIT_1B__ 0x90
|
|
%define __ALIGN_64BIT_GROUP__ 1
|
|
%elifidni %1,generic
|
|
%define __ALIGN_JMP_THRESHOLD__ 8
|
|
|
|
%define __ALIGN_16BIT_1B__ 0x90
|
|
%define __ALIGN_16BIT_2B__ 0x89,0xf6
|
|
%define __ALIGN_16BIT_3B__ 0x8d,0x74,0x00
|
|
%define __ALIGN_16BIT_4B__ 0x8d,0xb4,0x00,0x00
|
|
%define __ALIGN_16BIT_5B__ 0x8d,0xb4,0x00,0x00,0x90
|
|
%define __ALIGN_16BIT_6B__ 0x8d,0xb4,0x00,0x00,0x89,0xff
|
|
%define __ALIGN_16BIT_7B__ 0x8d,0xb4,0x00,0x00,0x8d,0x7d,0x00
|
|
%define __ALIGN_16BIT_8B__ 0x8d,0xb4,0x00,0x00,0x8d,0xbd,0x00,0x00
|
|
%define __ALIGN_16BIT_GROUP__ 8
|
|
|
|
%define __ALIGN_32BIT_1B__ 0x90
|
|
%define __ALIGN_32BIT_2B__ 0x89,0xf6
|
|
%define __ALIGN_32BIT_3B__ 0x8d,0x76,0x00
|
|
%define __ALIGN_32BIT_4B__ 0x8d,0x74,0x26,0x00
|
|
%define __ALIGN_32BIT_5B__ 0x90,0x8d,0x74,0x26,0x00
|
|
%define __ALIGN_32BIT_6B__ 0x8d,0xb6,0x00,0x00,0x00,0x00
|
|
%define __ALIGN_32BIT_7B__ 0x8d,0xb4,0x26,0x00,0x00,0x00,0x00
|
|
%define __ALIGN_32BIT_GROUP__ 7
|
|
|
|
%define __ALIGN_64BIT_1B__ 0x90
|
|
%define __ALIGN_64BIT_2B__ 0x66,0x90
|
|
%define __ALIGN_64BIT_3B__ 0x66,0x66,0x90
|
|
%define __ALIGN_64BIT_4B__ 0x66,0x66,0x66,0x90
|
|
%define __ALIGN_64BIT_GROUP__ 4
|
|
%elifidni %1,k8
|
|
%define __ALIGN_JMP_THRESHOLD__ 16
|
|
|
|
%define __ALIGN_16BIT_1B__ 0x90
|
|
%define __ALIGN_16BIT_2B__ 0x66,0x90
|
|
%define __ALIGN_16BIT_3B__ 0x66,0x66,0x90
|
|
%define __ALIGN_16BIT_4B__ 0x66,0x66,0x66,0x90
|
|
%define __ALIGN_16BIT_GROUP__ 4
|
|
|
|
%define __ALIGN_32BIT_1B__ 0x90
|
|
%define __ALIGN_32BIT_2B__ 0x66,0x90
|
|
%define __ALIGN_32BIT_3B__ 0x66,0x66,0x90
|
|
%define __ALIGN_32BIT_4B__ 0x66,0x66,0x66,0x90
|
|
%define __ALIGN_32BIT_GROUP__ 4
|
|
|
|
%define __ALIGN_64BIT_1B__ 0x90
|
|
%define __ALIGN_64BIT_2B__ 0x66,0x90
|
|
%define __ALIGN_64BIT_3B__ 0x66,0x66,0x90
|
|
%define __ALIGN_64BIT_4B__ 0x66,0x66,0x66,0x90
|
|
%define __ALIGN_64BIT_GROUP__ 4
|
|
%elifidni %1,k7
|
|
%define __ALIGN_JMP_THRESHOLD__ 16
|
|
|
|
%define __ALIGN_16BIT_1B__ 0x90
|
|
%define __ALIGN_16BIT_2B__ 0x66,0x90
|
|
%define __ALIGN_16BIT_3B__ 0x66,0x66,0x90
|
|
%define __ALIGN_16BIT_4B__ 0x66,0x66,0x66,0x90
|
|
%define __ALIGN_64BIT_GROUP__ 4
|
|
|
|
%define __ALIGN_32BIT_1B__ 0x90
|
|
%define __ALIGN_32BIT_2B__ 0x8b,0xc0
|
|
%define __ALIGN_32BIT_3B__ 0x8d,0x04,0x20
|
|
%define __ALIGN_32BIT_4B__ 0x8d,0x44,0x20,0x00
|
|
%define __ALIGN_32BIT_5B__ 0x8d,0x44,0x20,0x00,0x90
|
|
%define __ALIGN_32BIT_6B__ 0x8d,0x80,0x00,0x00,0x00,0x00
|
|
%define __ALIGN_32BIT_7B__ 0x8d,0x04,0x05,0x00,0x00,0x00,0x00
|
|
%define __ALIGN_32BIT_GROUP__ 7
|
|
|
|
%define __ALIGN_64BIT_1B__ 0x90
|
|
%define __ALIGN_64BIT_2B__ 0x66,0x90
|
|
%define __ALIGN_64BIT_3B__ 0x66,0x66,0x90
|
|
%define __ALIGN_64BIT_4B__ 0x66,0x66,0x66,0x90
|
|
%define __ALIGN_64BIT_GROUP__ 4
|
|
%elifidni %1,p6
|
|
%define __ALIGN_JMP_THRESHOLD__ 16
|
|
|
|
%define __ALIGN_16BIT_1B__ 0x90
|
|
%define __ALIGN_16BIT_2B__ 0x66,0x90
|
|
%define __ALIGN_16BIT_3B__ 0x0f,0x1f,0x00
|
|
%define __ALIGN_16BIT_4B__ 0x0f,0x1f,0x40,0x00
|
|
%define __ALIGN_16BIT_GROUP__ 4
|
|
|
|
%define __ALIGN_32BIT_1B__ 0x90
|
|
%define __ALIGN_32BIT_2B__ 0x66,0x90
|
|
%define __ALIGN_32BIT_3B__ 0x0f,0x1f,0x00
|
|
%define __ALIGN_32BIT_4B__ 0x0f,0x1f,0x40,0x00
|
|
%define __ALIGN_32BIT_5B__ 0x0f,0x1f,0x44,0x00,0x00
|
|
%define __ALIGN_32BIT_6B__ 0x66,0x0f,0x1f,0x44,0x00,0x00
|
|
%define __ALIGN_32BIT_7B__ 0x0f,0x1f,0x80,0x00,0x00,0x00,0x00
|
|
%define __ALIGN_32BIT_8B__ 0x0f,0x1f,0x84,0x00,0x00,0x00,0x00,0x00
|
|
%define __ALIGN_32BIT_GROUP__ 8
|
|
|
|
%define __ALIGN_64BIT_1B__ 0x90
|
|
%define __ALIGN_64BIT_2B__ 0x66,0x90
|
|
%define __ALIGN_64BIT_3B__ 0x0f,0x1f,0x00
|
|
%define __ALIGN_64BIT_4B__ 0x0f,0x1f,0x40,0x00
|
|
%define __ALIGN_64BIT_5B__ 0x0f,0x1f,0x44,0x00,0x00
|
|
%define __ALIGN_64BIT_6B__ 0x66,0x0f,0x1f,0x44,0x00,0x00
|
|
%define __ALIGN_64BIT_7B__ 0x0f,0x1f,0x80,0x00,0x00,0x00,0x00
|
|
%define __ALIGN_64BIT_8B__ 0x0f,0x1f,0x84,0x00,0x00,0x00,0x00,0x00
|
|
%define __ALIGN_64BIT_GROUP__ 8
|
|
%else
|
|
%error unknown alignment mode: %1
|
|
%endif
|
|
%ifnempty %2
|
|
%ifidni %2,nojmp
|
|
%xdefine __ALIGN_JMP_THRESHOLD__ -1
|
|
%else
|
|
%xdefine __ALIGN_JMP_THRESHOLD__ %2
|
|
%endif
|
|
%endif
|
|
%xdefine __ALIGNMODE__ %1,__ALIGN_JMP_THRESHOLD__
|
|
%endmacro
|
|
|
|
%unimacro align 1-2+.nolist
|
|
%imacro align 1-2+.nolist
|
|
sectalign %1 ; align a segment as well
|
|
%ifnempty %2
|
|
times (((%1) - (($-$$) % (%1))) % (%1)) %2
|
|
%elif __PASS__ == 0 || __PASS__ == 3
|
|
times (((%1) - (($-$$) % (%1))) % (%1)) nop
|
|
%else
|
|
%push
|
|
%assign %$pad (((%1) - (($-$$) % (%1))) % (%1))
|
|
%if __ALIGN_JMP_THRESHOLD__ != -1 && %$pad > __ALIGN_JMP_THRESHOLD__
|
|
jmp %$end
|
|
; We can't re-use %$pad here as $ will have changed!
|
|
times (((%1) - (($-$$) % (%1))) % (%1)) nop
|
|
%$end:
|
|
%else
|
|
times (%$pad / __ALIGN_%[__BITS__]BIT_GROUP__) \
|
|
db __ALIGN_%[__BITS__]BIT_%[__ALIGN_%[__BITS__]BIT_GROUP__]B__
|
|
%assign %$pad %$pad % __ALIGN_%[__BITS__]BIT_GROUP__
|
|
%if %$pad > 0
|
|
db __ALIGN_%[__BITS__]BIT_%[%$pad]B__
|
|
%endif
|
|
%endif
|
|
%pop
|
|
%endif
|
|
%endmacro
|
|
|
|
alignmode generic
|