mirror of
				https://git.code.sf.net/p/libpng/code.git
				synced 2025-07-10 18:04:09 +02:00 
			
		
		
		
	
		
			
				
	
	
		
			6039 lines
		
	
	
		
			260 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			6039 lines
		
	
	
		
			260 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
| 
 | ||
| /* pnggccrd.c - mixed C/assembler version of utilities to read a PNG file
 | ||
|  *
 | ||
|  * For Intel/AMD x86 or x86-64 CPU (Pentium-MMX or later) and GNU C compiler.
 | ||
|  *
 | ||
|  * Last changed in libpng 1.2.19 July 31, 2007
 | ||
|  * For conditions of distribution and use, see copyright notice in png.h
 | ||
|  * Copyright (c) 1998 Intel Corporation
 | ||
|  * Copyright (c) 1999-2002,2007 Greg Roelofs
 | ||
|  * Copyright (c) 1998-2007 Glenn Randers-Pehrson
 | ||
|  *
 | ||
|  * Based on MSVC code contributed by Nirav Chhatrapati, Intel Corp., 1998.
 | ||
|  * Interface to libpng contributed by Gilles Vollant, 1999.
 | ||
|  * GNU C port by Greg Roelofs, 1999-2001.
 | ||
|  *
 | ||
|  * References:
 | ||
|  *
 | ||
|  *     http://www.intel.com/drg/pentiumII/appnotes/916/916.htm
 | ||
|  *     http://www.intel.com/drg/pentiumII/appnotes/923/923.htm
 | ||
|  *       [Intel's performance analysis of the MMX vs. non-MMX code;
 | ||
|  *        moved/deleted as of 2006, but text and some graphs still
 | ||
|  *        available via WayBack Machine at archive.org]
 | ||
|  *
 | ||
|  *     http://www.ibiblio.org/gferg/ldp/GCC-Inline-Assembly-HOWTO.html
 | ||
|  *     http://sam.zoy.org/blog/2007-04-13-shlib-with-non-pic-code-have-inline-assembly-and-pic-mix-well
 | ||
|  *     http://gcc.gnu.org/onlinedocs/gcc/Extended-Asm.html
 | ||
|  *     http://gcc.gnu.org/onlinedocs/gcc/Variable-Attributes.html
 | ||
|  *     http://gcc.gnu.org/onlinedocs/gcc/Function-Attributes.html
 | ||
|  *     AMD64 Architecture Programmer's Manual, volumes 1 and 5
 | ||
|  *       [http://www.amd.com/us-en/Processors/TechnicalResources/0,,30_182_739_7044,00.html]
 | ||
|  *     Intel 64 and IA-32 Software Developer's Manuals
 | ||
|  *       [http://developer.intel.com/products/processor/manuals/]
 | ||
|  *
 | ||
|  * png_read_filter_row_mmx_*() were converted in place with intel2gas 1.3.1:
 | ||
|  *
 | ||
|  *     intel2gas -mdI pnggccrd.c.partially-msvc -o pnggccrd.c
 | ||
|  *
 | ||
|  * and then cleaned up by hand.  See http://hermes.terminal.at/intel2gas/ .
 | ||
|  *
 | ||
|  * NOTE:  A sufficiently recent version of GNU as (or as.exe under DOS/Windows)
 | ||
|  * is required to assemble the newer asm instructions such as movq.  (Version
 | ||
|  * 2.5.2l.15 is definitely too old.)  See ftp://ftp.gnu.org/pub/gnu/binutils/ .
 | ||
|  */
 | ||
| 
 | ||
| /*
 | ||
|  * PORTING NOTES AND CHANGELOG (mostly by Greg Roelofs)
 | ||
|  * ===========================
 | ||
|  *
 | ||
|  * 19991006:
 | ||
|  *  - fixed sign error in post-MMX cleanup code (16- & 32-bit cases)
 | ||
|  *
 | ||
|  * 19991007:
 | ||
|  *  - additional optimizations (possible or definite):
 | ||
|  *     x [DONE] write MMX code for 64-bit case (pixel_bytes == 8) [not tested]
 | ||
|  *     - write MMX code for 48-bit case (pixel_bytes == 6)
 | ||
|  *     - figure out what's up with 24-bit case (pixel_bytes == 3):
 | ||
|  *        why subtract 8 from width_mmx in the pass 4/5 case?
 | ||
|  *        (only width_mmx case) (near line 2335)
 | ||
|  *     x [DONE] replace pixel_bytes within each block with the true
 | ||
|  *        constant value (or are compilers smart enough to do that?)
 | ||
|  *     - rewrite all MMX interlacing code so it's aligned with
 | ||
|  *        the *beginning* of the row buffer, not the end.  This
 | ||
|  *        would not only allow one to eliminate half of the memory
 | ||
|  *        writes for odd passes (that is, pass == odd), it may also
 | ||
|  *        eliminate some unaligned-data-access exceptions (assuming
 | ||
|  *        there's a penalty for not aligning 64-bit accesses on
 | ||
|  *        64-bit boundaries).  The only catch is that the "leftover"
 | ||
|  *        pixel(s) at the end of the row would have to be saved,
 | ||
|  *        but there are enough unused MMX registers in every case,
 | ||
|  *        so this is not a problem.  A further benefit is that the
 | ||
|  *        post-MMX cleanup code (C code) in at least some of the
 | ||
|  *        cases could be done within the assembler block.
 | ||
|  *  x [DONE] the "v3 v2 v1 v0 v7 v6 v5 v4" comments are confusing,
 | ||
|  *     inconsistent, and don't match the MMX Programmer's Reference
 | ||
|  *     Manual conventions anyway.  They should be changed to
 | ||
|  *     "b7 b6 b5 b4 b3 b2 b1 b0," where b0 indicates the byte that
 | ||
|  *     was lowest in memory (i.e., corresponding to a left pixel)
 | ||
|  *     and b7 is the byte that was highest (i.e., a right pixel).
 | ||
|  *
 | ||
|  * 19991016:
 | ||
|  *  - Brennan's Guide notwithstanding, gcc under Linux does *not*
 | ||
|  *     want globals prefixed by underscores when referencing them--
 | ||
|  *     i.e., if the variable is const4, then refer to it as const4,
 | ||
|  *     not _const4.  This seems to be a djgpp-specific requirement.
 | ||
|  *     Also, such variables apparently *must* be declared outside
 | ||
|  *     of functions; neither static nor automatic variables work if
 | ||
|  *     defined within the scope of a single function, but both
 | ||
|  *     static and truly global (multi-module) variables work fine.
 | ||
|  *
 | ||
|  * 19991017:
 | ||
|  *  - replaced pixel_bytes in each png_memcpy() call with constant value for
 | ||
|  *     inlining (png_do_read_interlace() "non-MMX/modified C code" block)
 | ||
|  *
 | ||
|  * 19991023:
 | ||
|  *  - fixed png_combine_row() non-MMX replication bug (odd passes only?)
 | ||
|  *  - switched from string-concatenation-with-macros to cleaner method of
 | ||
|  *     renaming global variables for djgpp--i.e., always use prefixes in
 | ||
|  *     inlined assembler code (== strings) and conditionally rename the
 | ||
|  *     variables, not the other way around.  Hence _const4, _mask8_0, etc.
 | ||
|  *
 | ||
|  * 19991024:
 | ||
|  *  - fixed mmxsupport()/png_do_read_interlace() first-row bug
 | ||
|  *     This one was severely weird:  even though mmxsupport() doesn't touch
 | ||
|  *     ebx (where "row" pointer was stored), it nevertheless managed to zero
 | ||
|  *     the register (even in static/non-fPIC code--see below), which in turn
 | ||
|  *     caused png_do_read_interlace() to return prematurely on the first row of
 | ||
|  *     interlaced images (i.e., without expanding the interlaced pixels).
 | ||
|  *     Inspection of the generated assembly code didn't turn up any clues,
 | ||
|  *     although it did point at a minor optimization (i.e., get rid of
 | ||
|  *     mmx_supported_local variable and just use eax).  Possibly the CPUID
 | ||
|  *     instruction is more destructive than it looks?  (Not yet checked.)
 | ||
|  *  - "info gcc" was next to useless, so compared fPIC and non-fPIC assembly
 | ||
|  *     listings...  Apparently register spillage has to do with ebx, since
 | ||
|  *     it's used to index the global offset table.  Commenting it out of the
 | ||
|  *     input-reg lists in png_combine_row() eliminated compiler barfage, so
 | ||
|  *     ifdef'd with __PIC__ macro:  if defined, use a global for unmask
 | ||
|  *
 | ||
|  * 19991107:
 | ||
|  *  - verified CPUID clobberage:  12-char string constant ("GenuineIntel",
 | ||
|  *     "AuthenticAMD", etc.) placed in ebx:ecx:edx.  Still need to polish.
 | ||
|  *
 | ||
|  * 19991120:
 | ||
|  *  - made "diff" variable (now "_dif") global to simplify conversion of
 | ||
|  *     filtering routines (running out of regs, sigh).  "diff" is still used
 | ||
|  *     in interlacing routines, however.
 | ||
|  *  - fixed up both versions of mmxsupport() (ORIG_THAT_USED_TO_CLOBBER_EBX
 | ||
|  *     macro determines which is used); original not yet tested.
 | ||
|  *
 | ||
|  * 20000213:
 | ||
|  *  - when compiling with gcc, be sure to use  -fomit-frame-pointer
 | ||
|  *
 | ||
|  * 20000319:
 | ||
|  *  - fixed a register-name typo in png_do_read_interlace(), default (MMX) case,
 | ||
|  *     pass == 4 or 5, that caused visible corruption of interlaced images
 | ||
|  *
 | ||
|  * 20000623:
 | ||
|  *  - Various problems were reported with gcc 2.95.2 in the Cygwin environment,
 | ||
|  *     many of the form "forbidden register 0 (ax) was spilled for class AREG."
 | ||
|  *     This is explained at http://gcc.gnu.org/fom_serv/cache/23.html, and
 | ||
|  *     Chuck Wilson supplied a patch involving dummy output registers.  See
 | ||
|  *     http://sourceforge.net/bugs/?func=detailbug&bug_id=108741&group_id=5624
 | ||
|  *     for the original (anonymous) SourceForge bug report.
 | ||
|  *
 | ||
|  * 20000706:
 | ||
|  *  - Chuck Wilson passed along these remaining gcc 2.95.2 errors:
 | ||
|  *       pnggccrd.c: In function `png_combine_row':
 | ||
|  *       pnggccrd.c:525: more than 10 operands in `asm'
 | ||
|  *       pnggccrd.c:669: more than 10 operands in `asm'
 | ||
|  *       pnggccrd.c:828: more than 10 operands in `asm'
 | ||
|  *       pnggccrd.c:994: more than 10 operands in `asm'
 | ||
|  *       pnggccrd.c:1177: more than 10 operands in `asm'
 | ||
|  *     They are all the same problem and can be worked around by using the
 | ||
|  *     global _unmask variable unconditionally, not just in the -fPIC case.
 | ||
|  *     Reportedly earlier versions of gcc also have the problem with more than
 | ||
|  *     10 operands; they just don't report it.  Much strangeness ensues, etc.
 | ||
|  *
 | ||
|  * 20000729:
 | ||
|  *  - enabled png_read_filter_row_mmx_up() (shortest remaining unconverted
 | ||
|  *     MMX routine); began converting png_read_filter_row_mmx_sub()
 | ||
|  *  - to finish remaining sections:
 | ||
|  *     - clean up indentation and comments
 | ||
|  *     - preload local variables
 | ||
|  *     - add output and input regs (order of former determines numerical
 | ||
|  *        mapping of latter)
 | ||
|  *     - avoid all usage of ebx (including bx, bh, bl) register [20000823]
 | ||
|  *     - remove "$" from addressing of Shift and Mask variables [20000823]
 | ||
|  *
 | ||
|  * 20000731:
 | ||
|  *  - global union vars causing segfaults in png_read_filter_row_mmx_sub()?
 | ||
|  *
 | ||
|  * 20000822:
 | ||
|  *  - ARGH, stupid png_read_filter_row_mmx_sub() segfault only happens with
 | ||
|  *     shared-library (-fPIC) version!  Code works just fine as part of static
 | ||
|  *     library.  Damn damn damn damn damn, should have tested that sooner.
 | ||
|  *     ebx is getting clobbered again (explicitly this time); need to save it
 | ||
|  *     on stack or rewrite asm code to avoid using it altogether.  Blargh!
 | ||
|  *
 | ||
|  * 20000823:
 | ||
|  *  - first section was trickiest; all remaining sections have ebx -> edx now.
 | ||
|  *     (-fPIC works again.)  Also added missing underscores to various Shift*
 | ||
|  *     and *Mask* globals and got rid of leading "$" signs.
 | ||
|  *
 | ||
|  * 20000826:
 | ||
|  *  - added visual separators to help navigate microscopic printed copies
 | ||
|  *     (http://pobox.com/~newt/code/gpr-latest.zip, mode 10); started working
 | ||
|  *     on png_read_filter_row_mmx_avg()
 | ||
|  *
 | ||
|  * 20000828:
 | ||
|  *  - finished png_read_filter_row_mmx_avg():  only Paeth left! (930 lines...)
 | ||
|  *     What the hell, did png_read_filter_row_mmx_paeth(), too.  Comments not
 | ||
|  *     cleaned up/shortened in either routine, but functionality is complete
 | ||
|  *     and seems to be working fine.
 | ||
|  *
 | ||
|  * 20000829:
 | ||
|  *  - ahhh, figured out last(?) bit of gcc/gas asm-fu:  if register is listed
 | ||
|  *     as an input reg (with dummy output variables, etc.), then it *cannot*
 | ||
|  *     also appear in the clobber list or gcc 2.95.2 will barf.  The solution
 | ||
|  *     is simple enough...
 | ||
|  *
 | ||
|  * 20000914:
 | ||
|  *  - bug in png_read_filter_row_mmx_avg():  16-bit grayscale not handled
 | ||
|  *     correctly (but 48-bit RGB just fine)
 | ||
|  *
 | ||
|  * 20000916:
 | ||
|  *  - fixed bug in png_read_filter_row_mmx_avg(), bpp == 2 case; three errors:
 | ||
|  *     - "_ShiftBpp.use = 24;"      should have been   "_ShiftBpp.use = 16;"
 | ||
|  *     - "_ShiftRem.use = 40;"      should have been   "_ShiftRem.use = 48;"
 | ||
|  *     - "psllq _ShiftRem, %%mm2"   should have been   "psrlq _ShiftRem, %%mm2"
 | ||
|  *
 | ||
|  * 20010101:
 | ||
|  *  - added new png_init_mmx_flags() function (here only because it needs to
 | ||
|  *     call mmxsupport(), which should probably become global png_mmxsupport());
 | ||
|  *     modified other MMX routines to run conditionally (png_ptr->asm_flags)
 | ||
|  *
 | ||
|  * 20010103:
 | ||
|  *  - renamed mmxsupport() to png_mmx_support(), with auto-set of mmx_supported,
 | ||
|  *     and made it public; moved png_init_mmx_flags() to png.c as internal func
 | ||
|  *
 | ||
|  * 20010104:
 | ||
|  *  - removed dependency on png_read_filter_row_c() (C code already duplicated
 | ||
|  *     within MMX version of png_read_filter_row()) so no longer necessary to
 | ||
|  *     compile it into pngrutil.o
 | ||
|  *
 | ||
|  * 20010310:
 | ||
|  *  - fixed buffer-overrun bug in png_combine_row() C code (non-MMX)
 | ||
|  *
 | ||
|  * 20010808:
 | ||
|  *  - added PNG_THREAD_UNSAFE_OK around code using global variables [GR-P]
 | ||
|  *
 | ||
|  * 20011124:
 | ||
|  *  - fixed missing save of Eflag in png_mmx_support() [Maxim Sobolev]
 | ||
|  *
 | ||
|  * 20020304:
 | ||
|  *  - eliminated incorrect use of width_mmx in pixel_bytes == 8 case
 | ||
|  *
 | ||
|  * 20020407:
 | ||
|  *  - fixed insufficient preservation of ebx register [Sami Farin]
 | ||
|  *
 | ||
|  * 20040724:
 | ||
|  *  - more tinkering with clobber list at lines 4529 and 5033 to get it to
 | ||
|  *     compile with gcc 3.4 [GR-P]
 | ||
|  *
 | ||
|  * 20040809:
 | ||
|  *  - added "rim" definitions for CONST4 and CONST6 [GR-P]
 | ||
|  *
 | ||
|  * 20060303:
 | ||
|  *  - added "OS2" to list of systems that don't need leading underscores [GR-P]
 | ||
|  *
 | ||
|  * 20060320:
 | ||
|  *  - made PIC-compliant [Christian Aichinger]
 | ||
|  *
 | ||
|  * 20070313:
 | ||
|  *  - finally applied Giuseppe Ghib<69>'s 64-bit patch of 20060803 (completely
 | ||
|  *     overlooked Dylan Alex Simon's similar patch of 20060414, oops...)
 | ||
|  *
 | ||
|  * 20070524:
 | ||
|  *  - fixed link failure caused by asm-only variables being optimized out
 | ||
|  *     (identified by Dimitri of Trolltech) with __attribute__((used)), which
 | ||
|  *     also gets rid of warnings => nuked ugly png_squelch_warnings() hack
 | ||
|  *  - dropped redundant ifdef
 | ||
|  *  - moved png_mmx_support() back up where originally intended (as in
 | ||
|  *     pngvcrd.c), using __attribute__((noinline)) in extra prototype
 | ||
|  *
 | ||
|  * 20070527:
 | ||
|  *  - revised png_combine_row() to reuse mask in lieu of external _unmask
 | ||
|  *  - moved 32-bit (RGBA) case to top of png_combine_row():  most common
 | ||
|  *  - just about ready to give up on x86-64 -fPIC mode; can't even access 16
 | ||
|  *     _mask*_* constants without triggering link error on shared library:
 | ||
|  *       /usr/bin/ld: pnggccrd.pic.o: relocation R_X86_64_32S against `a local
 | ||
|  *         symbol' can not be used when making a shared object; recompile with
 | ||
|  *         -fPIC
 | ||
|  *       pnggccrd.pic.o: could not read symbols: Bad value
 | ||
|  *       ("objdump -x pnggccrd.pic.o | grep rodata" to verify)
 | ||
|  *     [might be able to work around by doing within assembly code whatever
 | ||
|  *     -fPIC does, but given problems to date, seems like long shot...]
 | ||
|  *     [relevant ifdefs:  __x86_64__ && __PIC__ => C code only]
 | ||
|  *  - changed #if 0 to #ifdef PNG_CLOBBER_MMX_REGS_SUPPORTED in case gcc ever
 | ||
|  *     supports MMX regs (%mm0, etc.) in clobber list (not supported by gcc
 | ||
|  *     2.7.2.3, 2.91.66 (egcs 1.1.2), 3.x, or 4.1.2)
 | ||
|  *
 | ||
|  * 20070603:
 | ||
|  *  - revised png_combine_row() to use @GOTPCREL(%%rip) addressing on _c64
 | ||
|  *     struct of _mask*_* constants for x86-64 -fPIC; see sam.zoy.org link
 | ||
|  *     above for details
 | ||
|  *  - moved _const4 and _const6 into _c64 struct, renamed to _amask5_3_0 and
 | ||
|  *     _amask7_1_0, respectively
 | ||
|  *  - can't figure out how to use _c64._mask*_* vars within asm code, so still
 | ||
|  *     need single variables for non-x86-64/-fPIC half :-(
 | ||
|  *  - replaced various __PIC__ ifdefs with *_GOT_ebx macros
 | ||
|  *  - moved _LBCarryMask and _HBClearMask into _c64 struct
 | ||
|  *  - conditionally replaced _p*temp variables with %r11d-%r13d (via p*_TEMP
 | ||
|  *     and CLOBBER_r1*d macros)
 | ||
|  *
 | ||
|  * 20070604:
 | ||
|  *  - replaced all _ActiveMask and _ActiveMaskEnd with new _amask*_*_* consts
 | ||
|  *     (_amask naming convention:  numbers of 00-bytes, ff-bytes, 00-bytes)
 | ||
|  *    - _ActiveMask     // (10) // avg/paeth/sub; read-only; consts; movq/pand
 | ||
|  *       0x0000000000ffffffLL (bpp 3, avg)      _amask5_3_0
 | ||
|  *       0xffffffffffffffffLL (bpp 4, 6, avg)   _amask0_8_0
 | ||
|  *       0x000000000000ffffLL (bpp 2, avg)      _amask6_2_0
 | ||
|  *       0x0000000000ffffffLL (bpp 3, paeth)    _amask5_3_0
 | ||
|  *       0x00000000ffffffffLL (bpp 6, paeth)    _amask4_4_0
 | ||
|  *       0x00000000ffffffffLL (bpp 4, paeth)    _amask4_4_0
 | ||
|  *       0x00000000ffffffffLL (bpp 8, paeth)    _amask4_4_0
 | ||
|  *       0x0000ffffff000000LL (bpp 3, sub)      _amask2_3_3
 | ||
|  *       0x00000000ffff0000LL (bpp 2, sub)      _amask4_2_2
 | ||
|  *    - _ActiveMaskEnd  // (1)  // paeth only; read-only; const; pand
 | ||
|  *       0xffff000000000000LL (bpp 3, paeth)    _amask0_2_6
 | ||
|  *  - changed all "#if defined(__x86_64__) // later // && defined(__PIC__)"
 | ||
|  *     lines to "#ifdef PNG_x86_64_USE_GOTPCREL" for easier/safer testing
 | ||
|  *
 | ||
|  * 20070605:
 | ||
|  *  - merged PNG_x86_64_USE_GOTPCREL, non-PNG_x86_64_USE_GOTPCREL code via
 | ||
|  *     *MASK* and LOAD/RESTORE macros
 | ||
|  *
 | ||
|  * 20070607:
 | ||
|  *  - replaced all constant instances of _ShiftBpp, _ShiftRem with immediates
 | ||
|  *     (still have two shared cases in avg, sub routines)
 | ||
|  *
 | ||
|  * 20070609:
 | ||
|  *  - replaced remaining instances of _ShiftBpp, _ShiftRem with immediates
 | ||
|  *     (split sub and avg 4/6-bpp cases into separate blocks)
 | ||
|  *  - fixed paeth bug due to clobbered r11/r12/r13 regs
 | ||
|  *
 | ||
|  * 20070610:
 | ||
|  *  - made global "_dif" variable (avg/paeth/sub routines) local again (now
 | ||
|  *     "diff"--see 19991120 entry above), using register constraints
 | ||
|  *  - note that %ebp in clobber list doesn't actually work, at least for 32-bit
 | ||
|  *     version and gcc 4.1.2; must save and restore manually.  (Seems to work
 | ||
|  *     OK for 64-bit version and gcc 3.4.3, but gcc may not be using ebp/rbp
 | ||
|  *     in that case.)
 | ||
|  *  - started replacing direct _MMXLength accesses with register constraints
 | ||
|  *
 | ||
|  * 20070612:
 | ||
|  *  - continued replacing direct _MMXLength accesses with register constraints
 | ||
|  *
 | ||
|  * 20070613:
 | ||
|  *  - finished replacing direct _MMXLength accesses with register constraints;
 | ||
|  *     switched to local variable (and renamed back to MMXLength)
 | ||
|  *
 | ||
|  * 20070614:
 | ||
|  *  - fixed sub bpp = 1 bug
 | ||
|  *  - started replacing direct _FullLength accesses with register constraints
 | ||
|  *
 | ||
|  * 20070615:
 | ||
|  *  - fixed 64-bit paeth bpp 3 crash bug (misplaced LOAD_GOT_rbp)
 | ||
|  *  - fixed 64-bit paeth bpp 1/2 and cleanup-block crash bugs (misplaced
 | ||
|  *     RESTORE_r11_r12_r13)
 | ||
|  *  - slightly optimized avg/paeth cleanup blocks and paeth bpp 1/2 block
 | ||
|  *     (save/restore ebx only if needed)
 | ||
|  *  - continued replacing direct _FullLength accesses with register constraints
 | ||
|  *
 | ||
|  * 20070616:
 | ||
|  *  - finished replacing direct _FullLength accesses with register constraints
 | ||
|  *     (*ugly* conditional clobber-separator macros for avg and paeth, sigh)
 | ||
|  *
 | ||
|  * 20070618:
 | ||
|  *  - fixed misplaced PNG_THREAD_UNSAFE_OK endif (was missing LOAD_GOT_rbp/
 | ||
|  *     RESTORE_rbp in 32-bit thread-safe case)
 | ||
|  *  - changed all "ifdef *" to "if defined(*)" [GR-P]
 | ||
|  *
 | ||
|  * 20070619:
 | ||
|  *  - rearranged most bitdepth-related case statements to put most frequent
 | ||
|  *     cases at top (24-bit, 32-bit, 8-bit, rest)
 | ||
|  *
 | ||
|  * 20070623:
 | ||
|  *  - cleaned up png_debug() warnings/formatting
 | ||
|  *  - removed PNG_MMX_CODE_SUPPORTED ifdefs and added outer __GNUC__ ifdef
 | ||
|  *     (module no longer used by non-x86/non-GCC builds as of libpng 1.2.19)
 | ||
|  *  - removed single libpng-1.2.x PNG_DEBUG dependency on 1.0.x png_struct
 | ||
|  *     member (row_buf_size)
 | ||
|  *  - rearranged pass-related if-blocks in png_do_read_interlace() to put most
 | ||
|  *     frequent cases (4, 5) at top [GR-P suggestion]
 | ||
|  *
 | ||
|  * 20070624-29:
 | ||
|  *  - fixed 64-bit crash bug:  pointers -> rsi/rdi, not esi/edi (switched to
 | ||
|  *     %0/%1/%2/%3/%4 notation; eliminated size suffixes from relevant add/
 | ||
|  *     inc/sub/mov instructions; changed dummy vars to pointers)
 | ||
|  *     - png_combine_row()
 | ||
|  *     - png_do_read_interlace()
 | ||
|  *     - png_read_filter_row_mmx_avg()
 | ||
|  *     - png_read_filter_row_mmx_paeth()
 | ||
|  *     - png_read_filter_row_mmx_sub()
 | ||
|  *     - png_read_filter_row_mmx_up()
 | ||
|  *  - NOTE:  this fix makes use of the fact that modifying a 32-bit reg (e.g.,
 | ||
|  *     %%ebx) clears the top half of its corresponding 64-bit reg (%%rbx), so
 | ||
|  *     it's safe to mix 32-bit operations with 64-bit base/index addressing
 | ||
|  *     (see new PSI/PAX/PBX/PDX/PBP/etc. "pointer-register" macros); applies
 | ||
|  *     also to clobber lists
 | ||
|  *
 | ||
|  * 20070630:
 | ||
|  *  - cleaned up formatting, macros, minor png_read_filter_row_mmx_sub() 8-bpp
 | ||
|  *     register-usage inefficiency
 | ||
|  *  - fixed 32-bit png_do_read_interlace() bug (was using pointer size for
 | ||
|  *     64-bit dummy values)
 | ||
|  *
 | ||
|  * 20070703:
 | ||
|  *  - added check for (manual) PIC macro to fix OpenBSD crash bug
 | ||
|  *
 | ||
|  * 20070717:
 | ||
|  *  - fixed 48-bit png_combine_row() bug (was acting like 32-bit):  copy 6
 | ||
|  *     bytes per pixel, not 4, and use stride of 6, not 4, in the second loop
 | ||
|  *     of interlace processing of 48-bit pixels [GR-P]
 | ||
|  *
 | ||
|  * 20070722:
 | ||
|  *  - fixed 64-bit png_uint_32 bug with MMXLength/FullLength temp vars
 | ||
|  *
 | ||
|  * [still broken:  tops of all row-filter blocks (input/output constraints);
 | ||
|  *  shows up on 64-bit dynamic (-fPIC) version with -O2, especially if debug-
 | ||
|  *  printfs enabled, but at right edge of odd-width images even if disabled]
 | ||
|  *
 | ||
|  *
 | ||
|  * STILL TO DO:
 | ||
|  *  - fix final thread-unsafe code using stack vars and pointer? (paeth top,
 | ||
|  *     default, bottom only:  default, bottom already 5 reg constraints; could
 | ||
|  *     replace bpp with pointer and group bpp/patemp/pbtemp/pctemp in array)
 | ||
|  *  - fix ebp/no-reg-constraint inefficiency (avg/paeth/sub top)
 | ||
|  *  - test png_do_read_interlace() 64-bit case (pixel_bytes == 8)
 | ||
|  *  - write MMX code for 48-bit case (pixel_bytes == 6)
 | ||
|  *  - figure out what's up with 24-bit case (pixel_bytes == 3):
 | ||
|  *     why subtract 8 from width_mmx in the pass 4/5 case?  due to
 | ||
|  *     odd number of bytes? (only width_mmx case) (near line 2335)
 | ||
|  *  - rewrite all MMX interlacing code so it's aligned with beginning
 | ||
|  *     of the row buffer, not the end (see 19991007 for details)
 | ||
|  *  - add error messages to any remaining bogus default cases
 | ||
|  *  - enable pixel_depth == 8 cases in png_read_filter_row()? (test speed)
 | ||
|  *  - try =r, etc., as reg constraints?  (would gcc use 64-bit ones on x86-64?)
 | ||
|  *  - need full, non-graphical, CRC-based test suite...  maybe autogenerate
 | ||
|  *     random data of various height/width/depth, compute CRCs, write (C
 | ||
|  *     funcs), read (asm/MMX), recompute CRCs, and compare?
 | ||
|  *  - write true x86-64 version using 128-bit "media instructions", %xmm0-15,
 | ||
|  *     and extra general-purpose registers
 | ||
|  */
 | ||
| 
 | ||
| #if defined(__GNUC__)
 | ||
| 
 | ||
| #define PNG_INTERNAL
 | ||
| #include "png.h"
 | ||
| 
 | ||
| 
 | ||
| /* for some inexplicable reason, gcc 3.3.5 on OpenBSD (and elsewhere?) does
 | ||
|  * *not* define __PIC__ when the -fPIC option is used, so we have to rely on
 | ||
|  * makefiles and whatnot to define the PIC macro explicitly */
 | ||
| #if defined(PIC) && !defined(__PIC__)   // (this can/should move to pngconf.h)
 | ||
| #  define __PIC__
 | ||
| #endif
 | ||
| 
 | ||
| #if defined(PNG_ASSEMBLER_CODE_SUPPORTED) && defined(PNG_USE_PNGGCCRD)
 | ||
| 
 | ||
| /* if you want/need full thread-safety on x86-64 even when linking statically,
 | ||
|  * comment out the "&& defined(__PIC__)" part here: */
 | ||
| #if defined(__x86_64__) && defined(__PIC__)
 | ||
| #  define PNG_x86_64_USE_GOTPCREL            // GOTPCREL => full thread-safety
 | ||
| #  define PNG_CLOBBER_x86_64_REGS_SUPPORTED  // works as of gcc 3.4.3 ...
 | ||
| #endif
 | ||
| 
 | ||
| int PNGAPI png_mmx_support(void);
 | ||
| 
 | ||
| #if defined(PNG_USE_LOCAL_ARRAYS)
 | ||
| static PNG_CONST int FARDATA png_pass_start[7] = {0, 4, 0, 2, 0, 1, 0};
 | ||
| static PNG_CONST int FARDATA png_pass_inc[7]   = {8, 8, 4, 4, 2, 2, 1};
 | ||
| static PNG_CONST int FARDATA png_pass_width[7] = {8, 4, 4, 2, 2, 1, 1};
 | ||
| #endif
 | ||
| 
 | ||
| /* djgpp, Win32, Cygwin, and OS2 add their own underscores to global variables,
 | ||
|  * so define them without: */
 | ||
| #if defined(__DJGPP__) || defined(WIN32) || defined(__CYGWIN__) || \
 | ||
|     defined(__OS2__)
 | ||
| #  define _mmx_supported  mmx_supported
 | ||
| #  define _mask8_0        mask8_0
 | ||
| #  define _mask16_1       mask16_1
 | ||
| #  define _mask16_0       mask16_0
 | ||
| #  define _mask24_2       mask24_2
 | ||
| #  define _mask24_1       mask24_1
 | ||
| #  define _mask24_0       mask24_0
 | ||
| #  define _mask32_3       mask32_3
 | ||
| #  define _mask32_2       mask32_2
 | ||
| #  define _mask32_1       mask32_1
 | ||
| #  define _mask32_0       mask32_0
 | ||
| #  define _mask48_5       mask48_5
 | ||
| #  define _mask48_4       mask48_4
 | ||
| #  define _mask48_3       mask48_3
 | ||
| #  define _mask48_2       mask48_2
 | ||
| #  define _mask48_1       mask48_1
 | ||
| #  define _mask48_0       mask48_0
 | ||
| #  define _amask5_3_0     amask5_3_0
 | ||
| #  define _amask7_1_0     amask7_1_0
 | ||
| #  define _LBCarryMask    LBCarryMask
 | ||
| #  define _HBClearMask    HBClearMask
 | ||
| #  define _amask0_8_0     amask0_8_0
 | ||
| #  define _amask6_2_0     amask6_2_0
 | ||
| #  define _amask4_4_0     amask4_4_0
 | ||
| #  define _amask0_2_6     amask0_2_6
 | ||
| #  define _amask2_3_3     amask2_3_3
 | ||
| #  define _amask4_2_2     amask4_2_2
 | ||
| #  if defined(PNG_THREAD_UNSAFE_OK)
 | ||
| #    define _patemp       patemp
 | ||
| #    define _pbtemp       pbtemp
 | ||
| #    define _pctemp       pctemp
 | ||
| #  endif
 | ||
| #endif // djgpp, Win32, Cygwin, OS2
 | ||
| 
 | ||
| 
 | ||
| /* These constants are used in the inlined MMX assembly code. */
 | ||
| 
 | ||
| typedef unsigned long long  ull;
 | ||
| 
 | ||
| #if defined(PNG_x86_64_USE_GOTPCREL)
 | ||
| static PNG_CONST struct {
 | ||
|     //ull _mask_array[26];
 | ||
| 
 | ||
|     // png_combine_row() constants:
 | ||
|     ull _mask8_0;
 | ||
|     ull _mask16_0, _mask16_1;
 | ||
|     ull _mask24_0, _mask24_1, _mask24_2;
 | ||
|     ull _mask32_0, _mask32_1, _mask32_2, _mask32_3;
 | ||
|     ull _mask48_0, _mask48_1, _mask48_2, _mask48_3, _mask48_4, _mask48_5;
 | ||
| 
 | ||
|     // png_do_read_interlace() constants:
 | ||
|     ull _amask5_3_0, _amask7_1_0;  // was _const4 and _const6, respectively
 | ||
| 
 | ||
|     // png_read_filter_row_mmx_avg() constants (also uses _amask5_3_0):
 | ||
|     ull _LBCarryMask, _HBClearMask;
 | ||
|     ull _amask0_8_0, _amask6_2_0;  // was ActiveMask for bpp 4/6 and 2 cases
 | ||
| 
 | ||
|     // png_read_filter_row_mmx_paeth() constants (also uses _amask5_3_0):
 | ||
|     ull _amask4_4_0, _amask0_2_6;  // was ActiveMask{,End} for bpp 6/4/8 and 3
 | ||
| 
 | ||
|     // png_read_filter_row_mmx_sub() constants:
 | ||
|     ull _amask2_3_3, _amask4_2_2;  // was ActiveMask for bpp 3 and 2 cases
 | ||
| 
 | ||
| } _c64 __attribute__((used, aligned(8))) = {
 | ||
| 
 | ||
|     // png_combine_row() constants:
 | ||
|     0x0102040810204080LL, // _mask8_0      offset 0
 | ||
| 
 | ||
|     0x1010202040408080LL, // _mask16_0     offset 8
 | ||
|     0x0101020204040808LL, // _mask16_1     offset 16
 | ||
| 
 | ||
|     0x2020404040808080LL, // _mask24_0     offset 24
 | ||
|     0x0408080810101020LL, // _mask24_1     offset 32
 | ||
|     0x0101010202020404LL, // _mask24_2     offset 40
 | ||
| 
 | ||
|     0x4040404080808080LL, // _mask32_0     offset 48
 | ||
|     0x1010101020202020LL, // _mask32_1     offset 56
 | ||
|     0x0404040408080808LL, // _mask32_2     offset 64
 | ||
|     0x0101010102020202LL, // _mask32_3     offset 72
 | ||
| 
 | ||
|     0x4040808080808080LL, // _mask48_0     offset 80
 | ||
|     0x2020202040404040LL, // _mask48_1     offset 88
 | ||
|     0x1010101010102020LL, // _mask48_2     offset 96
 | ||
|     0x0404080808080808LL, // _mask48_3     offset 104
 | ||
|     0x0202020204040404LL, // _mask48_4     offset 112
 | ||
|     0x0101010101010202LL, // _mask48_5     offset 120
 | ||
| 
 | ||
|     // png_do_read_interlace() constants:
 | ||
|     0x0000000000FFFFFFLL, // _amask5_3_0   offset 128  (bpp 3, avg/paeth) const4
 | ||
|     0x00000000000000FFLL, // _amask7_1_0   offset 136                     const6
 | ||
| 
 | ||
|     // png_read_filter_row_mmx_avg() constants:
 | ||
|     0x0101010101010101LL, // _LBCarryMask  offset 144
 | ||
|     0x7F7F7F7F7F7F7F7FLL, // _HBClearMask  offset 152
 | ||
|     0xFFFFFFFFFFFFFFFFLL, // _amask0_8_0   offset 160  (bpp 4/6, avg)
 | ||
|     0x000000000000FFFFLL, // _amask6_2_0   offset 168  (bpp 2,   avg)
 | ||
| 
 | ||
|     // png_read_filter_row_mmx_paeth() constants:
 | ||
|     0x00000000FFFFFFFFLL, // _amask4_4_0   offset 176  (bpp 6/4/8, paeth)
 | ||
|     0xFFFF000000000000LL, // _amask0_2_6   offset 184  (bpp 3, paeth)   A.M.End
 | ||
| 
 | ||
|     // png_read_filter_row_mmx_sub() constants:
 | ||
|     0x0000FFFFFF000000LL, // _amask2_3_3   offset 192  (bpp 3, sub)
 | ||
|     0x00000000FFFF0000LL, // _amask4_2_2   offset 200  (bpp 2, sub)
 | ||
| 
 | ||
| };
 | ||
| 
 | ||
| #define MASK8_0        "(%%rbp)"
 | ||
| #define MASK16_0       "8(%%rbp)"
 | ||
| #define MASK16_1       "16(%%rbp)"
 | ||
| #define MASK24_0       "24(%%rbp)"
 | ||
| #define MASK24_1       "32(%%rbp)"
 | ||
| #define MASK24_2       "40(%%rbp)"
 | ||
| #define MASK32_0       "48(%%rbp)"
 | ||
| #define MASK32_1       "56(%%rbp)"
 | ||
| #define MASK32_2       "64(%%rbp)"
 | ||
| #define MASK32_3       "72(%%rbp)"
 | ||
| #define MASK48_0       "80(%%rbp)"
 | ||
| #define MASK48_1       "88(%%rbp)"
 | ||
| #define MASK48_2       "96(%%rbp)"
 | ||
| #define MASK48_3       "104(%%rbp)"
 | ||
| #define MASK48_4       "112(%%rbp)"
 | ||
| #define MASK48_5       "120(%%rbp)"
 | ||
| #define AMASK5_3_0     "128(%%rbp)"
 | ||
| #define AMASK7_1_0     "136(%%rbp)"
 | ||
| #define LB_CARRY_MASK  "144(%%rbp)"
 | ||
| #define HB_CLEAR_MASK  "152(%%rbp)"
 | ||
| #define AMASK0_8_0     "160(%%rbp)"
 | ||
| #define AMASK6_2_0     "168(%%rbp)"
 | ||
| #define AMASK4_4_0     "176(%%rbp)"
 | ||
| #define AMASK0_2_6     "184(%%rbp)"
 | ||
| #define AMASK2_3_3     "192(%%rbp)"
 | ||
| #define AMASK4_2_2     "200(%%rbp)"
 | ||
| 
 | ||
| #else // !PNG_x86_64_USE_GOTPCREL
 | ||
| 
 | ||
| static PNG_CONST ull _mask8_0  __attribute__((used, aligned(8))) = 0x0102040810204080LL;
 | ||
| 
 | ||
| static PNG_CONST ull _mask16_1 __attribute__((used, aligned(8))) = 0x0101020204040808LL;
 | ||
| static PNG_CONST ull _mask16_0 __attribute__((used, aligned(8))) = 0x1010202040408080LL;
 | ||
| 
 | ||
| static PNG_CONST ull _mask24_2 __attribute__((used, aligned(8))) = 0x0101010202020404LL;
 | ||
| static PNG_CONST ull _mask24_1 __attribute__((used, aligned(8))) = 0x0408080810101020LL;
 | ||
| static PNG_CONST ull _mask24_0 __attribute__((used, aligned(8))) = 0x2020404040808080LL;
 | ||
| 
 | ||
| static PNG_CONST ull _mask32_3 __attribute__((used, aligned(8))) = 0x0101010102020202LL;
 | ||
| static PNG_CONST ull _mask32_2 __attribute__((used, aligned(8))) = 0x0404040408080808LL;
 | ||
| static PNG_CONST ull _mask32_1 __attribute__((used, aligned(8))) = 0x1010101020202020LL;
 | ||
| static PNG_CONST ull _mask32_0 __attribute__((used, aligned(8))) = 0x4040404080808080LL;
 | ||
| 
 | ||
| static PNG_CONST ull _mask48_5 __attribute__((used, aligned(8))) = 0x0101010101010202LL;
 | ||
| static PNG_CONST ull _mask48_4 __attribute__((used, aligned(8))) = 0x0202020204040404LL;
 | ||
| static PNG_CONST ull _mask48_3 __attribute__((used, aligned(8))) = 0x0404080808080808LL;
 | ||
| static PNG_CONST ull _mask48_2 __attribute__((used, aligned(8))) = 0x1010101010102020LL;
 | ||
| static PNG_CONST ull _mask48_1 __attribute__((used, aligned(8))) = 0x2020202040404040LL;
 | ||
| static PNG_CONST ull _mask48_0 __attribute__((used, aligned(8))) = 0x4040808080808080LL;
 | ||
| 
 | ||
| // png_do_read_interlace() constants:
 | ||
| static PNG_CONST ull _amask5_3_0  __attribute__((aligned(8))) = 0x0000000000FFFFFFLL;  // was _const4
 | ||
| static PNG_CONST ull _amask7_1_0  __attribute__((aligned(8))) = 0x00000000000000FFLL;  // was _const6
 | ||
| 
 | ||
| // png_read_filter_row_mmx_avg() constants:
 | ||
| static PNG_CONST ull _LBCarryMask __attribute__((used, aligned(8))) = 0x0101010101010101LL;
 | ||
| static PNG_CONST ull _HBClearMask __attribute__((used, aligned(8))) = 0x7f7f7f7f7f7f7f7fLL;
 | ||
| static PNG_CONST ull _amask0_8_0  __attribute__((used, aligned(8))) = 0xFFFFFFFFFFFFFFFFLL;
 | ||
| static PNG_CONST ull _amask6_2_0  __attribute__((used, aligned(8))) = 0x000000000000FFFFLL;
 | ||
| 
 | ||
| // png_read_filter_row_mmx_paeth() constants:
 | ||
| static PNG_CONST ull _amask4_4_0  __attribute__((used, aligned(8))) = 0x00000000FFFFFFFFLL;
 | ||
| static PNG_CONST ull _amask0_2_6  __attribute__((used, aligned(8))) = 0xFFFF000000000000LL;
 | ||
| 
 | ||
| // png_read_filter_row_mmx_sub() constants:
 | ||
| static PNG_CONST ull _amask2_3_3  __attribute__((used, aligned(8))) = 0x0000FFFFFF000000LL;
 | ||
| static PNG_CONST ull _amask4_2_2  __attribute__((used, aligned(8))) = 0x00000000FFFF0000LL;
 | ||
| 
 | ||
| #define MASK8_0        "_mask8_0"
 | ||
| #define MASK16_0       "_mask16_0"
 | ||
| #define MASK16_1       "_mask16_1"
 | ||
| #define MASK24_0       "_mask24_0"
 | ||
| #define MASK24_1       "_mask24_1"
 | ||
| #define MASK24_2       "_mask24_2"
 | ||
| #define MASK32_0       "_mask32_0"
 | ||
| #define MASK32_1       "_mask32_1"
 | ||
| #define MASK32_2       "_mask32_2"
 | ||
| #define MASK32_3       "_mask32_3"
 | ||
| #define MASK48_0       "_mask48_0"
 | ||
| #define MASK48_1       "_mask48_1"
 | ||
| #define MASK48_2       "_mask48_2"
 | ||
| #define MASK48_3       "_mask48_3"
 | ||
| #define MASK48_4       "_mask48_4"
 | ||
| #define MASK48_5       "_mask48_5"
 | ||
| #define AMASK5_3_0     "_amask5_3_0"
 | ||
| #define AMASK7_1_0     "_amask7_1_0"
 | ||
| #define LB_CARRY_MASK  "_LBCarryMask"
 | ||
| #define HB_CLEAR_MASK  "_HBClearMask"
 | ||
| #define AMASK0_8_0     "_amask0_8_0"
 | ||
| #define AMASK6_2_0     "_amask6_2_0"
 | ||
| #define AMASK4_4_0     "_amask4_4_0"
 | ||
| #define AMASK0_2_6     "_amask0_2_6"
 | ||
| #define AMASK2_3_3     "_amask2_3_3"
 | ||
| #define AMASK4_2_2     "_amask4_2_2"
 | ||
| 
 | ||
| #endif // ?PNG_x86_64_USE_GOTPCREL
 | ||
| 
 | ||
| 
 | ||
| #if defined(PNG_HAVE_MMX_READ_FILTER_ROW) || defined(PNG_HAVE_MMX_COMBINE_ROW)
 | ||
| 
 | ||
| // this block is specific to png_read_filter_row_mmx_paeth() except for
 | ||
| // LOAD_GOT_rbp and RESTORE_rbp, which are also used in png_combine_row()
 | ||
| #if defined(PNG_x86_64_USE_GOTPCREL)
 | ||
| #  define pa_TEMP                "%%r11d"
 | ||
| #  define pb_TEMP                "%%r12d"
 | ||
| #  define pc_TEMP                "%%r13d"
 | ||
| #  if defined(PNG_CLOBBER_x86_64_REGS_SUPPORTED)  // works as of gcc 3.4.3 ...
 | ||
| #    define SAVE_r11_r12_r13
 | ||
| #    define RESTORE_r11_r12_r13
 | ||
| #    define _CLOBBER_r11_r12_r13 ,"%r11", "%r12", "%r13"
 | ||
| #    define CLOBBER_r11_r12_r13  "%r11", "%r12", "%r13"
 | ||
| #  else // !PNG_CLOBBER_x86_64_REGS_SUPPORTED
 | ||
| #    define SAVE_r11_r12_r13     "pushq %%r11  \n\t" \
 | ||
|                                  "pushq %%r12  \n\t" \
 | ||
|                                  "pushq %%r13  \n\t"  // "normally 0-extended"
 | ||
| #    define RESTORE_r11_r12_r13  "popq  %%r13  \n\t" \
 | ||
|                                  "popq  %%r12  \n\t" \
 | ||
|                                  "popq  %%r11  \n\t"
 | ||
| #    define _CLOBBER_r11_r12_r13
 | ||
| #    define CLOBBER_r11_r12_r13
 | ||
| #  endif
 | ||
| #  define LOAD_GOT_rbp           "pushq %%rbp                        \n\t" \
 | ||
|                                  "movq  _c64@GOTPCREL(%%rip), %%rbp  \n\t"
 | ||
| #  define RESTORE_rbp            "popq  %%rbp                        \n\t"
 | ||
| #else // 32-bit and/or non-PIC
 | ||
| #  if defined(PNG_THREAD_UNSAFE_OK)
 | ||
|      // These variables are used in png_read_filter_row_mmx_paeth() and would be
 | ||
|      //   local variables if not for gcc-inline-assembly addressing limitations
 | ||
|      //   (some apparently related to ELF format, others to CPU type).
 | ||
|      //
 | ||
|      // WARNING: Their presence defeats the thread-safety of libpng.
 | ||
|      static int                     _patemp  __attribute__((used));
 | ||
|      static int                     _pbtemp  __attribute__((used));
 | ||
|      static int                     _pctemp  __attribute__((used));
 | ||
| #    define pa_TEMP                "_patemp"
 | ||
| #    define pb_TEMP                "_pbtemp"  // temp variables for
 | ||
| #    define pc_TEMP                "_pctemp"  //  Paeth routine
 | ||
| #    define SAVE_r11_r12_r13
 | ||
| #    define RESTORE_r11_r12_r13
 | ||
| #    define _CLOBBER_r11_r12_r13   // not using regs => not clobbering
 | ||
| #    define CLOBBER_r11_r12_r13
 | ||
| #  endif // PNG_THREAD_UNSAFE_OK
 | ||
| #  define LOAD_GOT_rbp
 | ||
| #  define RESTORE_rbp
 | ||
| #endif
 | ||
| 
 | ||
| #if defined(__x86_64__)
 | ||
| #  define SAVE_ebp
 | ||
| #  define RESTORE_ebp
 | ||
| #  define _CLOBBER_ebp         ,"%ebp"
 | ||
| #  define CLOBBER_ebp          "%ebp"
 | ||
| #  define SAVE_FullLength      "movl %%eax, %%r15d  \n\t"
 | ||
| #  define RESTORE_FullLength   "movl %%r15d, "     // may go into eax or ecx
 | ||
| #  if defined(PNG_CLOBBER_x86_64_REGS_SUPPORTED)   // works as of gcc 3.4.3 ...
 | ||
| #    define SAVE_r15
 | ||
| #    define RESTORE_r15
 | ||
| #    define _CLOBBER_r15       ,"%r15"
 | ||
| #  else
 | ||
| #    define SAVE_r15           "pushq %%r15  \n\t"
 | ||
| #    define RESTORE_r15        "popq  %%r15  \n\t"
 | ||
| #    define _CLOBBER_r15
 | ||
| #  endif
 | ||
| #  define PBP                  "%%rbp"             // regs used for 64-bit
 | ||
| #  define PAX                  "%%rax"             //  pointers or in
 | ||
| #  define PBX                  "%%rbx"             //  combination with
 | ||
| #  define PCX                  "%%rcx"             //  64-bit pointer-regs
 | ||
| #  define PDX                  "%%rdx"             //  (base/index pairs,
 | ||
| #  define PSI                  "%%rsi"             //  add/sub/mov pairs)
 | ||
| #  define CLEAR_BOTTOM_3_BITS  "and  $0xfffffffffffffff8, "
 | ||
| #else
 | ||
| #  define SAVE_ebp             "pushl %%ebp \n\t"  // clobber list doesn't work
 | ||
| #  define RESTORE_ebp          "popl  %%ebp \n\t"  //  for %ebp on 32-bit; not
 | ||
| #  define _CLOBBER_ebp                             //  clear why not
 | ||
| #  define CLOBBER_ebp
 | ||
| #  define SAVE_FullLength      "pushl %%eax \n\t"
 | ||
| #  define RESTORE_FullLength   "popl "             // eax (avg) or ecx (paeth)
 | ||
| #  define SAVE_r15
 | ||
| #  define RESTORE_r15
 | ||
| #  define _CLOBBER_r15
 | ||
| #  define PBP                  "%%ebp"             // regs used for or in
 | ||
| #  define PAX                  "%%eax"             //  combination with
 | ||
| #  define PBX                  "%%ebx"             //  "normal," 32-bit
 | ||
| #  define PCX                  "%%ecx"             //  pointers
 | ||
| #  define PDX                  "%%edx"
 | ||
| #  define PSI                  "%%esi"
 | ||
| #  define CLEAR_BOTTOM_3_BITS  "and  $0xfffffff8, "
 | ||
| #endif
 | ||
| 
 | ||
| // CLOB_COMMA_ebx_ebp:  need comma ONLY if both CLOBBER_ebp and CLOBBER_GOT_ebx
 | ||
| //                      have values, i.e., only if __x86_64__ AND !__PIC__
 | ||
| #if defined(__x86_64__) && !defined(__PIC__)
 | ||
| #  define CLOB_COMMA_ebx_ebp    , // clobbering both ebp and ebx => need comma
 | ||
| #else
 | ||
| #  define CLOB_COMMA_ebx_ebp
 | ||
| #endif
 | ||
| 
 | ||
| // CLOB_COMMA_ebX_r1X:  need comma UNLESS both CLOBBER_ebp and CLOBBER_GOT_ebx
 | ||
| //                   are empty OR CLOBBER_r11_r12_r13 is empty--i.e., NO comma
 | ||
| //                   if (!__x86_64__ AND __PIC__) OR !(PNG_x86_64_USE_GOTPCREL
 | ||
| //                   AND PNG_CLOBBER_x86_64_REGS_SUPPORTED)   (double sigh...)
 | ||
| #if (!defined(__x86_64__) && defined(__PIC__)) || \
 | ||
|     !defined(PNG_x86_64_USE_GOTPCREL) || \
 | ||
|     !defined(PNG_CLOBBER_x86_64_REGS_SUPPORTED)
 | ||
| #  define CLOB_COMMA_ebX_r1X
 | ||
| #else
 | ||
| #  define CLOB_COMMA_ebX_r1X    , // clobbering (ebp OR ebx) AND r11_r12_r13
 | ||
| #endif
 | ||
| 
 | ||
| // CLOB_COLON_ebx_ebp:  need colon unless CLOBBER_ebp and CLOBBER_GOT_ebx are
 | ||
| //                      BOTH empty--i.e., NO colon if (!__x86_64__ AND __PIC__)
 | ||
| // CLOB_COLON_ebx_ebp_r1X:  if, in addition, CLOBBER_r11_r12_r13 is empty, then
 | ||
| //                          no colon for Paeth blocks, either--i.e., NO colon
 | ||
| //                          if !(PNG_x86_64_USE_GOTPCREL AND
 | ||
| //                               PNG_CLOBBER_x86_64_REGS_SUPPORTED)
 | ||
| #if (!defined(__x86_64__) && defined(__PIC__))
 | ||
| #  define CLOB_COLON_ebx_ebp
 | ||
| #  if !(defined(PNG_x86_64_USE_GOTPCREL) && \
 | ||
|         defined(PNG_CLOBBER_x86_64_REGS_SUPPORTED))
 | ||
| #    define CLOB_COLON_ebx_ebp_r1X
 | ||
| #  else
 | ||
| #    define CLOB_COLON_ebx_ebp_r1X  : // clobbering ebp OR ebx OR r11_r12_r13
 | ||
| #  endif
 | ||
| #else
 | ||
| #  define CLOB_COLON_ebx_ebp        : // clobbering ebp OR ebx
 | ||
| #  define CLOB_COLON_ebx_ebp_r1X    : // clobbering ebp OR ebx OR r11_r12_r13
 | ||
| #endif
 | ||
| 
 | ||
| #endif // PNG_HAVE_MMX_READ_FILTER_ROW
 | ||
| 
 | ||
| #if defined(__PIC__)  // macros to save, restore index to Global Offset Table
 | ||
| #  if defined(__x86_64__)
 | ||
| #    define SAVE_GOT_ebx     "pushq %%rbx \n\t"
 | ||
| #    define RESTORE_GOT_ebx  "popq  %%rbx \n\t"
 | ||
| #  else
 | ||
| #    define SAVE_GOT_ebx     "pushl %%ebx \n\t"
 | ||
| #    define RESTORE_GOT_ebx  "popl  %%ebx \n\t"
 | ||
| #  endif
 | ||
| #  define _CLOBBER_GOT_ebx   // explicitly saved, restored => not clobbered
 | ||
| #  define CLOBBER_GOT_ebx
 | ||
| #else
 | ||
| #  define SAVE_GOT_ebx
 | ||
| #  define RESTORE_GOT_ebx
 | ||
| #  define _CLOBBER_GOT_ebx   ,"%ebx"
 | ||
| #  define CLOBBER_GOT_ebx    "%ebx"
 | ||
| #endif
 | ||
| 
 | ||
| #if defined(PNG_HAVE_MMX_COMBINE_ROW) || defined(PNG_HAVE_MMX_READ_INTERLACE)
 | ||
| #  define BPP2  2
 | ||
| #  define BPP3  3  // bytes per pixel (a.k.a. pixel_bytes)
 | ||
| #  define BPP4  4  // (defined only to help avoid cut-and-paste errors)
 | ||
| #  define BPP6  6
 | ||
| #  define BPP8  8
 | ||
| #endif
 | ||
| 
 | ||
| 
 | ||
| 
 | ||
| static int _mmx_supported = 2; // 0: no MMX; 1: MMX supported; 2: not tested
 | ||
| 
 | ||
| /*===========================================================================*/
 | ||
| /*                                                                           */
 | ||
| /*                      P N G _ M M X _ S U P P O R T                        */
 | ||
| /*                                                                           */
 | ||
| /*===========================================================================*/
 | ||
| 
 | ||
| // GRR NOTES:  (1) the following code assumes 386 or better (pushfl/popfl)
 | ||
| //             (2) all instructions compile with gcc 2.7.2.3 and later
 | ||
| //           x (3) the function is moved down here to prevent gcc from
 | ||
| //           x      inlining it in multiple places and then barfing be-
 | ||
| //           x      cause the ".NOT_SUPPORTED" label is multiply defined
 | ||
| //                  [need to retest with gcc 2.7.2.3]
 | ||
| 
 | ||
| // GRR 20070524:  This declaration apparently is compatible with but supersedes
 | ||
| //   the one in png.h; in any case, the generated object file is slightly
 | ||
| //   smaller.  It is unnecessary with gcc 4.1.2, but gcc 2.x apparently
 | ||
| //   replicated the ".NOT_SUPPORTED" label in each location the function was
 | ||
| //   inlined, leading to compilation errors due to the "multiply defined"
 | ||
| //   label.  Old workaround was to leave the function at the end of this
 | ||
| //   file; new one (still testing) is to use a gcc-specific function attribute
 | ||
| //   to prevent local inlining.
 | ||
| int PNGAPI
 | ||
| png_mmx_support(void) __attribute__((noinline));
 | ||
| 
 | ||
| int PNGAPI
 | ||
| png_mmx_support(void)
 | ||
| {
 | ||
| #if defined(PNG_MMX_CODE_SUPPORTED)  // superfluous, but what the heck
 | ||
|     int result;
 | ||
|     __asm__ __volatile__ (
 | ||
| #if defined(__x86_64__)
 | ||
|         "pushq %%rbx          \n\t"  // rbx gets clobbered by CPUID instruction
 | ||
|         "pushq %%rcx          \n\t"  // so does rcx...
 | ||
|         "pushq %%rdx          \n\t"  // ...and rdx (but rcx & rdx safe on Linux)
 | ||
|         "pushfq               \n\t"  // save Eflag to stack
 | ||
|         "popq %%rax           \n\t"  // get Eflag from stack into rax
 | ||
|         "movq %%rax, %%rcx    \n\t"  // make another copy of Eflag in rcx
 | ||
|         "xorl $0x200000, %%eax \n\t" // toggle ID bit in Eflag (i.e., bit 21)
 | ||
|         "pushq %%rax          \n\t"  // save modified Eflag back to stack
 | ||
|         "popfq                \n\t"  // restore modified value to Eflag reg
 | ||
|         "pushfq               \n\t"  // save Eflag to stack
 | ||
|         "popq %%rax           \n\t"  // get Eflag from stack
 | ||
|         "pushq %%rcx          \n\t"  // save original Eflag to stack
 | ||
|         "popfq                \n\t"  // restore original Eflag
 | ||
| #else
 | ||
|         "pushl %%ebx          \n\t"  // ebx gets clobbered by CPUID instruction
 | ||
|         "pushl %%ecx          \n\t"  // so does ecx...
 | ||
|         "pushl %%edx          \n\t"  // ...and edx (but ecx & edx safe on Linux)
 | ||
|         "pushfl               \n\t"  // save Eflag to stack
 | ||
|         "popl %%eax           \n\t"  // get Eflag from stack into eax
 | ||
|         "movl %%eax, %%ecx    \n\t"  // make another copy of Eflag in ecx
 | ||
|         "xorl $0x200000, %%eax \n\t" // toggle ID bit in Eflag (i.e., bit 21)
 | ||
|         "pushl %%eax          \n\t"  // save modified Eflag back to stack
 | ||
|         "popfl                \n\t"  // restore modified value to Eflag reg
 | ||
|         "pushfl               \n\t"  // save Eflag to stack
 | ||
|         "popl %%eax           \n\t"  // get Eflag from stack
 | ||
|         "pushl %%ecx          \n\t"  // save original Eflag to stack
 | ||
|         "popfl                \n\t"  // restore original Eflag
 | ||
| #endif
 | ||
|         "xorl %%ecx, %%eax    \n\t"  // compare new Eflag with original Eflag
 | ||
|         "jz 0f                \n\t"  // if same, CPUID instr. is not supported
 | ||
| 
 | ||
|         "xorl %%eax, %%eax    \n\t"  // set eax to zero
 | ||
| //      ".byte  0x0f, 0xa2    \n\t"  // CPUID instruction (two-byte opcode)
 | ||
|         "cpuid                \n\t"  // get the CPU identification info
 | ||
|         "cmpl $1, %%eax       \n\t"  // make sure eax return non-zero value
 | ||
|         "jl 0f                \n\t"  // if eax is zero, MMX is not supported
 | ||
| 
 | ||
|         "xorl %%eax, %%eax    \n\t"  // set eax to zero and...
 | ||
|         "incl %%eax           \n\t"  // ...increment eax to 1.  This pair is
 | ||
|                                      // faster than the instruction "mov eax, 1"
 | ||
|         "cpuid                \n\t"  // get the CPU identification info again
 | ||
|         "andl $0x800000, %%edx \n\t" // mask out all bits but MMX bit (23)
 | ||
|         "cmpl $0, %%edx       \n\t"  // 0 = MMX not supported
 | ||
|         "jz 0f                \n\t"  // non-zero = yes, MMX IS supported
 | ||
| 
 | ||
|         "movl $1, %%eax       \n\t"  // set return value to 1
 | ||
|         "jmp  1f              \n\t"  // DONE:  have MMX support
 | ||
| 
 | ||
|     "0:                       \n\t"  // .NOT_SUPPORTED: target label for jump instructions
 | ||
|         "movl $0, %%eax       \n\t"  // set return value to 0
 | ||
|     "1:                       \n\t"  // .RETURN: target label for jump instructions
 | ||
| #if defined(__x86_64__)
 | ||
|         "popq %%rdx           \n\t"  // restore rdx
 | ||
|         "popq %%rcx           \n\t"  // restore rcx
 | ||
|         "popq %%rbx           \n\t"  // restore rbx
 | ||
| #else
 | ||
|         "popl %%edx           \n\t"  // restore edx
 | ||
|         "popl %%ecx           \n\t"  // restore ecx
 | ||
|         "popl %%ebx           \n\t"  // restore ebx
 | ||
| #endif
 | ||
| 
 | ||
| //      "ret                  \n\t"  // DONE:  no MMX support
 | ||
|                                      // (fall through to standard C "ret")
 | ||
| 
 | ||
|         : "=a" (result)              // output list
 | ||
| 
 | ||
|         :                            // any variables used on input (none)
 | ||
| 
 | ||
|                                      // no clobber list
 | ||
| //      , "%ebx", "%ecx", "%edx"     // GRR:  we handle these manually
 | ||
| //      , "memory"   // if write to a variable gcc thought was in a reg
 | ||
| //      , "cc"       // "condition codes" (flag bits)
 | ||
|     );
 | ||
|     _mmx_supported = result;
 | ||
| #else
 | ||
|     _mmx_supported = 0;
 | ||
| #endif /* PNG_MMX_CODE_SUPPORTED */
 | ||
| 
 | ||
|     return _mmx_supported;
 | ||
| }
 | ||
| 
 | ||
| 
 | ||
| /*===========================================================================*/
 | ||
| /*                                                                           */
 | ||
| /*                       P N G _ C O M B I N E _ R O W                       */
 | ||
| /*                                                                           */
 | ||
| /*===========================================================================*/
 | ||
| 
 | ||
| #if defined(PNG_HAVE_MMX_COMBINE_ROW)
 | ||
| 
 | ||
| /* Combines the row recently read in with the previous row.
 | ||
|    This routine takes care of alpha and transparency if requested.
 | ||
|    This routine also handles the two methods of progressive display
 | ||
|    of interlaced images, depending on the mask value.
 | ||
|    The mask value describes which pixels are to be combined with
 | ||
|    the row.  The pattern always repeats every 8 pixels, so just 8
 | ||
|    bits are needed.  A one indicates the pixel is to be combined; a
 | ||
|    zero indicates the pixel is to be skipped.  This is in addition
 | ||
|    to any alpha or transparency value associated with the pixel.
 | ||
|    If you want all pixels to be combined, pass 0xff (255) in mask. */
 | ||
| 
 | ||
| /* Use this routine for the x86 platform - it uses a faster MMX routine
 | ||
|    if the machine supports MMX. */
 | ||
| 
 | ||
| void /* PRIVATE */
 | ||
| png_combine_row(png_structp png_ptr, png_bytep row, int mask)
 | ||
| {
 | ||
|    int dummy_value_a;    // fix 'forbidden register spilled' error
 | ||
|    int dummy_value_c;
 | ||
|    int dummy_value_d;
 | ||
|    png_bytep dummy_value_S;
 | ||
|    png_bytep dummy_value_D;
 | ||
| 
 | ||
|    png_debug(1, "in png_combine_row (pnggccrd.c)\n");
 | ||
| 
 | ||
|    if (_mmx_supported == 2) {
 | ||
| #if !defined(PNG_1_0_X)
 | ||
|        /* this should have happened in png_init_mmx_flags() already */
 | ||
|        png_warning(png_ptr, "asm_flags may not have been initialized");
 | ||
| #endif
 | ||
|        png_mmx_support();
 | ||
|    }
 | ||
| 
 | ||
|    if (mask == 0xff)
 | ||
|    {
 | ||
|       png_debug(2,"mask == 0xff:  doing single png_memcpy()\n");
 | ||
|       png_memcpy(row, png_ptr->row_buf + 1,
 | ||
|        (png_size_t)PNG_ROWBYTES(png_ptr->row_info.pixel_depth,png_ptr->width));
 | ||
|    }
 | ||
|    else   /* (png_combine_row() is never called with mask == 0) */
 | ||
|    {
 | ||
|       switch (png_ptr->row_info.pixel_depth)
 | ||
|       {
 | ||
|          case 24:       /* png_ptr->row_info.pixel_depth */
 | ||
|          {
 | ||
|             png_bytep srcptr;
 | ||
|             png_bytep dstptr;
 | ||
| 
 | ||
| #if !defined(PNG_1_0_X)
 | ||
|             if (png_ptr->asm_flags & PNG_ASM_FLAG_MMX_READ_COMBINE_ROW)
 | ||
| #else
 | ||
|             if (_mmx_supported)
 | ||
| #endif
 | ||
|             {
 | ||
|                png_uint_32 len;
 | ||
|                int diff;
 | ||
| 
 | ||
|                srcptr = png_ptr->row_buf + 1;
 | ||
|                dstptr = row;
 | ||
|                len  = png_ptr->width & ~7;          // reduce to multiple of 8
 | ||
|                diff = (int) (png_ptr->width & 7);   // amount lost
 | ||
| 
 | ||
|                __asm__ __volatile__ (
 | ||
|                   "not       %%edx            \n\t" // mask => unmask
 | ||
|                   "movd      %%edx, %%mm7     \n\t" // load bit pattern
 | ||
|                   "not       %%edx            \n\t" // unmask => mask for later
 | ||
|                   "psubb     %%mm6, %%mm6     \n\t" // zero mm6
 | ||
|                   "punpcklbw %%mm7, %%mm7     \n\t"
 | ||
|                   "punpcklwd %%mm7, %%mm7     \n\t"
 | ||
|                   "punpckldq %%mm7, %%mm7     \n\t" // fill reg with 8 masks
 | ||
| 
 | ||
|                   LOAD_GOT_rbp
 | ||
|                   "movq   " MASK24_0 ", %%mm0 \n\t" // _mask24_0 -> mm0
 | ||
|                   "movq   " MASK24_1 ", %%mm1 \n\t" // _mask24_1 -> mm1
 | ||
|                   "movq   " MASK24_2 ", %%mm2 \n\t" // _mask24_2 -> mm2
 | ||
|                   RESTORE_rbp
 | ||
| 
 | ||
|                   "pand      %%mm7, %%mm0     \n\t"
 | ||
|                   "pand      %%mm7, %%mm1     \n\t"
 | ||
|                   "pand      %%mm7, %%mm2     \n\t"
 | ||
| 
 | ||
|                   "pcmpeqb   %%mm6, %%mm0     \n\t"
 | ||
|                   "pcmpeqb   %%mm6, %%mm1     \n\t"
 | ||
|                   "pcmpeqb   %%mm6, %%mm2     \n\t"
 | ||
| 
 | ||
| // preload        "movl      len, %%ecx       \n\t" // load length of line
 | ||
| // preload        "movl      srcptr, %3       \n\t" // load source
 | ||
| // preload        "movl      dstptr, %4       \n\t" // load dest
 | ||
| 
 | ||
|                   "cmpl      $0, %%ecx        \n\t"
 | ||
|                   "jz        mainloop24end    \n\t"
 | ||
| 
 | ||
|                 "mainloop24:                  \n\t"
 | ||
|                   "movq      (%3), %%mm4      \n\t"
 | ||
|                   "pand      %%mm0, %%mm4     \n\t"
 | ||
|                   "movq      %%mm0, %%mm6     \n\t"
 | ||
|                   "movq      (%4), %%mm7      \n\t"
 | ||
|                   "pandn     %%mm7, %%mm6     \n\t"
 | ||
|                   "por       %%mm6, %%mm4     \n\t"
 | ||
|                   "movq      %%mm4, (%4)      \n\t"
 | ||
| 
 | ||
|                   "movq      8(%3), %%mm5     \n\t"
 | ||
|                   "pand      %%mm1, %%mm5     \n\t"
 | ||
|                   "movq      %%mm1, %%mm7     \n\t"
 | ||
|                   "movq      8(%4), %%mm6     \n\t"
 | ||
|                   "pandn     %%mm6, %%mm7     \n\t"
 | ||
|                   "por       %%mm7, %%mm5     \n\t"
 | ||
|                   "movq      %%mm5, 8(%4)     \n\t"
 | ||
| 
 | ||
|                   "movq      16(%3), %%mm6    \n\t"
 | ||
|                   "pand      %%mm2, %%mm6     \n\t"
 | ||
|                   "movq      %%mm2, %%mm4     \n\t"
 | ||
|                   "movq      16(%4), %%mm7    \n\t"
 | ||
|                   "pandn     %%mm7, %%mm4     \n\t"
 | ||
|                   "por       %%mm4, %%mm6     \n\t"
 | ||
|                   "movq      %%mm6, 16(%4)    \n\t"
 | ||
| 
 | ||
|                   "add       $24, %3          \n\t" // inc by 24 bytes processed
 | ||
|                   "add       $24, %4          \n\t"
 | ||
|                   "subl      $8, %%ecx        \n\t" // dec by 8 pixels processed
 | ||
| 
 | ||
|                   "ja        mainloop24       \n\t"
 | ||
| 
 | ||
|                 "mainloop24end:               \n\t"
 | ||
| // preload        "movl      diff, %%ecx      \n\t" // (diff is in eax)
 | ||
|                   "movl      %%eax, %%ecx     \n\t"
 | ||
|                   "cmpl      $0, %%ecx        \n\t"
 | ||
|                   "jz        end24            \n\t"
 | ||
| // preload        "movl      mask, %%edx      \n\t"
 | ||
|                   "sall      $24, %%edx       \n\t" // make low byte, high byte
 | ||
| 
 | ||
|                 "secondloop24:                \n\t"
 | ||
|                   "sall      %%edx            \n\t" // move high bit to CF
 | ||
|                   "jnc       skip24           \n\t" // if CF = 0
 | ||
|                   "movw      (%3), %%ax       \n\t"
 | ||
|                   "movw      %%ax, (%4)       \n\t"
 | ||
|                   "xorl      %%eax, %%eax     \n\t"
 | ||
|                   "movb      2(%3), %%al      \n\t"
 | ||
|                   "movb      %%al, 2(%4)      \n\t"
 | ||
| 
 | ||
|                 "skip24:                      \n\t"
 | ||
|                   "add       $3, %3           \n\t"
 | ||
|                   "add       $3, %4           \n\t"
 | ||
|                   "decl      %%ecx            \n\t"
 | ||
|                   "jnz       secondloop24     \n\t"
 | ||
| 
 | ||
|                 "end24:                       \n\t"
 | ||
|                   "EMMS                       \n\t" // DONE
 | ||
| 
 | ||
|                   : "=a" (dummy_value_a),           // output regs (dummy)
 | ||
|                     "=d" (dummy_value_d),
 | ||
|                     "=c" (dummy_value_c),
 | ||
|                     "=S" (dummy_value_S),
 | ||
|                     "=D" (dummy_value_D)
 | ||
| 
 | ||
|                   : "0" (diff),        // eax       // input regs
 | ||
|                     "1" (mask),        // edx
 | ||
|                     "2" (len),         // ecx
 | ||
| // was (unmask)     "b"    RESERVED    // ebx       // Global Offset Table idx
 | ||
|                     "3" (srcptr),      // esi/rsi
 | ||
|                     "4" (dstptr)       // edi/rdi
 | ||
| 
 | ||
| #if defined(PNG_CLOBBER_MMX_REGS_SUPPORTED)
 | ||
|                   : "%mm0", "%mm1", "%mm2"          // clobber list
 | ||
|                   , "%mm4", "%mm5", "%mm6", "%mm7"
 | ||
| #endif
 | ||
|                );
 | ||
|             }
 | ||
|             else /* not _mmx_supported - use modified C routine */
 | ||
|             {
 | ||
|                register png_uint_32 i;
 | ||
|                png_uint_32 initial_val = BPP3 * png_pass_start[png_ptr->pass];
 | ||
|                  /* png.c:  png_pass_start[] = {0, 4, 0, 2, 0, 1, 0}; */
 | ||
|                register int stride = BPP3 * png_pass_inc[png_ptr->pass];
 | ||
|                  /* png.c:  png_pass_inc[] = {8, 8, 4, 4, 2, 2, 1}; */
 | ||
|                register int rep_bytes = BPP3 * png_pass_width[png_ptr->pass];
 | ||
|                  /* png.c:  png_pass_width[] = {8, 4, 4, 2, 2, 1, 1}; */
 | ||
|                png_uint_32 len = png_ptr->width &~7;  /* reduce to mult. of 8 */
 | ||
|                int diff = (int) (png_ptr->width & 7); /* amount lost */
 | ||
|                register png_uint_32 final_val = BPP3 * len;   /* GRR bugfix */
 | ||
| 
 | ||
|                srcptr = png_ptr->row_buf + 1 + initial_val;
 | ||
|                dstptr = row + initial_val;
 | ||
| 
 | ||
|                for (i = initial_val; i < final_val; i += stride)
 | ||
|                {
 | ||
|                   png_memcpy(dstptr, srcptr, rep_bytes);
 | ||
|                   srcptr += stride;
 | ||
|                   dstptr += stride;
 | ||
|                }
 | ||
|                if (diff)  /* number of leftover pixels:  3 for pngtest */
 | ||
|                {
 | ||
|                   final_val += diff*BPP3;
 | ||
|                   for (; i < final_val; i += stride)
 | ||
|                   {
 | ||
|                      if (rep_bytes > (int)(final_val-i))
 | ||
|                         rep_bytes = (int)(final_val-i);
 | ||
|                      png_memcpy(dstptr, srcptr, rep_bytes);
 | ||
|                      srcptr += stride;
 | ||
|                      dstptr += stride;
 | ||
|                   }
 | ||
|                }
 | ||
|             } /* end of else (_mmx_supported) */
 | ||
| 
 | ||
|             break;
 | ||
|          }       /* end 24 bpp */
 | ||
| 
 | ||
|          // formerly claimed to be most common case (combining 32-bit RGBA),
 | ||
|          // but almost certainly less common than 24-bit RGB case
 | ||
|          case 32:       /* png_ptr->row_info.pixel_depth */
 | ||
|          {
 | ||
|             png_bytep srcptr;
 | ||
|             png_bytep dstptr;
 | ||
| 
 | ||
| #if !defined(PNG_1_0_X)
 | ||
|             if (png_ptr->asm_flags & PNG_ASM_FLAG_MMX_READ_COMBINE_ROW)
 | ||
| #else
 | ||
|             if (_mmx_supported)
 | ||
| #endif
 | ||
|             {
 | ||
|                png_uint_32 len;
 | ||
|                int diff;
 | ||
| 
 | ||
|                srcptr = png_ptr->row_buf + 1;
 | ||
|                dstptr = row;
 | ||
|                len  = png_ptr->width & ~7;          // reduce to multiple of 8
 | ||
|                diff = (int) (png_ptr->width & 7);   // amount lost
 | ||
| 
 | ||
|                __asm__ __volatile__ (
 | ||
|                   "not       %%edx            \n\t" // mask => unmask
 | ||
|                   "movd      %%edx, %%mm7     \n\t" // load bit pattern
 | ||
|                   "not       %%edx            \n\t" // unmask => mask for later
 | ||
|                   "psubb     %%mm6, %%mm6     \n\t" // zero mm6
 | ||
|                   "punpcklbw %%mm7, %%mm7     \n\t"
 | ||
|                   "punpcklwd %%mm7, %%mm7     \n\t"
 | ||
|                   "punpckldq %%mm7, %%mm7     \n\t" // fill reg with 8 masks
 | ||
| 
 | ||
|                   LOAD_GOT_rbp
 | ||
|                   "movq   " MASK32_0 ", %%mm0 \n\t" // _mask32_0
 | ||
|                   "movq   " MASK32_1 ", %%mm1 \n\t" // _mask32_1
 | ||
|                   "movq   " MASK32_2 ", %%mm2 \n\t" // _mask32_2
 | ||
|                   "movq   " MASK32_3 ", %%mm3 \n\t" // _mask32_3
 | ||
|                   RESTORE_rbp
 | ||
| 
 | ||
|                   "pand      %%mm7, %%mm0     \n\t"
 | ||
|                   "pand      %%mm7, %%mm1     \n\t"
 | ||
|                   "pand      %%mm7, %%mm2     \n\t"
 | ||
|                   "pand      %%mm7, %%mm3     \n\t"
 | ||
| 
 | ||
|                   "pcmpeqb   %%mm6, %%mm0     \n\t"
 | ||
|                   "pcmpeqb   %%mm6, %%mm1     \n\t"
 | ||
|                   "pcmpeqb   %%mm6, %%mm2     \n\t"
 | ||
|                   "pcmpeqb   %%mm6, %%mm3     \n\t"
 | ||
| 
 | ||
| // preload        "movl      len, %%ecx       \n\t" // load length of line
 | ||
| // preload        "movl      srcptr, %3       \n\t" // load source
 | ||
| // preload        "movl      dstptr, %4       \n\t" // load dest
 | ||
| 
 | ||
|                   "cmpl      $0, %%ecx        \n\t" // lcr
 | ||
|                   "jz        mainloop32end    \n\t"
 | ||
| 
 | ||
|                 "mainloop32:                  \n\t"
 | ||
|                   "movq      (%3), %%mm4      \n\t"
 | ||
|                   "pand      %%mm0, %%mm4     \n\t"
 | ||
|                   "movq      %%mm0, %%mm6     \n\t"
 | ||
|                   "movq      (%4), %%mm7      \n\t"
 | ||
|                   "pandn     %%mm7, %%mm6     \n\t"
 | ||
|                   "por       %%mm6, %%mm4     \n\t"
 | ||
|                   "movq      %%mm4, (%4)      \n\t"
 | ||
| 
 | ||
|                   "movq      8(%3), %%mm5     \n\t"
 | ||
|                   "pand      %%mm1, %%mm5     \n\t"
 | ||
|                   "movq      %%mm1, %%mm7     \n\t"
 | ||
|                   "movq      8(%4), %%mm6     \n\t"
 | ||
|                   "pandn     %%mm6, %%mm7     \n\t"
 | ||
|                   "por       %%mm7, %%mm5     \n\t"
 | ||
|                   "movq      %%mm5, 8(%4)     \n\t"
 | ||
| 
 | ||
|                   "movq      16(%3), %%mm6    \n\t"
 | ||
|                   "pand      %%mm2, %%mm6     \n\t"
 | ||
|                   "movq      %%mm2, %%mm4     \n\t"
 | ||
|                   "movq      16(%4), %%mm7    \n\t"
 | ||
|                   "pandn     %%mm7, %%mm4     \n\t"
 | ||
|                   "por       %%mm4, %%mm6     \n\t"
 | ||
|                   "movq      %%mm6, 16(%4)    \n\t"
 | ||
| 
 | ||
|                   "movq      24(%3), %%mm7    \n\t"
 | ||
|                   "pand      %%mm3, %%mm7     \n\t"
 | ||
|                   "movq      %%mm3, %%mm5     \n\t"
 | ||
|                   "movq      24(%4), %%mm4    \n\t"
 | ||
|                   "pandn     %%mm4, %%mm5     \n\t"
 | ||
|                   "por       %%mm5, %%mm7     \n\t"
 | ||
|                   "movq      %%mm7, 24(%4)    \n\t"
 | ||
| 
 | ||
|                   "add       $32, %3          \n\t" // inc by 32 bytes processed
 | ||
|                   "add       $32, %4          \n\t"
 | ||
|                   "subl      $8, %%ecx        \n\t" // dec by 8 pixels processed
 | ||
|                   "ja        mainloop32       \n\t"
 | ||
| 
 | ||
|                 "mainloop32end:               \n\t"
 | ||
| // preload        "movl      diff, %%ecx      \n\t" // (diff is in eax)
 | ||
|                   "movl      %%eax, %%ecx     \n\t"
 | ||
|                   "cmpl      $0, %%ecx        \n\t"
 | ||
|                   "jz        end32            \n\t"
 | ||
| // preload        "movl      mask, %%edx      \n\t"
 | ||
|                   "sall      $24, %%edx       \n\t" // low byte => high byte
 | ||
| 
 | ||
|                 "secondloop32:                \n\t"
 | ||
|                   "sall      %%edx            \n\t" // move high bit to CF
 | ||
|                   "jnc       skip32           \n\t" // if CF = 0
 | ||
|                   "movl      (%3), %%eax      \n\t"
 | ||
|                   "movl      %%eax, (%4)      \n\t"
 | ||
| 
 | ||
|                 "skip32:                      \n\t"
 | ||
|                   "add       $4, %3           \n\t"
 | ||
|                   "add       $4, %4           \n\t"
 | ||
|                   "decl      %%ecx            \n\t"
 | ||
|                   "jnz       secondloop32     \n\t"
 | ||
| 
 | ||
|                 "end32:                       \n\t"
 | ||
|                   "EMMS                       \n\t" // DONE
 | ||
| 
 | ||
|                   : "=a" (dummy_value_a),           // output regs (dummy)
 | ||
|                     "=d" (dummy_value_d),
 | ||
|                     "=c" (dummy_value_c),
 | ||
|                     "=S" (dummy_value_S),
 | ||
|                     "=D" (dummy_value_D)
 | ||
| 
 | ||
|                   : "0" (diff),        // eax       // input regs
 | ||
|                     "1" (mask),        // edx
 | ||
|                     "2" (len),         // ecx
 | ||
| // was (unmask)     "b"    RESERVED    // ebx       // Global Offset Table idx
 | ||
|                     "3" (srcptr),      // esi/rsi
 | ||
|                     "4" (dstptr)       // edi/rdi
 | ||
| 
 | ||
| #if defined(PNG_CLOBBER_MMX_REGS_SUPPORTED)
 | ||
|                   : "%mm0", "%mm1", "%mm2", "%mm3"  // clobber list
 | ||
|                   , "%mm4", "%mm5", "%mm6", "%mm7"
 | ||
| #endif
 | ||
|                );
 | ||
|             }
 | ||
|             else /* not _mmx_supported - use modified C routine */
 | ||
|             {
 | ||
|                register png_uint_32 i;
 | ||
|                png_uint_32 initial_val = BPP4 * png_pass_start[png_ptr->pass];
 | ||
|                  /* png.c:  png_pass_start[] = {0, 4, 0, 2, 0, 1, 0}; */
 | ||
|                register int stride = BPP4 * png_pass_inc[png_ptr->pass];
 | ||
|                  /* png.c:  png_pass_inc[] = {8, 8, 4, 4, 2, 2, 1}; */
 | ||
|                register int rep_bytes = BPP4 * png_pass_width[png_ptr->pass];
 | ||
|                  /* png.c:  png_pass_width[] = {8, 4, 4, 2, 2, 1, 1}; */
 | ||
|                png_uint_32 len = png_ptr->width &~7;  /* reduce to mult. of 8 */
 | ||
|                int diff = (int) (png_ptr->width & 7); /* amount lost */
 | ||
|                register png_uint_32 final_val = BPP4 * len;   /* GRR bugfix */
 | ||
| 
 | ||
|                srcptr = png_ptr->row_buf + 1 + initial_val;
 | ||
|                dstptr = row + initial_val;
 | ||
| 
 | ||
|                for (i = initial_val; i < final_val; i += stride)
 | ||
|                {
 | ||
|                   png_memcpy(dstptr, srcptr, rep_bytes);
 | ||
|                   srcptr += stride;
 | ||
|                   dstptr += stride;
 | ||
|                }
 | ||
|                if (diff)  /* number of leftover pixels:  3 for pngtest */
 | ||
|                {
 | ||
|                   final_val += diff*BPP4;
 | ||
|                   for (; i < final_val; i += stride)
 | ||
|                   {
 | ||
|                      if (rep_bytes > (int)(final_val-i))
 | ||
|                         rep_bytes = (int)(final_val-i);
 | ||
|                      png_memcpy(dstptr, srcptr, rep_bytes);
 | ||
|                      srcptr += stride;
 | ||
|                      dstptr += stride;
 | ||
|                   }
 | ||
|                }
 | ||
|             } /* end of else (_mmx_supported) */
 | ||
| 
 | ||
|             break;
 | ||
|          }       /* end 32 bpp */
 | ||
| 
 | ||
|          case 8:        /* png_ptr->row_info.pixel_depth */
 | ||
|          {
 | ||
|             png_bytep srcptr;
 | ||
|             png_bytep dstptr;
 | ||
| 
 | ||
| #if !defined(PNG_1_0_X)
 | ||
|             if (png_ptr->asm_flags & PNG_ASM_FLAG_MMX_READ_COMBINE_ROW)
 | ||
| #else
 | ||
|             if (_mmx_supported)
 | ||
| #endif
 | ||
|             {
 | ||
|                png_uint_32 len;
 | ||
|                int diff;
 | ||
| 
 | ||
|                srcptr = png_ptr->row_buf + 1;
 | ||
|                dstptr = row;
 | ||
|                len  = png_ptr->width & ~7;          // reduce to multiple of 8
 | ||
|                diff = (int) (png_ptr->width & 7);   // amount lost
 | ||
| 
 | ||
|                __asm__ __volatile__ (
 | ||
|                   "not       %%edx            \n\t" // mask => unmask
 | ||
|                   "movd      %%edx, %%mm7     \n\t" // load bit pattern
 | ||
|                   "not       %%edx            \n\t" // unmask => mask for later
 | ||
|                   "psubb     %%mm6, %%mm6     \n\t" // zero mm6
 | ||
|                   "punpcklbw %%mm7, %%mm7     \n\t"
 | ||
|                   "punpcklwd %%mm7, %%mm7     \n\t"
 | ||
|                   "punpckldq %%mm7, %%mm7     \n\t" // fill reg with 8 masks
 | ||
| 
 | ||
|                   LOAD_GOT_rbp
 | ||
|                   "movq   " MASK8_0 ", %%mm0  \n\t" // _mask8_0 -> mm0
 | ||
|                   RESTORE_rbp
 | ||
| 
 | ||
|                   "pand      %%mm7, %%mm0     \n\t" // nonzero if keep byte
 | ||
|                   "pcmpeqb   %%mm6, %%mm0     \n\t" // zeros->1s, v versa
 | ||
| 
 | ||
| // preload        "movl      len, %%ecx       \n\t" // load length of line
 | ||
| // preload        "movl      srcptr, %3       \n\t" // load source
 | ||
| // preload        "movl      dstptr, %4       \n\t" // load dest
 | ||
| 
 | ||
|                   "cmpl      $0, %%ecx        \n\t" // len == 0 ?
 | ||
|                   "je        mainloop8end     \n\t"
 | ||
| 
 | ||
|                 "mainloop8:                   \n\t"
 | ||
|                   "movq      (%3), %%mm4      \n\t" // *srcptr
 | ||
|                   "pand      %%mm0, %%mm4     \n\t"
 | ||
|                   "movq      %%mm0, %%mm6     \n\t"
 | ||
|                   "pandn     (%4), %%mm6      \n\t" // *dstptr
 | ||
|                   "por       %%mm6, %%mm4     \n\t"
 | ||
|                   "movq      %%mm4, (%4)      \n\t"
 | ||
|                   "add       $8, %3           \n\t" // inc by 8 bytes processed
 | ||
|                   "add       $8, %4           \n\t"
 | ||
|                   "subl      $8, %%ecx        \n\t" // dec by 8 pixels processed
 | ||
|                   "ja        mainloop8        \n\t"
 | ||
| 
 | ||
|                 "mainloop8end:                \n\t"
 | ||
| // preload        "movl      diff, %%ecx      \n\t" // (diff is in eax)
 | ||
|                   "movl      %%eax, %%ecx     \n\t"
 | ||
|                   "cmpl      $0, %%ecx        \n\t"
 | ||
|                   "jz        end8             \n\t"
 | ||
| // preload        "movl      mask, %%edx      \n\t"
 | ||
|                   "sall      $24, %%edx       \n\t" // make low byte, high byte
 | ||
| 
 | ||
|                 "secondloop8:                 \n\t"
 | ||
|                   "sall      %%edx            \n\t" // move high bit to CF
 | ||
|                   "jnc       skip8            \n\t" // if CF = 0
 | ||
|                   "movb      (%3), %%al       \n\t"
 | ||
|                   "movb      %%al, (%4)       \n\t"
 | ||
| 
 | ||
|                 "skip8:                       \n\t"
 | ||
|                   "inc       %3               \n\t"
 | ||
|                   "inc       %4               \n\t"
 | ||
|                   "decl      %%ecx            \n\t"
 | ||
|                   "jnz       secondloop8      \n\t"
 | ||
| 
 | ||
|                 "end8:                        \n\t"
 | ||
|                   "EMMS                       \n\t" // DONE
 | ||
| 
 | ||
|                   : "=a" (dummy_value_a),           // output regs (dummy)
 | ||
|                     "=d" (dummy_value_d),
 | ||
|                     "=c" (dummy_value_c),
 | ||
|                     "=S" (dummy_value_S),
 | ||
|                     "=D" (dummy_value_D)
 | ||
| 
 | ||
|                   : "0" (diff),        // eax       // input regs
 | ||
|                     "1" (mask),        // edx
 | ||
|                     "2" (len),         // ecx
 | ||
| // was (unmask)     "b"    RESERVED    // ebx       // Global Offset Table idx
 | ||
|                     "3" (srcptr),      // esi/rsi
 | ||
|                     "4" (dstptr)       // edi/rdi
 | ||
| 
 | ||
| #if defined(PNG_CLOBBER_MMX_REGS_SUPPORTED)
 | ||
|                   : "%mm0", "%mm4", "%mm6", "%mm7"  // clobber list
 | ||
| #endif
 | ||
|                );
 | ||
|             }
 | ||
|             else /* not _mmx_supported - use modified C routine */
 | ||
|             {
 | ||
|                register png_uint_32 i;
 | ||
|                png_uint_32 initial_val = png_pass_start[png_ptr->pass];
 | ||
|                  /* png.c:  png_pass_start[] = {0, 4, 0, 2, 0, 1, 0}; */
 | ||
|                register int stride = png_pass_inc[png_ptr->pass];
 | ||
|                  /* png.c:  png_pass_inc[] = {8, 8, 4, 4, 2, 2, 1}; */
 | ||
|                register int rep_bytes = png_pass_width[png_ptr->pass];
 | ||
|                  /* png.c:  png_pass_width[] = {8, 4, 4, 2, 2, 1, 1}; */
 | ||
|                png_uint_32 len = png_ptr->width &~7;  /* reduce to mult. of 8 */
 | ||
|                int diff = (int) (png_ptr->width & 7); /* amount lost */
 | ||
|                register png_uint_32 final_val = len;  /* GRR bugfix */
 | ||
| 
 | ||
|                srcptr = png_ptr->row_buf + 1 + initial_val;
 | ||
|                dstptr = row + initial_val;
 | ||
| 
 | ||
|                for (i = initial_val; i < final_val; i += stride)
 | ||
|                {
 | ||
|                   png_memcpy(dstptr, srcptr, rep_bytes);
 | ||
|                   srcptr += stride;
 | ||
|                   dstptr += stride;
 | ||
|                }
 | ||
|                if (diff)  /* number of leftover pixels:  3 for pngtest */
 | ||
|                {
 | ||
|                   final_val += diff /* *BPP1 */ ;
 | ||
|                   for (; i < final_val; i += stride)
 | ||
|                   {
 | ||
|                      if (rep_bytes > (int)(final_val-i))
 | ||
|                         rep_bytes = (int)(final_val-i);
 | ||
|                      png_memcpy(dstptr, srcptr, rep_bytes);
 | ||
|                      srcptr += stride;
 | ||
|                      dstptr += stride;
 | ||
|                   }
 | ||
|                }
 | ||
| 
 | ||
|             } /* end of else (_mmx_supported) */
 | ||
| 
 | ||
|             break;
 | ||
|          }       /* end 8 bpp */
 | ||
| 
 | ||
|          case 1:        /* png_ptr->row_info.pixel_depth */
 | ||
|          {
 | ||
|             png_bytep sp;
 | ||
|             png_bytep dp;
 | ||
|             int s_inc, s_start, s_end;
 | ||
|             int m;
 | ||
|             int shift;
 | ||
|             png_uint_32 i;
 | ||
| 
 | ||
|             sp = png_ptr->row_buf + 1;
 | ||
|             dp = row;
 | ||
|             m = 0x80;
 | ||
| #if defined(PNG_READ_PACKSWAP_SUPPORTED)
 | ||
|             if (png_ptr->transformations & PNG_PACKSWAP)
 | ||
|             {
 | ||
|                s_start = 0;
 | ||
|                s_end = 7;
 | ||
|                s_inc = 1;
 | ||
|             }
 | ||
|             else
 | ||
| #endif
 | ||
|             {
 | ||
|                s_start = 7;
 | ||
|                s_end = 0;
 | ||
|                s_inc = -1;
 | ||
|             }
 | ||
| 
 | ||
|             shift = s_start;
 | ||
| 
 | ||
|             for (i = 0; i < png_ptr->width; i++)
 | ||
|             {
 | ||
|                if (m & mask)
 | ||
|                {
 | ||
|                   int value;
 | ||
| 
 | ||
|                   value = (*sp >> shift) & 0x1;
 | ||
|                   *dp &= (png_byte)((0x7f7f >> (7 - shift)) & 0xff);
 | ||
|                   *dp |= (png_byte)(value << shift);
 | ||
|                }
 | ||
| 
 | ||
|                if (shift == s_end)
 | ||
|                {
 | ||
|                   shift = s_start;
 | ||
|                   sp++;
 | ||
|                   dp++;
 | ||
|                }
 | ||
|                else
 | ||
|                   shift += s_inc;
 | ||
| 
 | ||
|                if (m == 1)
 | ||
|                   m = 0x80;
 | ||
|                else
 | ||
|                   m >>= 1;
 | ||
|             }
 | ||
|             break;
 | ||
|          }       /* end 1 bpp */
 | ||
| 
 | ||
|          case 2:        /* png_ptr->row_info.pixel_depth */
 | ||
|          {
 | ||
|             png_bytep sp;
 | ||
|             png_bytep dp;
 | ||
|             int s_start, s_end, s_inc;
 | ||
|             int m;
 | ||
|             int shift;
 | ||
|             png_uint_32 i;
 | ||
|             int value;
 | ||
| 
 | ||
|             sp = png_ptr->row_buf + 1;
 | ||
|             dp = row;
 | ||
|             m = 0x80;
 | ||
| #if defined(PNG_READ_PACKSWAP_SUPPORTED)
 | ||
|             if (png_ptr->transformations & PNG_PACKSWAP)
 | ||
|             {
 | ||
|                s_start = 0;
 | ||
|                s_end = 6;
 | ||
|                s_inc = 2;
 | ||
|             }
 | ||
|             else
 | ||
| #endif
 | ||
|             {
 | ||
|                s_start = 6;
 | ||
|                s_end = 0;
 | ||
|                s_inc = -2;
 | ||
|             }
 | ||
| 
 | ||
|             shift = s_start;
 | ||
| 
 | ||
|             for (i = 0; i < png_ptr->width; i++)
 | ||
|             {
 | ||
|                if (m & mask)
 | ||
|                {
 | ||
|                   value = (*sp >> shift) & 0x3;
 | ||
|                   *dp &= (png_byte)((0x3f3f >> (6 - shift)) & 0xff);
 | ||
|                   *dp |= (png_byte)(value << shift);
 | ||
|                }
 | ||
| 
 | ||
|                if (shift == s_end)
 | ||
|                {
 | ||
|                   shift = s_start;
 | ||
|                   sp++;
 | ||
|                   dp++;
 | ||
|                }
 | ||
|                else
 | ||
|                   shift += s_inc;
 | ||
|                if (m == 1)
 | ||
|                   m = 0x80;
 | ||
|                else
 | ||
|                   m >>= 1;
 | ||
|             }
 | ||
|             break;
 | ||
|          }       /* end 2 bpp */
 | ||
| 
 | ||
|          case 4:        /* png_ptr->row_info.pixel_depth */
 | ||
|          {
 | ||
|             png_bytep sp;
 | ||
|             png_bytep dp;
 | ||
|             int s_start, s_end, s_inc;
 | ||
|             int m;
 | ||
|             int shift;
 | ||
|             png_uint_32 i;
 | ||
|             int value;
 | ||
| 
 | ||
|             sp = png_ptr->row_buf + 1;
 | ||
|             dp = row;
 | ||
|             m = 0x80;
 | ||
| #if defined(PNG_READ_PACKSWAP_SUPPORTED)
 | ||
|             if (png_ptr->transformations & PNG_PACKSWAP)
 | ||
|             {
 | ||
|                s_start = 0;
 | ||
|                s_end = 4;
 | ||
|                s_inc = 4;
 | ||
|             }
 | ||
|             else
 | ||
| #endif
 | ||
|             {
 | ||
|                s_start = 4;
 | ||
|                s_end = 0;
 | ||
|                s_inc = -4;
 | ||
|             }
 | ||
| 
 | ||
|             shift = s_start;
 | ||
| 
 | ||
|             for (i = 0; i < png_ptr->width; i++)
 | ||
|             {
 | ||
|                if (m & mask)
 | ||
|                {
 | ||
|                   value = (*sp >> shift) & 0xf;
 | ||
|                   *dp &= (png_byte)((0xf0f >> (4 - shift)) & 0xff);
 | ||
|                   *dp |= (png_byte)(value << shift);
 | ||
|                }
 | ||
| 
 | ||
|                if (shift == s_end)
 | ||
|                {
 | ||
|                   shift = s_start;
 | ||
|                   sp++;
 | ||
|                   dp++;
 | ||
|                }
 | ||
|                else
 | ||
|                   shift += s_inc;
 | ||
|                if (m == 1)
 | ||
|                   m = 0x80;
 | ||
|                else
 | ||
|                   m >>= 1;
 | ||
|             }
 | ||
|             break;
 | ||
|          }       /* end 4 bpp */
 | ||
| 
 | ||
|          case 16:       /* png_ptr->row_info.pixel_depth */
 | ||
|          {
 | ||
|             png_bytep srcptr;
 | ||
|             png_bytep dstptr;
 | ||
| 
 | ||
| #if !defined(PNG_1_0_X)
 | ||
|             if (png_ptr->asm_flags & PNG_ASM_FLAG_MMX_READ_COMBINE_ROW)
 | ||
| #else
 | ||
|             if (_mmx_supported)
 | ||
| #endif
 | ||
|             {
 | ||
|                png_uint_32 len;
 | ||
|                int diff;
 | ||
| 
 | ||
|                srcptr = png_ptr->row_buf + 1;
 | ||
|                dstptr = row;
 | ||
|                len  = png_ptr->width & ~7;          // reduce to multiple of 8
 | ||
|                diff = (int) (png_ptr->width & 7);   // amount lost
 | ||
| 
 | ||
|                __asm__ __volatile__ (
 | ||
|                   "not       %%edx            \n\t" // mask => unmask
 | ||
|                   "movd      %%edx, %%mm7     \n\t" // load bit pattern
 | ||
|                   "not       %%edx            \n\t" // unmask => mask for later
 | ||
|                   "psubb     %%mm6, %%mm6     \n\t" // zero mm6
 | ||
|                   "punpcklbw %%mm7, %%mm7     \n\t"
 | ||
|                   "punpcklwd %%mm7, %%mm7     \n\t"
 | ||
|                   "punpckldq %%mm7, %%mm7     \n\t" // fill reg with 8 masks
 | ||
| 
 | ||
|                   LOAD_GOT_rbp
 | ||
|                   "movq   " MASK16_0 ", %%mm0 \n\t" // _mask16_0 -> mm0
 | ||
|                   "movq   " MASK16_1 ", %%mm1 \n\t" // _mask16_1 -> mm1
 | ||
|                   RESTORE_rbp
 | ||
| 
 | ||
|                   "pand      %%mm7, %%mm0     \n\t"
 | ||
|                   "pand      %%mm7, %%mm1     \n\t"
 | ||
| 
 | ||
|                   "pcmpeqb   %%mm6, %%mm0     \n\t"
 | ||
|                   "pcmpeqb   %%mm6, %%mm1     \n\t"
 | ||
| 
 | ||
| // preload        "movl      len, %%ecx       \n\t" // load length of line
 | ||
| // preload        "movl      srcptr, %3       \n\t" // load source
 | ||
| // preload        "movl      dstptr, %4       \n\t" // load dest
 | ||
| 
 | ||
|                   "cmpl      $0, %%ecx        \n\t"
 | ||
|                   "jz        mainloop16end    \n\t"
 | ||
| 
 | ||
|                 "mainloop16:                  \n\t"
 | ||
|                   "movq      (%3), %%mm4      \n\t"
 | ||
|                   "pand      %%mm0, %%mm4     \n\t"
 | ||
|                   "movq      %%mm0, %%mm6     \n\t"
 | ||
|                   "movq      (%4), %%mm7      \n\t"
 | ||
|                   "pandn     %%mm7, %%mm6     \n\t"
 | ||
|                   "por       %%mm6, %%mm4     \n\t"
 | ||
|                   "movq      %%mm4, (%4)      \n\t"
 | ||
| 
 | ||
|                   "movq      8(%3), %%mm5     \n\t"
 | ||
|                   "pand      %%mm1, %%mm5     \n\t"
 | ||
|                   "movq      %%mm1, %%mm7     \n\t"
 | ||
|                   "movq      8(%4), %%mm6     \n\t"
 | ||
|                   "pandn     %%mm6, %%mm7     \n\t"
 | ||
|                   "por       %%mm7, %%mm5     \n\t"
 | ||
|                   "movq      %%mm5, 8(%4)     \n\t"
 | ||
| 
 | ||
|                   "add       $16, %3          \n\t" // inc by 16 bytes processed
 | ||
|                   "add       $16, %4          \n\t"
 | ||
|                   "subl      $8, %%ecx        \n\t" // dec by 8 pixels processed
 | ||
|                   "ja        mainloop16       \n\t"
 | ||
| 
 | ||
|                 "mainloop16end:               \n\t"
 | ||
| // preload        "movl      diff, %%ecx      \n\t" // (diff is in eax)
 | ||
|                   "movl      %%eax, %%ecx     \n\t"
 | ||
|                   "cmpl      $0, %%ecx        \n\t"
 | ||
|                   "jz        end16            \n\t"
 | ||
| // preload        "movl      mask, %%edx      \n\t"
 | ||
|                   "sall      $24, %%edx       \n\t" // make low byte, high byte
 | ||
| 
 | ||
|                 "secondloop16:                \n\t"
 | ||
|                   "sall      %%edx            \n\t" // move high bit to CF
 | ||
|                   "jnc       skip16           \n\t" // if CF = 0
 | ||
|                   "movw      (%3), %%ax       \n\t"
 | ||
|                   "movw      %%ax, (%4)       \n\t"
 | ||
| 
 | ||
|                 "skip16:                      \n\t"
 | ||
|                   "add       $2, %3           \n\t"
 | ||
|                   "add       $2, %4           \n\t"
 | ||
|                   "decl      %%ecx            \n\t"
 | ||
|                   "jnz       secondloop16     \n\t"
 | ||
| 
 | ||
|                 "end16:                       \n\t"
 | ||
|                   "EMMS                       \n\t" // DONE
 | ||
| 
 | ||
|                   : "=a" (dummy_value_a),           // output regs (dummy)
 | ||
|                     "=d" (dummy_value_d),
 | ||
|                     "=c" (dummy_value_c),
 | ||
|                     "=S" (dummy_value_S),
 | ||
|                     "=D" (dummy_value_D)
 | ||
| 
 | ||
|                   : "0" (diff),        // eax       // input regs
 | ||
|                     "1" (mask),        // edx
 | ||
|                     "2" (len),         // ecx
 | ||
| // was (unmask)     "b"    RESERVED    // ebx       // Global Offset Table idx
 | ||
|                     "3" (srcptr),      // esi/rsi
 | ||
|                     "4" (dstptr)       // edi/rdi
 | ||
| 
 | ||
| #if defined(PNG_CLOBBER_MMX_REGS_SUPPORTED)
 | ||
|                   : "%mm0", "%mm1", "%mm4"          // clobber list
 | ||
|                   , "%mm5", "%mm6", "%mm7"
 | ||
| #endif
 | ||
|                );
 | ||
|             }
 | ||
|             else /* not _mmx_supported - use modified C routine */
 | ||
|             {
 | ||
|                register png_uint_32 i;
 | ||
|                png_uint_32 initial_val = BPP2 * png_pass_start[png_ptr->pass];
 | ||
|                  /* png.c:  png_pass_start[] = {0, 4, 0, 2, 0, 1, 0}; */
 | ||
|                register int stride = BPP2 * png_pass_inc[png_ptr->pass];
 | ||
|                  /* png.c:  png_pass_inc[] = {8, 8, 4, 4, 2, 2, 1}; */
 | ||
|                register int rep_bytes = BPP2 * png_pass_width[png_ptr->pass];
 | ||
|                  /* png.c:  png_pass_width[] = {8, 4, 4, 2, 2, 1, 1}; */
 | ||
|                png_uint_32 len = png_ptr->width &~7;  /* reduce to mult. of 8 */
 | ||
|                int diff = (int) (png_ptr->width & 7); /* amount lost */
 | ||
|                register png_uint_32 final_val = BPP2 * len;   /* GRR bugfix */
 | ||
| 
 | ||
|                srcptr = png_ptr->row_buf + 1 + initial_val;
 | ||
|                dstptr = row + initial_val;
 | ||
| 
 | ||
|                for (i = initial_val; i < final_val; i += stride)
 | ||
|                {
 | ||
|                   png_memcpy(dstptr, srcptr, rep_bytes);
 | ||
|                   srcptr += stride;
 | ||
|                   dstptr += stride;
 | ||
|                }
 | ||
|                if (diff)  /* number of leftover pixels:  3 for pngtest */
 | ||
|                {
 | ||
|                   final_val += diff*BPP2;
 | ||
|                   for (; i < final_val; i += stride)
 | ||
|                   {
 | ||
|                      if (rep_bytes > (int)(final_val-i))
 | ||
|                         rep_bytes = (int)(final_val-i);
 | ||
|                      png_memcpy(dstptr, srcptr, rep_bytes);
 | ||
|                      srcptr += stride;
 | ||
|                      dstptr += stride;
 | ||
|                   }
 | ||
|                }
 | ||
|             } /* end of else (_mmx_supported) */
 | ||
| 
 | ||
|             break;
 | ||
|          }       /* end 16 bpp */
 | ||
| 
 | ||
|          case 48:       /* png_ptr->row_info.pixel_depth */
 | ||
|          {
 | ||
|             png_bytep srcptr;
 | ||
|             png_bytep dstptr;
 | ||
| 
 | ||
| #if !defined(PNG_1_0_X)
 | ||
|             if (png_ptr->asm_flags & PNG_ASM_FLAG_MMX_READ_COMBINE_ROW)
 | ||
| #else
 | ||
|             if (_mmx_supported)
 | ||
| #endif
 | ||
|             {
 | ||
|                png_uint_32 len;
 | ||
|                int diff;
 | ||
| 
 | ||
|                srcptr = png_ptr->row_buf + 1;
 | ||
|                dstptr = row;
 | ||
|                len  = png_ptr->width & ~7;          // reduce to multiple of 8
 | ||
|                diff = (int) (png_ptr->width & 7);   // amount lost
 | ||
| 
 | ||
|                __asm__ __volatile__ (
 | ||
|                   "not       %%edx            \n\t" // mask => unmask
 | ||
|                   "movd      %%edx, %%mm7     \n\t" // load bit pattern
 | ||
|                   "not       %%edx            \n\t" // unmask => mask for later
 | ||
|                   "psubb     %%mm6, %%mm6     \n\t" // zero mm6
 | ||
|                   "punpcklbw %%mm7, %%mm7     \n\t"
 | ||
|                   "punpcklwd %%mm7, %%mm7     \n\t"
 | ||
|                   "punpckldq %%mm7, %%mm7     \n\t" // fill reg with 8 masks
 | ||
| 
 | ||
|                   LOAD_GOT_rbp
 | ||
|                   "movq   " MASK48_0 ", %%mm0 \n\t" // _mask48_0 -> mm0
 | ||
|                   "movq   " MASK48_1 ", %%mm1 \n\t" // _mask48_1 -> mm1
 | ||
|                   "movq   " MASK48_2 ", %%mm2 \n\t" // _mask48_2 -> mm2
 | ||
|                   "movq   " MASK48_3 ", %%mm3 \n\t" // _mask48_3 -> mm3
 | ||
|                   "movq   " MASK48_4 ", %%mm4 \n\t" // _mask48_4 -> mm4
 | ||
|                   "movq   " MASK48_5 ", %%mm5 \n\t" // _mask48_5 -> mm5
 | ||
|                   RESTORE_rbp
 | ||
| 
 | ||
|                   "pand      %%mm7, %%mm0     \n\t"
 | ||
|                   "pand      %%mm7, %%mm1     \n\t"
 | ||
|                   "pand      %%mm7, %%mm2     \n\t"
 | ||
|                   "pand      %%mm7, %%mm3     \n\t"
 | ||
|                   "pand      %%mm7, %%mm4     \n\t"
 | ||
|                   "pand      %%mm7, %%mm5     \n\t"
 | ||
| 
 | ||
|                   "pcmpeqb   %%mm6, %%mm0     \n\t"
 | ||
|                   "pcmpeqb   %%mm6, %%mm1     \n\t"
 | ||
|                   "pcmpeqb   %%mm6, %%mm2     \n\t"
 | ||
|                   "pcmpeqb   %%mm6, %%mm3     \n\t"
 | ||
|                   "pcmpeqb   %%mm6, %%mm4     \n\t"
 | ||
|                   "pcmpeqb   %%mm6, %%mm5     \n\t"
 | ||
| 
 | ||
| // preload        "movl      len, %%ecx       \n\t" // load length of line
 | ||
| // preload        "movl      srcptr, %3       \n\t" // load source
 | ||
| // preload        "movl      dstptr, %4       \n\t" // load dest
 | ||
| 
 | ||
|                   "cmpl      $0, %%ecx        \n\t"
 | ||
|                   "jz        mainloop48end    \n\t"
 | ||
| 
 | ||
|                 "mainloop48:                  \n\t"
 | ||
|                   "movq      (%3), %%mm7      \n\t"
 | ||
|                   "pand      %%mm0, %%mm7     \n\t"
 | ||
|                   "movq      %%mm0, %%mm6     \n\t"
 | ||
|                   "pandn     (%4), %%mm6      \n\t"
 | ||
|                   "por       %%mm6, %%mm7     \n\t"
 | ||
|                   "movq      %%mm7, (%4)      \n\t"
 | ||
| 
 | ||
|                   "movq      8(%3), %%mm6     \n\t"
 | ||
|                   "pand      %%mm1, %%mm6     \n\t"
 | ||
|                   "movq      %%mm1, %%mm7     \n\t"
 | ||
|                   "pandn     8(%4), %%mm7     \n\t"
 | ||
|                   "por       %%mm7, %%mm6     \n\t"
 | ||
|                   "movq      %%mm6, 8(%4)     \n\t"
 | ||
| 
 | ||
|                   "movq      16(%3), %%mm6    \n\t"
 | ||
|                   "pand      %%mm2, %%mm6     \n\t"
 | ||
|                   "movq      %%mm2, %%mm7     \n\t"
 | ||
|                   "pandn     16(%4), %%mm7    \n\t"
 | ||
|                   "por       %%mm7, %%mm6     \n\t"
 | ||
|                   "movq      %%mm6, 16(%4)    \n\t"
 | ||
| 
 | ||
|                   "movq      24(%3), %%mm7    \n\t"
 | ||
|                   "pand      %%mm3, %%mm7     \n\t"
 | ||
|                   "movq      %%mm3, %%mm6     \n\t"
 | ||
|                   "pandn     24(%4), %%mm6    \n\t"
 | ||
|                   "por       %%mm6, %%mm7     \n\t"
 | ||
|                   "movq      %%mm7, 24(%4)    \n\t"
 | ||
| 
 | ||
|                   "movq      32(%3), %%mm6    \n\t"
 | ||
|                   "pand      %%mm4, %%mm6     \n\t"
 | ||
|                   "movq      %%mm4, %%mm7     \n\t"
 | ||
|                   "pandn     32(%4), %%mm7    \n\t"
 | ||
|                   "por       %%mm7, %%mm6     \n\t"
 | ||
|                   "movq      %%mm6, 32(%4)    \n\t"
 | ||
| 
 | ||
|                   "movq      40(%3), %%mm7    \n\t"
 | ||
|                   "pand      %%mm5, %%mm7     \n\t"
 | ||
|                   "movq      %%mm5, %%mm6     \n\t"
 | ||
|                   "pandn     40(%4), %%mm6    \n\t"
 | ||
|                   "por       %%mm6, %%mm7     \n\t"
 | ||
|                   "movq      %%mm7, 40(%4)    \n\t"
 | ||
| 
 | ||
|                   "add       $48, %3          \n\t" // inc by 48 bytes processed
 | ||
|                   "add       $48, %4          \n\t"
 | ||
|                   "subl      $8, %%ecx        \n\t" // dec by 8 pixels processed
 | ||
| 
 | ||
|                   "ja        mainloop48       \n\t"
 | ||
| 
 | ||
|                 "mainloop48end:               \n\t"
 | ||
| // preload        "movl      diff, %%ecx      \n\t" // (diff is in eax)
 | ||
|                   "movl      %%eax, %%ecx     \n\t"
 | ||
|                   "cmpl      $0, %%ecx        \n\t"
 | ||
|                   "jz        end48            \n\t"
 | ||
| // preload        "movl      mask, %%edx      \n\t"
 | ||
|                   "sall      $24, %%edx       \n\t" // make low byte, high byte
 | ||
| 
 | ||
|                 "secondloop48:                \n\t"
 | ||
|                   "sall      %%edx            \n\t" // move high bit to CF
 | ||
|                   "jnc       skip48           \n\t" // if CF = 0
 | ||
|                   "movl      (%3), %%eax      \n\t"
 | ||
|                   "movl      %%eax, (%4)      \n\t"
 | ||
|                   "movw      4(%3), %%ax      \n\t" // GR-P bugfix 20070717
 | ||
|                   "movw      %%ax, 4(%4)      \n\t" // GR-P bugfix 20070717
 | ||
| 
 | ||
|                 "skip48:                      \n\t"
 | ||
|                   "add       $6, %3           \n\t" // GR-P bugfix 20070717
 | ||
|                   "add       $6, %4           \n\t" // GR-P bugfix 20070717
 | ||
|                   "decl      %%ecx            \n\t"
 | ||
|                   "jnz       secondloop48     \n\t"
 | ||
| 
 | ||
|                 "end48:                       \n\t"
 | ||
|                   "EMMS                       \n\t" // DONE
 | ||
| 
 | ||
|                   : "=a" (dummy_value_a),           // output regs (dummy)
 | ||
|                     "=d" (dummy_value_d),
 | ||
|                     "=c" (dummy_value_c),
 | ||
|                     "=S" (dummy_value_S),
 | ||
|                     "=D" (dummy_value_D)
 | ||
| 
 | ||
|                   : "0" (diff),        // eax       // input regs
 | ||
|                     "1" (mask),        // edx
 | ||
|                     "2" (len),         // ecx
 | ||
| // was (unmask)     "b"    RESERVED    // ebx       // Global Offset Table idx
 | ||
|                     "3" (srcptr),      // esi/rsi
 | ||
|                     "4" (dstptr)       // edi/rdi
 | ||
| 
 | ||
| #if defined(PNG_CLOBBER_MMX_REGS_SUPPORTED)
 | ||
|                   : "%mm0", "%mm1", "%mm2", "%mm3"  // clobber list
 | ||
|                   , "%mm4", "%mm5", "%mm6", "%mm7"
 | ||
| #endif
 | ||
|                );
 | ||
|             }
 | ||
|             else /* not _mmx_supported - use modified C routine */
 | ||
|             {
 | ||
|                register png_uint_32 i;
 | ||
|                png_uint_32 initial_val = BPP6 * png_pass_start[png_ptr->pass];
 | ||
|                  /* png.c:  png_pass_start[] = {0, 4, 0, 2, 0, 1, 0}; */
 | ||
|                register int stride = BPP6 * png_pass_inc[png_ptr->pass];
 | ||
|                  /* png.c:  png_pass_inc[] = {8, 8, 4, 4, 2, 2, 1}; */
 | ||
|                register int rep_bytes = BPP6 * png_pass_width[png_ptr->pass];
 | ||
|                  /* png.c:  png_pass_width[] = {8, 4, 4, 2, 2, 1, 1}; */
 | ||
|                png_uint_32 len = png_ptr->width &~7;  /* reduce to mult. of 8 */
 | ||
|                int diff = (int) (png_ptr->width & 7); /* amount lost */
 | ||
|                register png_uint_32 final_val = BPP6 * len;   /* GRR bugfix */
 | ||
| 
 | ||
|                srcptr = png_ptr->row_buf + 1 + initial_val;
 | ||
|                dstptr = row + initial_val;
 | ||
| 
 | ||
|                for (i = initial_val; i < final_val; i += stride)
 | ||
|                {
 | ||
|                   png_memcpy(dstptr, srcptr, rep_bytes);
 | ||
|                   srcptr += stride;
 | ||
|                   dstptr += stride;
 | ||
|                }
 | ||
|                if (diff)  /* number of leftover pixels:  3 for pngtest */
 | ||
|                {
 | ||
|                   final_val += diff*BPP6;
 | ||
|                   for (; i < final_val; i += stride)
 | ||
|                   {
 | ||
|                      if (rep_bytes > (int)(final_val-i))
 | ||
|                         rep_bytes = (int)(final_val-i);
 | ||
|                      png_memcpy(dstptr, srcptr, rep_bytes);
 | ||
|                      srcptr += stride;
 | ||
|                      dstptr += stride;
 | ||
|                   }
 | ||
|                }
 | ||
|             } /* end of else (_mmx_supported) */
 | ||
| 
 | ||
|             break;
 | ||
|          }       /* end 48 bpp */
 | ||
| 
 | ||
|          case 64:       /* png_ptr->row_info.pixel_depth */
 | ||
|          {
 | ||
|             png_bytep srcptr;
 | ||
|             png_bytep dstptr;
 | ||
|             register png_uint_32 i;
 | ||
|             png_uint_32 initial_val = BPP8 * png_pass_start[png_ptr->pass];
 | ||
|               /* png.c:  png_pass_start[] = {0, 4, 0, 2, 0, 1, 0}; */
 | ||
|             register int stride = BPP8 * png_pass_inc[png_ptr->pass];
 | ||
|               /* png.c:  png_pass_inc[] = {8, 8, 4, 4, 2, 2, 1}; */
 | ||
|             register int rep_bytes = BPP8 * png_pass_width[png_ptr->pass];
 | ||
|               /* png.c:  png_pass_width[] = {8, 4, 4, 2, 2, 1, 1}; */
 | ||
|             png_uint_32 len = png_ptr->width &~7;  /* reduce to mult. of 8 */
 | ||
|             int diff = (int) (png_ptr->width & 7); /* amount lost */
 | ||
|             register png_uint_32 final_val = BPP8 * len;   /* GRR bugfix */
 | ||
| 
 | ||
|             srcptr = png_ptr->row_buf + 1 + initial_val;
 | ||
|             dstptr = row + initial_val;
 | ||
| 
 | ||
|             for (i = initial_val; i < final_val; i += stride)
 | ||
|             {
 | ||
|                png_memcpy(dstptr, srcptr, rep_bytes);
 | ||
|                srcptr += stride;
 | ||
|                dstptr += stride;
 | ||
|             }
 | ||
|             if (diff)  /* number of leftover pixels:  3 for pngtest */
 | ||
|             {
 | ||
|                final_val += diff*BPP8;
 | ||
|                for (; i < final_val; i += stride)
 | ||
|                {
 | ||
|                   if (rep_bytes > (int)(final_val-i))
 | ||
|                      rep_bytes = (int)(final_val-i);
 | ||
|                   png_memcpy(dstptr, srcptr, rep_bytes);
 | ||
|                   srcptr += stride;
 | ||
|                   dstptr += stride;
 | ||
|                }
 | ||
|             }
 | ||
| 
 | ||
|             break;
 | ||
|          }       /* end 64 bpp */
 | ||
| 
 | ||
|          default: /* png_ptr->row_info.pixel_depth != 1,2,4,8,16,24,32,48,64 */
 | ||
|          {
 | ||
|             // ERROR:  SHOULD NEVER BE REACHED
 | ||
| #if defined(PNG_DEBUG)
 | ||
|             png_debug(1, "Internal libpng logic error (GCC "
 | ||
|               "png_combine_row() pixel_depth)\n");
 | ||
| #endif
 | ||
|             break;
 | ||
|          }
 | ||
|       } /* end switch (png_ptr->row_info.pixel_depth) */
 | ||
| 
 | ||
|    } /* end if (non-trivial mask) */
 | ||
| 
 | ||
| } /* end png_combine_row() */
 | ||
| 
 | ||
| #endif /* PNG_HAVE_MMX_COMBINE_ROW */
 | ||
| 
 | ||
| 
 | ||
| 
 | ||
| 
 | ||
| /*===========================================================================*/
 | ||
| /*                                                                           */
 | ||
| /*                 P N G _ D O _ R E A D _ I N T E R L A C E                 */
 | ||
| /*                                                                           */
 | ||
| /*===========================================================================*/
 | ||
| 
 | ||
| #if defined(PNG_READ_INTERLACING_SUPPORTED)
 | ||
| #if defined(PNG_HAVE_MMX_READ_INTERLACE)
 | ||
| 
 | ||
| /* png_do_read_interlace() is called after any 16-bit to 8-bit conversion
 | ||
|  * has taken place.  [GRR: what other steps come before and/or after?]
 | ||
|  */
 | ||
| 
 | ||
| void /* PRIVATE */
 | ||
| png_do_read_interlace(png_structp png_ptr)
 | ||
| {
 | ||
|    png_row_infop row_info = &(png_ptr->row_info);
 | ||
|    png_bytep row = png_ptr->row_buf + 1;
 | ||
|    int pass = png_ptr->pass;
 | ||
| #if defined(PNG_READ_PACKSWAP_SUPPORTED)
 | ||
|    png_uint_32 transformations = png_ptr->transformations;
 | ||
| #endif
 | ||
| 
 | ||
|    png_debug(1, "in png_do_read_interlace (pnggccrd.c)\n");
 | ||
| 
 | ||
|    if (_mmx_supported == 2) {
 | ||
| #if !defined(PNG_1_0_X)
 | ||
|        /* this should have happened in png_init_mmx_flags() already */
 | ||
|        png_warning(png_ptr, "asm_flags may not have been initialized");
 | ||
| #endif
 | ||
|        png_mmx_support();
 | ||
|    }
 | ||
| 
 | ||
|    if (row != NULL && row_info != NULL)
 | ||
|    {
 | ||
|       png_uint_32 final_width;
 | ||
| 
 | ||
|       final_width = row_info->width * png_pass_inc[pass];
 | ||
| 
 | ||
|       switch (row_info->pixel_depth)
 | ||
|       {
 | ||
|          case 1:
 | ||
|          {
 | ||
|             png_bytep sp, dp;
 | ||
|             int sshift, dshift;
 | ||
|             int s_start, s_end, s_inc;
 | ||
|             png_byte v;
 | ||
|             png_uint_32 i;
 | ||
|             int j;
 | ||
| 
 | ||
|             sp = row + (png_size_t)((row_info->width - 1) >> 3);
 | ||
|             dp = row + (png_size_t)((final_width - 1) >> 3);
 | ||
| #if defined(PNG_READ_PACKSWAP_SUPPORTED)
 | ||
|             if (transformations & PNG_PACKSWAP)
 | ||
|             {
 | ||
|                sshift = (int)((row_info->width + 7) & 7);
 | ||
|                dshift = (int)((final_width + 7) & 7);
 | ||
|                s_start = 7;
 | ||
|                s_end = 0;
 | ||
|                s_inc = -1;
 | ||
|             }
 | ||
|             else
 | ||
| #endif
 | ||
|             {
 | ||
|                sshift = 7 - (int)((row_info->width + 7) & 7);
 | ||
|                dshift = 7 - (int)((final_width + 7) & 7);
 | ||
|                s_start = 0;
 | ||
|                s_end = 7;
 | ||
|                s_inc = 1;
 | ||
|             }
 | ||
| 
 | ||
|             for (i = row_info->width; i; i--)
 | ||
|             {
 | ||
|                v = (png_byte)((*sp >> sshift) & 0x1);
 | ||
|                for (j = 0; j < png_pass_inc[pass]; j++)
 | ||
|                {
 | ||
|                   *dp &= (png_byte)((0x7f7f >> (7 - dshift)) & 0xff);
 | ||
|                   *dp |= (png_byte)(v << dshift);
 | ||
|                   if (dshift == s_end)
 | ||
|                   {
 | ||
|                      dshift = s_start;
 | ||
|                      dp--;
 | ||
|                   }
 | ||
|                   else
 | ||
|                      dshift += s_inc;
 | ||
|                }
 | ||
|                if (sshift == s_end)
 | ||
|                {
 | ||
|                   sshift = s_start;
 | ||
|                   sp--;
 | ||
|                }
 | ||
|                else
 | ||
|                   sshift += s_inc;
 | ||
|             }
 | ||
|             break;
 | ||
|          }
 | ||
| 
 | ||
|          case 2:
 | ||
|          {
 | ||
|             png_bytep sp, dp;
 | ||
|             int sshift, dshift;
 | ||
|             int s_start, s_end, s_inc;
 | ||
|             png_uint_32 i;
 | ||
| 
 | ||
|             sp = row + (png_size_t)((row_info->width - 1) >> 2);
 | ||
|             dp = row + (png_size_t)((final_width - 1) >> 2);
 | ||
| #if defined(PNG_READ_PACKSWAP_SUPPORTED)
 | ||
|             if (transformations & PNG_PACKSWAP)
 | ||
|             {
 | ||
|                sshift = (png_size_t)(((row_info->width + 3) & 3) << 1);
 | ||
|                dshift = (png_size_t)(((final_width + 3) & 3) << 1);
 | ||
|                s_start = 6;
 | ||
|                s_end = 0;
 | ||
|                s_inc = -2;
 | ||
|             }
 | ||
|             else
 | ||
| #endif
 | ||
|             {
 | ||
|                sshift = (png_size_t)((3 - ((row_info->width + 3) & 3)) << 1);
 | ||
|                dshift = (png_size_t)((3 - ((final_width + 3) & 3)) << 1);
 | ||
|                s_start = 0;
 | ||
|                s_end = 6;
 | ||
|                s_inc = 2;
 | ||
|             }
 | ||
| 
 | ||
|             for (i = row_info->width; i; i--)
 | ||
|             {
 | ||
|                png_byte v;
 | ||
|                int j;
 | ||
| 
 | ||
|                v = (png_byte)((*sp >> sshift) & 0x3);
 | ||
|                for (j = 0; j < png_pass_inc[pass]; j++)
 | ||
|                {
 | ||
|                   *dp &= (png_byte)((0x3f3f >> (6 - dshift)) & 0xff);
 | ||
|                   *dp |= (png_byte)(v << dshift);
 | ||
|                   if (dshift == s_end)
 | ||
|                   {
 | ||
|                      dshift = s_start;
 | ||
|                      dp--;
 | ||
|                   }
 | ||
|                   else
 | ||
|                      dshift += s_inc;
 | ||
|                }
 | ||
|                if (sshift == s_end)
 | ||
|                {
 | ||
|                   sshift = s_start;
 | ||
|                   sp--;
 | ||
|                }
 | ||
|                else
 | ||
|                   sshift += s_inc;
 | ||
|             }
 | ||
|             break;
 | ||
|          }
 | ||
| 
 | ||
|          case 4:
 | ||
|          {
 | ||
|             png_bytep sp, dp;
 | ||
|             int sshift, dshift;
 | ||
|             int s_start, s_end, s_inc;
 | ||
|             png_uint_32 i;
 | ||
| 
 | ||
|             sp = row + (png_size_t)((row_info->width - 1) >> 1);
 | ||
|             dp = row + (png_size_t)((final_width - 1) >> 1);
 | ||
| #if defined(PNG_READ_PACKSWAP_SUPPORTED)
 | ||
|             if (transformations & PNG_PACKSWAP)
 | ||
|             {
 | ||
|                sshift = (png_size_t)(((row_info->width + 1) & 1) << 2);
 | ||
|                dshift = (png_size_t)(((final_width + 1) & 1) << 2);
 | ||
|                s_start = 4;
 | ||
|                s_end = 0;
 | ||
|                s_inc = -4;
 | ||
|             }
 | ||
|             else
 | ||
| #endif
 | ||
|             {
 | ||
|                sshift = (png_size_t)((1 - ((row_info->width + 1) & 1)) << 2);
 | ||
|                dshift = (png_size_t)((1 - ((final_width + 1) & 1)) << 2);
 | ||
|                s_start = 0;
 | ||
|                s_end = 4;
 | ||
|                s_inc = 4;
 | ||
|             }
 | ||
| 
 | ||
|             for (i = row_info->width; i; i--)
 | ||
|             {
 | ||
|                png_byte v;
 | ||
|                int j;
 | ||
| 
 | ||
|                v = (png_byte)((*sp >> sshift) & 0xf);
 | ||
|                for (j = 0; j < png_pass_inc[pass]; j++)
 | ||
|                {
 | ||
|                   *dp &= (png_byte)((0xf0f >> (4 - dshift)) & 0xff);
 | ||
|                   *dp |= (png_byte)(v << dshift);
 | ||
|                   if (dshift == s_end)
 | ||
|                   {
 | ||
|                      dshift = s_start;
 | ||
|                      dp--;
 | ||
|                   }
 | ||
|                   else
 | ||
|                      dshift += s_inc;
 | ||
|                }
 | ||
|                if (sshift == s_end)
 | ||
|                {
 | ||
|                   sshift = s_start;
 | ||
|                   sp--;
 | ||
|                }
 | ||
|                else
 | ||
|                   sshift += s_inc;
 | ||
|             }
 | ||
|             break;
 | ||
|          }
 | ||
| 
 | ||
|        /*====================================================================*/
 | ||
| 
 | ||
|          default: /* 8-bit or larger (this is where the routine is modified) */
 | ||
|          {
 | ||
|             png_bytep sptr, dp;
 | ||
|             png_uint_32 i;
 | ||
|             png_size_t pixel_bytes;
 | ||
|             int width = (int)row_info->width;
 | ||
| 
 | ||
|             pixel_bytes = (row_info->pixel_depth >> 3);
 | ||
| 
 | ||
|             /* point sptr at the last pixel in the pre-expanded row: */
 | ||
|             sptr = row + (width - 1) * pixel_bytes;
 | ||
| 
 | ||
|             /* point dp at the last pixel position in the expanded row: */
 | ||
|             dp = row + (final_width - 1) * pixel_bytes;
 | ||
| 
 | ||
|             /* New code by Nirav Chhatrapati - Intel Corporation */
 | ||
| 
 | ||
| #if !defined(PNG_1_0_X)
 | ||
|             if (png_ptr->asm_flags & PNG_ASM_FLAG_MMX_READ_INTERLACE)
 | ||
| #else
 | ||
|             if (_mmx_supported)
 | ||
| #endif
 | ||
|             {
 | ||
|                int dummy_value_c;        // fix 'forbidden register spilled'
 | ||
|                png_bytep dummy_value_S;
 | ||
|                png_bytep dummy_value_D;
 | ||
|                png_bytep dummy_value_a;
 | ||
|                png_bytep dummy_value_d;
 | ||
| 
 | ||
|                //--------------------------------------------------------------
 | ||
|                if (pixel_bytes == BPP3)
 | ||
|                {
 | ||
|                   if (((pass == 4) || (pass == 5)) && width)
 | ||
|                   {
 | ||
|                      int width_mmx = ((width >> 1) << 1) - 8;   // GRR:  huh?
 | ||
|                      if (width_mmx < 0)
 | ||
|                          width_mmx = 0;
 | ||
|                      width -= width_mmx;        // 8 or 9 pix, 24 or 27 bytes
 | ||
|                      if (width_mmx)
 | ||
|                      {
 | ||
|                         // png_pass_inc[] = {8, 8, 4, 4, 2, 2, 1};
 | ||
|                         // sptr points at last pixel in pre-expanded row
 | ||
|                         // dp points at last pixel position in expanded row
 | ||
|                         __asm__ __volatile__ (
 | ||
|                            "sub  $3, %1             \n\t"
 | ||
|                            "sub  $9, %2             \n\t"
 | ||
|                                         // (png_pass_inc[pass] + 1)*pixel_bytes
 | ||
| 
 | ||
|                         ".loop3_pass4:              \n\t"
 | ||
|                            "movq (%1), %%mm0        \n\t" // x x 5 4 3 2 1 0
 | ||
|                            "movq %%mm0, %%mm1       \n\t" // x x 5 4 3 2 1 0
 | ||
|                            "movq %%mm0, %%mm2       \n\t" // x x 5 4 3 2 1 0
 | ||
|                            "psllq $24, %%mm0        \n\t" // 4 3 2 1 0 z z z
 | ||
|                            "pand (%3), %%mm1        \n\t" // z z z z z 2 1 0
 | ||
|                            "psrlq $24, %%mm2        \n\t" // z z z x x 5 4 3
 | ||
|                            "por %%mm1, %%mm0        \n\t" // 4 3 2 1 0 2 1 0
 | ||
|                            "movq %%mm2, %%mm3       \n\t" // z z z x x 5 4 3
 | ||
|                            "psllq $8, %%mm2         \n\t" // z z x x 5 4 3 z
 | ||
|                            "movq %%mm0, (%2)        \n\t"
 | ||
|                            "psrlq $16, %%mm3        \n\t" // z z z z z x x 5
 | ||
|                            "pand (%4), %%mm3        \n\t" // z z z z z z z 5
 | ||
|                            "por %%mm3, %%mm2        \n\t" // z z x x 5 4 3 5
 | ||
|                            "sub  $6, %1             \n\t"
 | ||
|                            "movd %%mm2, 8(%2)       \n\t"
 | ||
|                            "sub  $12, %2            \n\t"
 | ||
|                            "subl $2, %%ecx          \n\t"
 | ||
|                            "jnz .loop3_pass4        \n\t"
 | ||
|                            "EMMS                    \n\t" // DONE
 | ||
| 
 | ||
|                            : "=c" (dummy_value_c),        // output regs (dummy)
 | ||
|                              "=S" (dummy_value_S),
 | ||
|                              "=D" (dummy_value_D),
 | ||
|                              "=a" (dummy_value_a),
 | ||
|                              "=d" (dummy_value_d)
 | ||
| 
 | ||
|                            : "0" (width_mmx),     // ecx  // input regs
 | ||
|                              "1" (sptr),          // esi/rsi
 | ||
|                              "2" (dp),            // edi/rdi
 | ||
| #if defined(PNG_x86_64_USE_GOTPCREL)     // formerly _const4 and _const6:
 | ||
|                              "3" (&_c64._amask5_3_0), // (0x0000000000FFFFFFLL)
 | ||
|                              "4" (&_c64._amask7_1_0)  // (0x00000000000000FFLL)
 | ||
| #else
 | ||
|                              "3" (&_amask5_3_0),  // eax (0x0000000000FFFFFFLL)
 | ||
|                              "4" (&_amask7_1_0)   // edx (0x00000000000000FFLL)
 | ||
| #endif
 | ||
| 
 | ||
| #if defined(PNG_CLOBBER_MMX_REGS_SUPPORTED)
 | ||
|                            : "%mm0", "%mm1"               // clobber list
 | ||
|                            , "%mm2", "%mm3"
 | ||
| #endif
 | ||
|                         );
 | ||
|                      }
 | ||
| 
 | ||
|                      sptr -= width_mmx*BPP3;
 | ||
|                      dp -= width_mmx*2*BPP3;
 | ||
|                      for (i = width; i; i--)
 | ||
|                      {
 | ||
|                         png_byte v[8];
 | ||
|                         int j;
 | ||
| 
 | ||
|                         png_memcpy(v, sptr, BPP3);
 | ||
|                         for (j = 0; j < png_pass_inc[pass]; j++)
 | ||
|                         {
 | ||
|                            png_memcpy(dp, v, BPP3);
 | ||
|                            dp -= BPP3;
 | ||
|                         }
 | ||
|                         sptr -= BPP3;
 | ||
|                      }
 | ||
|                   }
 | ||
|                   else if (((pass == 2) || (pass == 3)) && width)
 | ||
|                   {
 | ||
|                      __asm__ __volatile__ (
 | ||
|                         "sub  $9, %2             \n\t"
 | ||
|                                      // (png_pass_inc[pass] - 1)*pixel_bytes
 | ||
| 
 | ||
|                      ".loop3_pass2:              \n\t"
 | ||
|                         "movd (%1), %%mm0        \n\t" // x x x x x 2 1 0
 | ||
|                         "pand (%3), %%mm0        \n\t" // z z z z z 2 1 0
 | ||
|                         "movq %%mm0, %%mm1       \n\t" // z z z z z 2 1 0
 | ||
|                         "psllq $16, %%mm0        \n\t" // z z z 2 1 0 z z
 | ||
|                         "movq %%mm0, %%mm2       \n\t" // z z z 2 1 0 z z
 | ||
|                         "psllq $24, %%mm0        \n\t" // 2 1 0 z z z z z
 | ||
|                         "psrlq $8, %%mm1         \n\t" // z z z z z z 2 1
 | ||
|                         "por %%mm2, %%mm0        \n\t" // 2 1 0 2 1 0 z z
 | ||
|                         "por %%mm1, %%mm0        \n\t" // 2 1 0 2 1 0 2 1
 | ||
|                         "movq %%mm0, 4(%2)       \n\t"
 | ||
|                         "psrlq $16, %%mm0        \n\t" // z z 2 1 0 2 1 0
 | ||
|                         "sub  $3, %1             \n\t"
 | ||
|                         "movd %%mm0, (%2)        \n\t"
 | ||
|                         "sub  $12, %2            \n\t"
 | ||
|                         "decl %%ecx              \n\t"
 | ||
|                         "jnz .loop3_pass2        \n\t"
 | ||
|                         "EMMS                    \n\t" // DONE
 | ||
| 
 | ||
|                         : "=c" (dummy_value_c),        // output regs (dummy)
 | ||
|                           "=S" (dummy_value_S),
 | ||
|                           "=D" (dummy_value_D),
 | ||
|                           "=a" (dummy_value_a)
 | ||
| 
 | ||
|                         : "0" (width),         // ecx  // input regs
 | ||
|                           "1" (sptr),          // esi/rsi
 | ||
|                           "2" (dp),            // edi/rdi
 | ||
| #if defined(PNG_x86_64_USE_GOTPCREL)           // formerly _const4:
 | ||
|                           "3" (&_c64._amask5_3_0)  // (0x0000000000FFFFFFLL)
 | ||
| #else
 | ||
|                           "3" (&_amask5_3_0)   // eax (0x0000000000FFFFFFLL)
 | ||
| #endif
 | ||
| 
 | ||
| #if defined(CLOBBER_MMX_REGS_SUPPORTED)
 | ||
|                         : "%mm0", "%mm1", "%mm2"       // clobber list
 | ||
| #endif
 | ||
|                      );
 | ||
|                   }
 | ||
|                   else if (width)  // && ((pass == 0) || (pass == 1))
 | ||
|                   {
 | ||
|                      __asm__ __volatile__ (
 | ||
|                         "sub  $21, %2            \n\t"
 | ||
|                                      // (png_pass_inc[pass] - 1)*pixel_bytes
 | ||
| 
 | ||
|                      ".loop3_pass0:              \n\t"
 | ||
|                         "movd (%1), %%mm0        \n\t" // x x x x x 2 1 0
 | ||
|                         "pand (%3), %%mm0        \n\t" // z z z z z 2 1 0
 | ||
|                         "movq %%mm0, %%mm1       \n\t" // z z z z z 2 1 0
 | ||
|                         "psllq $16, %%mm0        \n\t" // z z z 2 1 0 z z
 | ||
|                         "movq %%mm0, %%mm2       \n\t" // z z z 2 1 0 z z
 | ||
|                         "psllq $24, %%mm0        \n\t" // 2 1 0 z z z z z
 | ||
|                         "psrlq $8, %%mm1         \n\t" // z z z z z z 2 1
 | ||
|                         "por %%mm2, %%mm0        \n\t" // 2 1 0 2 1 0 z z
 | ||
|                         "por %%mm1, %%mm0        \n\t" // 2 1 0 2 1 0 2 1
 | ||
|                         "movq %%mm0, %%mm3       \n\t" // 2 1 0 2 1 0 2 1
 | ||
|                         "psllq $16, %%mm0        \n\t" // 0 2 1 0 2 1 z z
 | ||
|                         "movq %%mm3, %%mm4       \n\t" // 2 1 0 2 1 0 2 1
 | ||
|                         "punpckhdq %%mm0, %%mm3  \n\t" // 0 2 1 0 2 1 0 2
 | ||
|                         "movq %%mm4, 16(%2)      \n\t"
 | ||
|                         "psrlq $32, %%mm0        \n\t" // z z z z 0 2 1 0
 | ||
|                         "movq %%mm3, 8(%2)       \n\t"
 | ||
|                         "punpckldq %%mm4, %%mm0  \n\t" // 1 0 2 1 0 2 1 0
 | ||
|                         "sub  $3, %1             \n\t"
 | ||
|                         "movq %%mm0, (%2)        \n\t"
 | ||
|                         "sub  $24, %2            \n\t"
 | ||
|                         "decl %%ecx              \n\t"
 | ||
|                         "jnz .loop3_pass0        \n\t"
 | ||
|                         "EMMS                    \n\t" // DONE
 | ||
| 
 | ||
|                         : "=c" (dummy_value_c),        // output regs (dummy)
 | ||
|                           "=S" (dummy_value_S),
 | ||
|                           "=D" (dummy_value_D),
 | ||
|                           "=a" (dummy_value_a)
 | ||
| 
 | ||
|                         : "0" (width),         // ecx  // input regs
 | ||
|                           "1" (sptr),          // esi/rsi
 | ||
|                           "2" (dp),            // edi/rdi
 | ||
| #if defined(PNG_x86_64_USE_GOTPCREL)           // formerly _const4:
 | ||
|                           "3" (&_c64._amask5_3_0)  // (0x0000000000FFFFFFLL)
 | ||
| #else
 | ||
|                           "3" (&_amask5_3_0)   // eax (0x0000000000FFFFFFLL)
 | ||
| #endif
 | ||
| 
 | ||
| #if defined(PNG_CLOBBER_MMX_REGS_SUPPORTED)
 | ||
|                         : "%mm0", "%mm1", "%mm2"       // clobber list
 | ||
|                         , "%mm3", "%mm4"
 | ||
| #endif
 | ||
|                      );
 | ||
|                   }
 | ||
|                } /* end of pixel_bytes == 3 */
 | ||
| 
 | ||
|                //--------------------------------------------------------------
 | ||
|                else if (pixel_bytes == BPP4)
 | ||
|                {
 | ||
|                   if (((pass == 4) || (pass == 5)) && width)
 | ||
|                   {
 | ||
|                      int width_mmx = ((width >> 1) << 1) ;
 | ||
|                      width -= width_mmx;        // 0,1 pixels => 0,4 bytes
 | ||
|                      if (width_mmx)
 | ||
|                      {
 | ||
|                         __asm__ __volatile__ (
 | ||
|                            "sub  $4, %1             \n\t"
 | ||
|                            "sub  $12, %2            \n\t"
 | ||
| 
 | ||
|                         ".loop4_pass4:              \n\t"
 | ||
|                            "movq (%1), %%mm0        \n\t" // 7 6 5 4 3 2 1 0
 | ||
|                            "movq %%mm0, %%mm1       \n\t" // 7 6 5 4 3 2 1 0
 | ||
|                            "punpckldq %%mm0, %%mm0  \n\t" // 3 2 1 0 3 2 1 0
 | ||
|                            "punpckhdq %%mm1, %%mm1  \n\t" // 7 6 5 4 7 6 5 4
 | ||
|                            "movq %%mm0, (%2)        \n\t"
 | ||
|                            "sub  $8, %1             \n\t"
 | ||
|                            "movq %%mm1, 8(%2)       \n\t"
 | ||
|                            "sub  $16, %2            \n\t"
 | ||
|                            "subl $2, %%ecx          \n\t"
 | ||
|                            "jnz .loop4_pass4        \n\t"
 | ||
|                            "EMMS                    \n\t" // DONE
 | ||
| 
 | ||
|                            : "=c" (dummy_value_c),        // output regs (dummy)
 | ||
|                              "=S" (dummy_value_S),
 | ||
|                              "=D" (dummy_value_D)
 | ||
| 
 | ||
|                            : "0" (width_mmx),     // ecx  // input regs
 | ||
|                              "1" (sptr),          // esi/rsi
 | ||
|                              "2" (dp)             // edi/rdi
 | ||
| 
 | ||
| #if defined(CLOBBER_MMX_REGS_SUPPORTED)
 | ||
|                            : "%mm0", "%mm1"               // clobber list
 | ||
| #endif
 | ||
|                         );
 | ||
|                      }
 | ||
| 
 | ||
|                      sptr -= (width_mmx*BPP4 - BPP4); // sign fixed
 | ||
|                      dp -= (width_mmx*2*BPP4 - BPP4); // sign fixed
 | ||
|                      for (i = width; i; i--)
 | ||
|                      {
 | ||
|                         png_byte v[8];
 | ||
|                         int j;
 | ||
|                         sptr -= BPP4;
 | ||
|                         png_memcpy(v, sptr, BPP4);
 | ||
|                         for (j = 0; j < png_pass_inc[pass]; j++)
 | ||
|                         {
 | ||
|                            dp -= BPP4;
 | ||
|                            png_memcpy(dp, v, BPP4);
 | ||
|                         }
 | ||
|                      }
 | ||
|                   }
 | ||
|                   else if (((pass == 2) || (pass == 3)) && width)
 | ||
|                   {
 | ||
|                      int width_mmx = ((width >> 1) << 1);
 | ||
|                      width -= width_mmx;        // 0,1 pixels => 0,4 bytes
 | ||
|                      if (width_mmx)
 | ||
|                      {
 | ||
|                         __asm__ __volatile__ (
 | ||
|                            "sub  $4, %1             \n\t"
 | ||
|                            "sub  $28, %2            \n\t"
 | ||
| 
 | ||
|                         ".loop4_pass2:              \n\t"
 | ||
|                            "movq (%1), %%mm0        \n\t" // 7 6 5 4 3 2 1 0
 | ||
|                            "movq %%mm0, %%mm1       \n\t" // 7 6 5 4 3 2 1 0
 | ||
|                            "punpckldq %%mm0, %%mm0  \n\t" // 3 2 1 0 3 2 1 0
 | ||
|                            "punpckhdq %%mm1, %%mm1  \n\t" // 7 6 5 4 7 6 5 4
 | ||
|                            "movq %%mm0, (%2)        \n\t"
 | ||
|                            "movq %%mm0, 8(%2)       \n\t"
 | ||
|                            "movq %%mm1, 16(%2)      \n\t"
 | ||
|                            "movq %%mm1, 24(%2)      \n\t"
 | ||
|                            "sub  $8, %1             \n\t"
 | ||
|                            "sub  $32, %2            \n\t"
 | ||
|                            "subl $2, %%ecx          \n\t"
 | ||
|                            "jnz .loop4_pass2        \n\t"
 | ||
|                            "EMMS                    \n\t" // DONE
 | ||
| 
 | ||
|                            : "=c" (dummy_value_c),        // output regs (dummy)
 | ||
|                              "=S" (dummy_value_S),
 | ||
|                              "=D" (dummy_value_D)
 | ||
| 
 | ||
|                            : "0" (width_mmx),     // ecx  // input regs
 | ||
|                              "1" (sptr),          // esi/rsi
 | ||
|                              "2" (dp)             // edi/rdi
 | ||
| 
 | ||
| #if defined(CLOBBER_MMX_REGS_SUPPORTED)
 | ||
|                            : "%mm0", "%mm1"               // clobber list
 | ||
| #endif
 | ||
|                         );
 | ||
|                      }
 | ||
| 
 | ||
|                      sptr -= (width_mmx*4 - 4); // sign fixed
 | ||
|                      dp -= (width_mmx*16 - 4);  // sign fixed
 | ||
|                      for (i = width; i; i--)
 | ||
|                      {
 | ||
|                         png_byte v[8];
 | ||
|                         int j;
 | ||
|                         sptr -= 4;
 | ||
|                         png_memcpy(v, sptr, 4);
 | ||
|                         for (j = 0; j < png_pass_inc[pass]; j++)
 | ||
|                         {
 | ||
|                            dp -= 4;
 | ||
|                            png_memcpy(dp, v, 4);
 | ||
|                         }
 | ||
|                      }
 | ||
|                   }
 | ||
|                   else if (width)  // && ((pass == 0) || (pass == 1))
 | ||
|                   {
 | ||
|                      int width_mmx = ((width >> 1) << 1);
 | ||
|                      width -= width_mmx;        // 0,1 pixels => 0,4 bytes
 | ||
|                      if (width_mmx)
 | ||
|                      {
 | ||
|                         __asm__ __volatile__ (
 | ||
|                            "sub  $4, %1             \n\t"
 | ||
|                            "sub  $60, %2            \n\t"
 | ||
| 
 | ||
|                         ".loop4_pass0:              \n\t"
 | ||
|                            "movq (%1), %%mm0        \n\t" // 7 6 5 4 3 2 1 0
 | ||
|                            "movq %%mm0, %%mm1       \n\t" // 7 6 5 4 3 2 1 0
 | ||
|                            "punpckldq %%mm0, %%mm0  \n\t" // 3 2 1 0 3 2 1 0
 | ||
|                            "punpckhdq %%mm1, %%mm1  \n\t" // 7 6 5 4 7 6 5 4
 | ||
|                            "movq %%mm0, (%2)        \n\t"
 | ||
|                            "movq %%mm0, 8(%2)       \n\t"
 | ||
|                            "movq %%mm0, 16(%2)      \n\t"
 | ||
|                            "movq %%mm0, 24(%2)      \n\t"
 | ||
|                            "movq %%mm1, 32(%2)      \n\t"
 | ||
|                            "movq %%mm1, 40(%2)      \n\t"
 | ||
|                            "movq %%mm1, 48(%2)      \n\t"
 | ||
|                            "sub  $8, %1             \n\t"
 | ||
|                            "movq %%mm1, 56(%2)      \n\t"
 | ||
|                            "sub  $64, %2            \n\t"
 | ||
|                            "subl $2, %%ecx          \n\t"
 | ||
|                            "jnz .loop4_pass0        \n\t"
 | ||
|                            "EMMS                    \n\t" // DONE
 | ||
| 
 | ||
|                            : "=c" (dummy_value_c),        // output regs (dummy)
 | ||
|                              "=S" (dummy_value_S),
 | ||
|                              "=D" (dummy_value_D)
 | ||
| 
 | ||
|                            : "0" (width_mmx),     // ecx  // input regs
 | ||
|                              "1" (sptr),          // esi/rsi
 | ||
|                              "2" (dp)             // edi/rdi
 | ||
| 
 | ||
| #if defined(PNG_CLOBBER_MMX_REGS_SUPPORTED)
 | ||
|                            : "%mm0", "%mm1"               // clobber list
 | ||
| #endif
 | ||
|                         );
 | ||
|                      }
 | ||
| 
 | ||
|                      sptr -= (width_mmx*4 - 4); // sign fixed
 | ||
|                      dp -= (width_mmx*32 - 4);  // sign fixed
 | ||
|                      for (i = width; i; i--)
 | ||
|                      {
 | ||
|                         png_byte v[8];
 | ||
|                         int j;
 | ||
|                         sptr -= 4;
 | ||
|                         png_memcpy(v, sptr, 4);
 | ||
|                         for (j = 0; j < png_pass_inc[pass]; j++)
 | ||
|                         {
 | ||
|                            dp -= 4;
 | ||
|                            png_memcpy(dp, v, 4);
 | ||
|                         }
 | ||
|                      }
 | ||
|                   }
 | ||
|                } /* end of pixel_bytes == 4 */
 | ||
| 
 | ||
|                //--------------------------------------------------------------
 | ||
|                else if (pixel_bytes == 1)
 | ||
|                {
 | ||
|                   if (((pass == 4) || (pass == 5)) && width)
 | ||
|                   {
 | ||
|                      int width_mmx = ((width >> 3) << 3);
 | ||
|                      width -= width_mmx;        // 0-3 pixels => 0-3 bytes
 | ||
|                      if (width_mmx)
 | ||
|                      {
 | ||
|                         __asm__ __volatile__ (
 | ||
|                            "sub  $7, %1             \n\t"
 | ||
|                            "sub  $15, %2            \n\t"
 | ||
| 
 | ||
|                         ".loop1_pass4:              \n\t"
 | ||
|                            "movq (%1), %%mm0        \n\t" // 7 6 5 4 3 2 1 0
 | ||
|                            "movq %%mm0, %%mm1       \n\t" // 7 6 5 4 3 2 1 0
 | ||
|                            "punpcklbw %%mm0, %%mm0  \n\t" // 3 3 2 2 1 1 0 0
 | ||
|                            "punpckhbw %%mm1, %%mm1  \n\t" // 7 7 6 6 5 5 4 4
 | ||
|                            "movq %%mm1, 8(%2)       \n\t"
 | ||
|                            "sub  $8, %1             \n\t"
 | ||
|                            "movq %%mm0, (%2)        \n\t"
 | ||
|                            "sub  $16, %2            \n\t"
 | ||
|                            "subl $8, %%ecx          \n\t"
 | ||
|                            "jnz .loop1_pass4        \n\t"
 | ||
|                            "EMMS                    \n\t" // DONE
 | ||
| 
 | ||
|                            : "=c" (dummy_value_c),        // output regs (dummy)
 | ||
|                              "=S" (dummy_value_S),
 | ||
|                              "=D" (dummy_value_D)
 | ||
| 
 | ||
|                            : "0" (width_mmx),     // ecx  // input regs
 | ||
|                              "1" (sptr),          // esi/rsi
 | ||
|                              "2" (dp)             // edi/rdi
 | ||
| 
 | ||
| #if defined(PNG_CLOBBER_MMX_REGS_SUPPORTED)
 | ||
|                            : "%mm0", "%mm1"               // clobber list
 | ||
| #endif
 | ||
|                         );
 | ||
|                      }
 | ||
| 
 | ||
|                      sptr -= width_mmx;
 | ||
|                      dp -= width_mmx*2;
 | ||
|                      for (i = width; i; i--)
 | ||
|                      {
 | ||
|                         int j;
 | ||
| 
 | ||
|                         for (j = 0; j < png_pass_inc[pass]; j++)
 | ||
|                         {
 | ||
|                            *dp-- = *sptr;
 | ||
|                         }
 | ||
|                         --sptr;
 | ||
|                      }
 | ||
|                   }
 | ||
|                   else if (((pass == 2) || (pass == 3)) && width)
 | ||
|                   {
 | ||
|                      int width_mmx = ((width >> 2) << 2);
 | ||
|                      width -= width_mmx;        // 0-3 pixels => 0-3 bytes
 | ||
|                      if (width_mmx)
 | ||
|                      {
 | ||
|                         __asm__ __volatile__ (
 | ||
|                            "sub  $3, %1             \n\t"
 | ||
|                            "sub  $15, %2            \n\t"
 | ||
| 
 | ||
|                         ".loop1_pass2:              \n\t"
 | ||
|                            "movd (%1), %%mm0        \n\t" // x x x x 3 2 1 0
 | ||
|                            "punpcklbw %%mm0, %%mm0  \n\t" // 3 3 2 2 1 1 0 0
 | ||
|                            "movq %%mm0, %%mm1       \n\t" // 3 3 2 2 1 1 0 0
 | ||
|                            "punpcklwd %%mm0, %%mm0  \n\t" // 1 1 1 1 0 0 0 0
 | ||
|                            "punpckhwd %%mm1, %%mm1  \n\t" // 3 3 3 3 2 2 2 2
 | ||
|                            "movq %%mm0, (%2)        \n\t"
 | ||
|                            "sub  $4, %1             \n\t"
 | ||
|                            "movq %%mm1, 8(%2)       \n\t"
 | ||
|                            "sub  $16, %2            \n\t"
 | ||
|                            "subl $4, %%ecx          \n\t"
 | ||
|                            "jnz .loop1_pass2        \n\t"
 | ||
|                            "EMMS                    \n\t" // DONE
 | ||
| 
 | ||
|                            : "=c" (dummy_value_c),        // output regs (dummy)
 | ||
|                              "=S" (dummy_value_S),
 | ||
|                              "=D" (dummy_value_D)
 | ||
| 
 | ||
|                            : "0" (width_mmx),     // ecx  // input regs
 | ||
|                              "1" (sptr),          // esi/rsi
 | ||
|                              "2" (dp)             // edi/rdi
 | ||
| 
 | ||
| #if defined(PNG_CLOBBER_MMX_REGS_SUPPORTED)
 | ||
|                            : "%mm0", "%mm1"               // clobber list
 | ||
| #endif
 | ||
|                         );
 | ||
|                      }
 | ||
| 
 | ||
|                      sptr -= width_mmx;
 | ||
|                      dp -= width_mmx*4;
 | ||
|                      for (i = width; i; i--)
 | ||
|                      {
 | ||
|                         int j;
 | ||
| 
 | ||
|                         for (j = 0; j < png_pass_inc[pass]; j++)
 | ||
|                         {
 | ||
|                            *dp-- = *sptr;
 | ||
|                         }
 | ||
|                         --sptr;
 | ||
|                      }
 | ||
|                   }
 | ||
|                   else if (width)  // && ((pass == 0) || (pass == 1))
 | ||
|                   {
 | ||
|                      int width_mmx = ((width >> 2) << 2);
 | ||
|                      width -= width_mmx;        // 0-3 pixels => 0-3 bytes
 | ||
|                      if (width_mmx)
 | ||
|                      {
 | ||
|                         __asm__ __volatile__ (
 | ||
|                            "sub  $3, %1             \n\t"
 | ||
|                            "sub  $31, %2            \n\t"
 | ||
| 
 | ||
|                         ".loop1_pass0:              \n\t"
 | ||
|                            "movd (%1), %%mm0        \n\t" // x x x x 3 2 1 0
 | ||
|                            "movq %%mm0, %%mm1       \n\t" // x x x x 3 2 1 0
 | ||
|                            "punpcklbw %%mm0, %%mm0  \n\t" // 3 3 2 2 1 1 0 0
 | ||
|                            "movq %%mm0, %%mm2       \n\t" // 3 3 2 2 1 1 0 0
 | ||
|                            "punpcklwd %%mm0, %%mm0  \n\t" // 1 1 1 1 0 0 0 0
 | ||
|                            "movq %%mm0, %%mm3       \n\t" // 1 1 1 1 0 0 0 0
 | ||
|                            "punpckldq %%mm0, %%mm0  \n\t" // 0 0 0 0 0 0 0 0
 | ||
|                            "punpckhdq %%mm3, %%mm3  \n\t" // 1 1 1 1 1 1 1 1
 | ||
|                            "movq %%mm0, (%2)        \n\t"
 | ||
|                            "punpckhwd %%mm2, %%mm2  \n\t" // 3 3 3 3 2 2 2 2
 | ||
|                            "movq %%mm3, 8(%2)       \n\t"
 | ||
|                            "movq %%mm2, %%mm4       \n\t" // 3 3 3 3 2 2 2 2
 | ||
|                            "punpckldq %%mm2, %%mm2  \n\t" // 2 2 2 2 2 2 2 2
 | ||
|                            "punpckhdq %%mm4, %%mm4  \n\t" // 3 3 3 3 3 3 3 3
 | ||
|                            "movq %%mm2, 16(%2)      \n\t"
 | ||
|                            "sub  $4, %1             \n\t"
 | ||
|                            "movq %%mm4, 24(%2)      \n\t"
 | ||
|                            "sub  $32, %2            \n\t"
 | ||
|                            "subl $4, %%ecx          \n\t"
 | ||
|                            "jnz .loop1_pass0        \n\t"
 | ||
|                            "EMMS                    \n\t" // DONE
 | ||
| 
 | ||
|                            : "=c" (dummy_value_c),        // output regs (dummy)
 | ||
|                              "=S" (dummy_value_S),
 | ||
|                              "=D" (dummy_value_D)
 | ||
| 
 | ||
|                            : "0" (width_mmx),     // ecx  // input regs
 | ||
|                              "1" (sptr),          // esi/rsi
 | ||
|                              "2" (dp)             // edi/rdi
 | ||
| 
 | ||
| #if defined(PNG_CLOBBER_MMX_REGS_SUPPORTED)
 | ||
|                            : "%mm0", "%mm1", "%mm2"       // clobber list
 | ||
|                            , "%mm3", "%mm4"
 | ||
| #endif
 | ||
|                         );
 | ||
|                      }
 | ||
| 
 | ||
|                      sptr -= width_mmx;
 | ||
|                      dp -= width_mmx*8;
 | ||
|                      for (i = width; i; i--)
 | ||
|                      {
 | ||
|                         int j;
 | ||
| 
 | ||
|                        /* I simplified this part in version 1.0.4e
 | ||
|                         * here and in several other instances where
 | ||
|                         * pixel_bytes == 1  -- GR-P
 | ||
|                         *
 | ||
|                         * Original code:
 | ||
|                         *
 | ||
|                         * png_byte v[8];
 | ||
|                         * png_memcpy(v, sptr, pixel_bytes);
 | ||
|                         * for (j = 0; j < png_pass_inc[pass]; j++)
 | ||
|                         * {
 | ||
|                         *    png_memcpy(dp, v, pixel_bytes);
 | ||
|                         *    dp -= pixel_bytes;
 | ||
|                         * }
 | ||
|                         * sptr -= pixel_bytes;
 | ||
|                         *
 | ||
|                         * Replacement code is in the next three lines:
 | ||
|                         */
 | ||
| 
 | ||
|                         for (j = 0; j < png_pass_inc[pass]; j++)
 | ||
|                         {
 | ||
|                            *dp-- = *sptr;
 | ||
|                         }
 | ||
|                         --sptr;
 | ||
|                      }
 | ||
|                   }
 | ||
|                } /* end of pixel_bytes == 1 */
 | ||
| 
 | ||
|                //--------------------------------------------------------------
 | ||
|                else if (pixel_bytes == BPP2)
 | ||
|                {
 | ||
|                   if (((pass == 4) || (pass == 5)) && width)
 | ||
|                   {
 | ||
|                      int width_mmx = ((width >> 1) << 1) ;
 | ||
|                      width -= width_mmx;        // 0,1 pixels => 0,2 bytes
 | ||
|                      if (width_mmx)
 | ||
|                      {
 | ||
|                         __asm__ __volatile__ (
 | ||
|                            "sub  $2, %1             \n\t"
 | ||
|                            "sub  $6, %2             \n\t"
 | ||
| 
 | ||
|                         ".loop2_pass4:              \n\t"
 | ||
|                            "movd (%1), %%mm0        \n\t" // x x x x 3 2 1 0
 | ||
|                            "punpcklwd %%mm0, %%mm0  \n\t" // 3 2 3 2 1 0 1 0
 | ||
|                            "sub  $4, %1             \n\t"
 | ||
|                            "movq %%mm0, (%2)        \n\t"
 | ||
|                            "sub  $8, %2             \n\t"
 | ||
|                            "subl $2, %%ecx          \n\t"
 | ||
|                            "jnz .loop2_pass4        \n\t"
 | ||
|                            "EMMS                    \n\t" // DONE
 | ||
| 
 | ||
|                            : "=c" (dummy_value_c),        // output regs (dummy)
 | ||
|                              "=S" (dummy_value_S),
 | ||
|                              "=D" (dummy_value_D)
 | ||
| 
 | ||
|                            : "0" (width_mmx),     // ecx  // input regs
 | ||
|                              "1" (sptr),          // esi/rsi
 | ||
|                              "2" (dp)             // edi/rdi
 | ||
| 
 | ||
| #if defined(PNG_CLOBBER_MMX_REGS_SUPPORTED)
 | ||
|                            : "%mm0"                       // clobber list
 | ||
| #endif
 | ||
|                         );
 | ||
|                      }
 | ||
| 
 | ||
|                      sptr -= (width_mmx*BPP2 - BPP2); // sign fixed
 | ||
|                      dp -= (width_mmx*2*BPP2 - BPP2); // sign fixed
 | ||
|                      for (i = width; i; i--)
 | ||
|                      {
 | ||
|                         png_byte v[8];
 | ||
|                         int j;
 | ||
|                         sptr -= BPP2;
 | ||
|                         png_memcpy(v, sptr, BPP2);
 | ||
|                         for (j = 0; j < png_pass_inc[pass]; j++)
 | ||
|                         {
 | ||
|                            dp -= BPP2;
 | ||
|                            png_memcpy(dp, v, BPP2);
 | ||
|                         }
 | ||
|                      }
 | ||
|                   }
 | ||
|                   else if (((pass == 2) || (pass == 3)) && width)
 | ||
|                   {
 | ||
|                      int width_mmx = ((width >> 1) << 1) ;
 | ||
|                      width -= width_mmx;        // 0,1 pixels => 0,2 bytes
 | ||
|                      if (width_mmx)
 | ||
|                      {
 | ||
|                         __asm__ __volatile__ (
 | ||
|                            "sub  $2, %1             \n\t"
 | ||
|                            "sub  $14, %2            \n\t"
 | ||
| 
 | ||
|                         ".loop2_pass2:              \n\t"
 | ||
|                            "movd (%1), %%mm0        \n\t" // x x x x 3 2 1 0
 | ||
|                            "punpcklwd %%mm0, %%mm0  \n\t" // 3 2 3 2 1 0 1 0
 | ||
|                            "movq %%mm0, %%mm1       \n\t" // 3 2 3 2 1 0 1 0
 | ||
|                            "punpckldq %%mm0, %%mm0  \n\t" // 1 0 1 0 1 0 1 0
 | ||
|                            "punpckhdq %%mm1, %%mm1  \n\t" // 3 2 3 2 3 2 3 2
 | ||
|                            "movq %%mm0, (%2)        \n\t"
 | ||
|                            "sub  $4, %1             \n\t"
 | ||
|                            "movq %%mm1, 8(%2)       \n\t"
 | ||
|                            "sub  $16, %2            \n\t"
 | ||
|                            "subl $2, %%ecx          \n\t"
 | ||
|                            "jnz .loop2_pass2        \n\t"
 | ||
|                            "EMMS                    \n\t" // DONE
 | ||
| 
 | ||
|                            : "=c" (dummy_value_c),        // output regs (dummy)
 | ||
|                              "=S" (dummy_value_S),
 | ||
|                              "=D" (dummy_value_D)
 | ||
| 
 | ||
|                            : "0" (width_mmx),     // ecx  // input regs
 | ||
|                              "1" (sptr),          // esi/rsi
 | ||
|                              "2" (dp)             // edi/rdi
 | ||
| 
 | ||
| #if defined(CLOBBER_MMX_REGS_SUPPORTED)
 | ||
|                            : "%mm0", "%mm1"               // clobber list
 | ||
| #endif
 | ||
|                         );
 | ||
|                      }
 | ||
| 
 | ||
|                      sptr -= (width_mmx*2 - 2); // sign fixed
 | ||
|                      dp -= (width_mmx*8 - 2);   // sign fixed
 | ||
|                      for (i = width; i; i--)
 | ||
|                      {
 | ||
|                         png_byte v[8];
 | ||
|                         int j;
 | ||
|                         sptr -= 2;
 | ||
|                         png_memcpy(v, sptr, 2);
 | ||
|                         for (j = 0; j < png_pass_inc[pass]; j++)
 | ||
|                         {
 | ||
|                            dp -= 2;
 | ||
|                            png_memcpy(dp, v, 2);
 | ||
|                         }
 | ||
|                      }
 | ||
|                   }
 | ||
|                   else if (width)  // && ((pass == 0) || (pass == 1))
 | ||
|                   {
 | ||
|                      int width_mmx = ((width >> 1) << 1);
 | ||
|                      width -= width_mmx;        // 0,1 pixels => 0,2 bytes
 | ||
|                      if (width_mmx)
 | ||
|                      {
 | ||
|                         __asm__ __volatile__ (
 | ||
|                            "sub  $2, %1             \n\t"
 | ||
|                            "sub  $30, %2            \n\t"
 | ||
| 
 | ||
|                         ".loop2_pass0:              \n\t"
 | ||
|                            "movd (%1), %%mm0        \n\t" // x x x x 3 2 1 0
 | ||
|                            "punpcklwd %%mm0, %%mm0  \n\t" // 3 2 3 2 1 0 1 0
 | ||
|                            "movq %%mm0, %%mm1       \n\t" // 3 2 3 2 1 0 1 0
 | ||
|                            "punpckldq %%mm0, %%mm0  \n\t" // 1 0 1 0 1 0 1 0
 | ||
|                            "punpckhdq %%mm1, %%mm1  \n\t" // 3 2 3 2 3 2 3 2
 | ||
|                            "movq %%mm0, (%2)        \n\t"
 | ||
|                            "movq %%mm0, 8(%2)       \n\t"
 | ||
|                            "movq %%mm1, 16(%2)      \n\t"
 | ||
|                            "sub  $4, %1             \n\t"
 | ||
|                            "movq %%mm1, 24(%2)      \n\t"
 | ||
|                            "sub  $32, %2            \n\t"
 | ||
|                            "subl $2, %%ecx          \n\t"
 | ||
|                            "jnz .loop2_pass0        \n\t"
 | ||
|                            "EMMS                    \n\t" // DONE
 | ||
| 
 | ||
|                            : "=c" (dummy_value_c),        // output regs (dummy)
 | ||
|                              "=S" (dummy_value_S),
 | ||
|                              "=D" (dummy_value_D)
 | ||
| 
 | ||
|                            : "0" (width_mmx),     // ecx  // input regs
 | ||
|                              "1" (sptr),          // esi/rsi
 | ||
|                              "2" (dp)             // edi/rdi
 | ||
| 
 | ||
| #if defined(CLOBBER_MMX_REGS_SUPPORTED)
 | ||
|                            : "%mm0", "%mm1"               // clobber list
 | ||
| #endif
 | ||
|                         );
 | ||
|                      }
 | ||
| 
 | ||
|                      sptr -= (width_mmx*2 - 2); // sign fixed
 | ||
|                      dp -= (width_mmx*16 - 2);  // sign fixed
 | ||
|                      for (i = width; i; i--)
 | ||
|                      {
 | ||
|                         png_byte v[8];
 | ||
|                         int j;
 | ||
|                         sptr -= 2;
 | ||
|                         png_memcpy(v, sptr, 2);
 | ||
|                         for (j = 0; j < png_pass_inc[pass]; j++)
 | ||
|                         {
 | ||
|                            dp -= 2;
 | ||
|                            png_memcpy(dp, v, 2);
 | ||
|                         }
 | ||
|                      }
 | ||
|                   }
 | ||
|                } /* end of pixel_bytes == 2 */
 | ||
| 
 | ||
|                //--------------------------------------------------------------
 | ||
|                else if (pixel_bytes == BPP8)
 | ||
|                {
 | ||
| // GRR TEST:  should work, but needs testing (special 64-bit version of rpng2?)
 | ||
|                   // GRR NOTE:  no need to combine passes here!
 | ||
|                   if (((pass == 4) || (pass == 5)) && width)
 | ||
|                   {
 | ||
|                      // source is 8-byte RRGGBBAA
 | ||
|                      // dest is 16-byte RRGGBBAA RRGGBBAA
 | ||
|                      __asm__ __volatile__ (
 | ||
|                         "sub  $8, %2             \n\t" // start of last block
 | ||
| 
 | ||
|                      ".loop8_pass4:              \n\t"
 | ||
|                         "movq (%1), %%mm0        \n\t" // 7 6 5 4 3 2 1 0
 | ||
|                         "movq %%mm0, (%2)        \n\t"
 | ||
|                         "sub  $8, %1             \n\t"
 | ||
|                         "movq %%mm0, 8(%2)       \n\t"
 | ||
|                         "sub  $16, %2            \n\t"
 | ||
|                         "decl %%ecx              \n\t"
 | ||
|                         "jnz .loop8_pass4        \n\t"
 | ||
|                         "EMMS                    \n\t" // DONE
 | ||
| 
 | ||
|                         : "=c" (dummy_value_c),        // output regs (dummy)
 | ||
|                           "=S" (dummy_value_S),
 | ||
|                           "=D" (dummy_value_D)
 | ||
| 
 | ||
|                         : "0" (width),         // ecx  // input regs
 | ||
|                           "1" (sptr),          // esi/rsi
 | ||
|                           "2" (dp)             // edi/rdi
 | ||
| 
 | ||
| #if defined(CLOBBER_MMX_REGS_SUPPORTED)
 | ||
|                         : "%mm0"                       // clobber list
 | ||
| #endif
 | ||
|                      );
 | ||
|                   }
 | ||
|                   else if (((pass == 2) || (pass == 3)) && width)
 | ||
|                   {
 | ||
|                      // source is 8-byte RRGGBBAA
 | ||
|                      // dest is 32-byte RRGGBBAA RRGGBBAA RRGGBBAA RRGGBBAA
 | ||
|                      // (recall that expansion is _in place_:  sptr and dp
 | ||
|                      //  both point at locations within same row buffer)
 | ||
|                      __asm__ __volatile__ (
 | ||
|                         "sub  $24, %2            \n\t" // start of last block
 | ||
| 
 | ||
|                      ".loop8_pass2:              \n\t"
 | ||
|                         "movq (%1), %%mm0        \n\t" // 7 6 5 4 3 2 1 0
 | ||
|                         "movq %%mm0, (%2)        \n\t"
 | ||
|                         "movq %%mm0, 8(%2)       \n\t"
 | ||
|                         "movq %%mm0, 16(%2)      \n\t"
 | ||
|                         "sub  $8, %1             \n\t"
 | ||
|                         "movq %%mm0, 24(%2)      \n\t"
 | ||
|                         "sub  $32, %2            \n\t"
 | ||
|                         "decl %%ecx              \n\t"
 | ||
|                         "jnz .loop8_pass2        \n\t"
 | ||
|                         "EMMS                    \n\t" // DONE
 | ||
| 
 | ||
|                         : "=c" (dummy_value_c),        // output regs (dummy)
 | ||
|                           "=S" (dummy_value_S),
 | ||
|                           "=D" (dummy_value_D)
 | ||
| 
 | ||
|                         : "0" (width),         // ecx  // input regs
 | ||
|                           "1" (sptr),          // esi/rsi
 | ||
|                           "2" (dp)             // edi/rdi
 | ||
| 
 | ||
| #if defined(PNG_CLOBBER_MMX_REGS_SUPPORTED)
 | ||
|                         : "%mm0"                       // clobber list
 | ||
| #endif
 | ||
|                      );
 | ||
|                   }
 | ||
|                   else if (width)  // && ((pass == 0) || (pass == 1))
 | ||
|                   {
 | ||
|                      // source is 8-byte RRGGBBAA
 | ||
|                      // dest is 64-byte RRGGBBAA RRGGBBAA RRGGBBAA RRGGBBAA ...
 | ||
|                      __asm__ __volatile__ (
 | ||
|                         "sub  $56, %2            \n\t" // start of last block
 | ||
| 
 | ||
|                      ".loop8_pass0:              \n\t"
 | ||
|                         "movq (%1), %%mm0        \n\t" // 7 6 5 4 3 2 1 0
 | ||
|                         "movq %%mm0, (%2)        \n\t"
 | ||
|                         "movq %%mm0, 8(%2)       \n\t"
 | ||
|                         "movq %%mm0, 16(%2)      \n\t"
 | ||
|                         "movq %%mm0, 24(%2)      \n\t"
 | ||
|                         "movq %%mm0, 32(%2)      \n\t"
 | ||
|                         "movq %%mm0, 40(%2)      \n\t"
 | ||
|                         "movq %%mm0, 48(%2)      \n\t"
 | ||
|                         "sub  $8, %1             \n\t"
 | ||
|                         "movq %%mm0, 56(%2)      \n\t"
 | ||
|                         "sub  $64, %2            \n\t"
 | ||
|                         "decl %%ecx              \n\t"
 | ||
|                         "jnz .loop8_pass0        \n\t"
 | ||
|                         "EMMS                    \n\t" // DONE
 | ||
| 
 | ||
|                         : "=c" (dummy_value_c),        // output regs (dummy)
 | ||
|                           "=S" (dummy_value_S),
 | ||
|                           "=D" (dummy_value_D)
 | ||
| 
 | ||
|                         : "0" (width),         // ecx  // input regs
 | ||
|                           "1" (sptr),          // esi/rsi
 | ||
|                           "2" (dp)             // edi/rdi
 | ||
| 
 | ||
| #if defined(PNG_CLOBBER_MMX_REGS_SUPPORTED)
 | ||
|                         : "%mm0"                       // clobber list
 | ||
| #endif
 | ||
|                      );
 | ||
|                   }
 | ||
|                } /* end of pixel_bytes == 8 */
 | ||
| 
 | ||
|                //--------------------------------------------------------------
 | ||
|                else if (pixel_bytes == BPP6)   // why no MMX for this case?
 | ||
|                {
 | ||
|                   for (i = width; i; i--)
 | ||
|                   {
 | ||
|                      png_byte v[8];
 | ||
|                      int j;
 | ||
|                      png_memcpy(v, sptr, BPP6);
 | ||
|                      for (j = 0; j < png_pass_inc[pass]; j++)
 | ||
|                      {
 | ||
|                         png_memcpy(dp, v, BPP6);
 | ||
|                         dp -= BPP6;
 | ||
|                      }
 | ||
|                      sptr -= BPP6;
 | ||
|                   }
 | ||
|                } /* end of pixel_bytes == 6 */
 | ||
| 
 | ||
|                //--------------------------------------------------------------
 | ||
|                else
 | ||
|                {
 | ||
|                   // ERROR:  SHOULD NEVER BE REACHED
 | ||
| #if defined(PNG_DEBUG)
 | ||
|                   png_debug(1, "Internal libpng logic error (GCC "
 | ||
|                     "png_do_read_interlace() _mmx_supported)\n");
 | ||
| #endif
 | ||
|                }
 | ||
| 
 | ||
|             } // end of _mmx_supported ========================================
 | ||
| 
 | ||
|             else /* MMX not supported:  use modified C code - takes advantage
 | ||
|                   *   of inlining of png_memcpy for a constant */
 | ||
|             {
 | ||
|                if (pixel_bytes == BPP3)
 | ||
|                {
 | ||
|                   for (i = width; i; i--)
 | ||
|                   {
 | ||
|                      png_byte v[8];
 | ||
|                      int j;
 | ||
|                      png_memcpy(v, sptr, BPP3);
 | ||
|                      for (j = 0; j < png_pass_inc[pass]; j++)
 | ||
|                      {
 | ||
|                         png_memcpy(dp, v, BPP3);
 | ||
|                         dp -= BPP3;
 | ||
|                      }
 | ||
|                      sptr -= BPP3;
 | ||
|                   }
 | ||
|                }
 | ||
|                else if (pixel_bytes == BPP4)
 | ||
|                {
 | ||
|                   for (i = width; i; i--)
 | ||
|                   {
 | ||
|                      png_byte v[8];
 | ||
|                      int j;
 | ||
|                      png_memcpy(v, sptr, BPP4);
 | ||
|                      for (j = 0; j < png_pass_inc[pass]; j++)
 | ||
|                      {
 | ||
| #if defined(PNG_DEBUG) && defined(PNG_1_0_X)  // row_buf_size gone in 1.2.x
 | ||
|                         if (dp < row || dp+3 > row+png_ptr->row_buf_size)
 | ||
|                         {
 | ||
|                            printf("dp out of bounds: row=%10p, dp=%10p, "
 | ||
|                              "rp=%10p\n", row, dp, row+png_ptr->row_buf_size);
 | ||
|                            printf("row_buf_size=%lu\n", png_ptr->row_buf_size);
 | ||
|                         }
 | ||
| #endif
 | ||
|                         png_memcpy(dp, v, BPP4);
 | ||
|                         dp -= BPP4;
 | ||
|                      }
 | ||
|                      sptr -= BPP4;
 | ||
|                   }
 | ||
|                }
 | ||
|                else if (pixel_bytes == 1)
 | ||
|                {
 | ||
|                   for (i = width; i; i--)
 | ||
|                   {
 | ||
|                      int j;
 | ||
|                      for (j = 0; j < png_pass_inc[pass]; j++)
 | ||
|                      {
 | ||
|                         *dp-- = *sptr;
 | ||
|                      }
 | ||
|                      --sptr;
 | ||
|                   }
 | ||
|                }
 | ||
|                else if (pixel_bytes == BPP2)
 | ||
|                {
 | ||
|                   for (i = width; i; i--)
 | ||
|                   {
 | ||
|                      png_byte v[8];
 | ||
|                      int j;
 | ||
|                      png_memcpy(v, sptr, BPP2);
 | ||
|                      for (j = 0; j < png_pass_inc[pass]; j++)
 | ||
|                      {
 | ||
|                         png_memcpy(dp, v, BPP2);
 | ||
|                         dp -= BPP2;
 | ||
|                      }
 | ||
|                      sptr -= BPP2;
 | ||
|                   }
 | ||
|                }
 | ||
|                else if (pixel_bytes == BPP6)
 | ||
|                {
 | ||
|                   for (i = width; i; i--)
 | ||
|                   {
 | ||
|                      png_byte v[8];
 | ||
|                      int j;
 | ||
|                      png_memcpy(v, sptr, BPP6);
 | ||
|                      for (j = 0; j < png_pass_inc[pass]; j++)
 | ||
|                      {
 | ||
|                         png_memcpy(dp, v, BPP6);
 | ||
|                         dp -= BPP6;
 | ||
|                      }
 | ||
|                      sptr -= BPP6;
 | ||
|                   }
 | ||
|                }
 | ||
|                else if (pixel_bytes == BPP8)
 | ||
|                {
 | ||
|                   for (i = width; i; i--)
 | ||
|                   {
 | ||
|                      png_byte v[8];
 | ||
|                      int j;
 | ||
|                      png_memcpy(v, sptr, BPP8);
 | ||
|                      for (j = 0; j < png_pass_inc[pass]; j++)
 | ||
|                      {
 | ||
|                         png_memcpy(dp, v, BPP8);
 | ||
|                         dp -= BPP8;
 | ||
|                      }
 | ||
|                      sptr -= BPP8;
 | ||
|                   }
 | ||
|                }
 | ||
|                else
 | ||
|                {
 | ||
|                   // ERROR:  SHOULD NEVER BE REACHED
 | ||
| #if defined(PNG_DEBUG)
 | ||
|                   png_debug(1, "Internal libpng logic error (GCC "
 | ||
|                     "png_do_read_interlace() !_mmx_supported)\n");
 | ||
| #endif
 | ||
|                }
 | ||
| 
 | ||
|             } /* end if (MMX not supported) */
 | ||
|             break;
 | ||
|          } /* end default (8-bit or larger) */
 | ||
|       } /* end switch (row_info->pixel_depth) */
 | ||
| 
 | ||
|       row_info->width = final_width;
 | ||
| 
 | ||
|       row_info->rowbytes = PNG_ROWBYTES(row_info->pixel_depth,final_width);
 | ||
|    }
 | ||
| 
 | ||
| } /* end png_do_read_interlace() */
 | ||
| 
 | ||
| #endif /* PNG_HAVE_MMX_READ_INTERLACE */
 | ||
| #endif /* PNG_READ_INTERLACING_SUPPORTED */
 | ||
| 
 | ||
| 
 | ||
| 
 | ||
| #if defined(PNG_HAVE_MMX_READ_FILTER_ROW)
 | ||
| #if defined(PNG_MMX_READ_FILTER_AVG_SUPPORTED)
 | ||
| 
 | ||
| //===========================================================================//
 | ||
| //                                                                           //
 | ||
| //           P N G _ R E A D _ F I L T E R _ R O W _ M M X _ A V G           //
 | ||
| //                                                                           //
 | ||
| //===========================================================================//
 | ||
| 
 | ||
| // Optimized code for PNG Average filter decoder
 | ||
| 
 | ||
| static void /* PRIVATE */
 | ||
| png_read_filter_row_mmx_avg(png_row_infop row_info, png_bytep row,
 | ||
|                             png_bytep prev_row)
 | ||
| {
 | ||
|    unsigned FullLength, MMXLength;  // png_uint_32 is actually 64-bit on x86-64
 | ||
|    int bpp;
 | ||
|    int dummy_value_a;
 | ||
|    int dummy_value_c;   // fix 'forbidden register 2 (cx) was spilled' error
 | ||
|    int dummy_value_d;
 | ||
|    png_bytep dummy_value_S;
 | ||
|    png_bytep dummy_value_D;
 | ||
|    int diff; //     __attribute__((used));
 | ||
| 
 | ||
|    bpp = (row_info->pixel_depth + 7) >> 3;  // calc number of bytes per pixel
 | ||
|    FullLength = row_info->rowbytes;         // number of bytes to filter
 | ||
| 
 | ||
|    __asm__ __volatile__ (
 | ||
|    "avg_top:                       \n\t"
 | ||
|       SAVE_GOT_ebx
 | ||
|       SAVE_r15
 | ||
|       SAVE_ebp
 | ||
|       // initialize address pointers and offset
 | ||
| //pre "movl row, %5                \n\t" // edi/rdi:  ptr to Avg(x)
 | ||
|       "xorl %%ebx, %%ebx           \n\t" // ebx:  x
 | ||
| //pre "movl prev_row, %4           \n\t" // esi/rsi:  ptr to Prior(x)
 | ||
|       "mov  %5, " PDX "            \n\t" // copy of row ptr...
 | ||
| //pre "subl bpp, " PDX "           \n\t" // (bpp is preloaded into ecx)
 | ||
|       "sub  " PCX "," PDX "        \n\t" // edx/rdx:  ptr to Raw(x-bpp)
 | ||
| //pre "movl FullLength, %%eax      \n\t" // bring in via eax...
 | ||
|       SAVE_FullLength                    // ...but store for later use
 | ||
|       "xorl %%eax, %%eax           \n\t"
 | ||
| 
 | ||
|       // Compute the Raw value for the first bpp bytes
 | ||
|       //    Raw(x) = Avg(x) + (Prior(x)/2)
 | ||
|    "avg_rlp:                       \n\t"
 | ||
|       "movb (%4," PBX ",), %%al    \n\t" // load al with Prior(x)
 | ||
|       "incl %%ebx                  \n\t"
 | ||
|       "shrb %%al                   \n\t" // divide by 2
 | ||
|       "addb -1(%5," PBX ",), %%al  \n\t" // add Avg(x); -1 to offset inc ebx
 | ||
| //pre "cmpl bpp, %%ebx             \n\t" // (bpp is preloaded into ecx)
 | ||
|       "cmpl %%ecx, %%ebx           \n\t"
 | ||
|       "movb %%al, -1(%5," PBX ",)  \n\t" // write Raw(x); -1 to offset inc ebx
 | ||
|       "jb avg_rlp                  \n\t" // mov does not affect flags
 | ||
| 
 | ||
|       // get # of bytes to alignment (32-bit mask _would_ be good enough
 | ||
|       // [computing delta], but 32-bit ops are zero-extended on 64-bit, argh)
 | ||
|       // (if swapped edx and ebp, could do 8-bit or 16-bit mask...FIXME?)
 | ||
|       "mov  %5, " PBP "            \n\t" // take start of row
 | ||
|       "add  " PBX "," PBP "        \n\t" // add bpp
 | ||
|       "add  $0xf, " PBP "          \n\t" // add 7+8 to incr past alignment bdry
 | ||
| //    "andl $0xfffffff8, %%ebp     \n\t" // mask to alignment boundary (32-bit!)
 | ||
|       CLEAR_BOTTOM_3_BITS  PBP    "\n\t" // mask to alignment boundary
 | ||
|       "sub  %5, " PBP "            \n\t" // subtract row ptr again => ebp =
 | ||
|       "jz avg_go                   \n\t" //  target value of ebx at alignment
 | ||
| 
 | ||
|       "xorl %%ecx, %%ecx           \n\t"
 | ||
| 
 | ||
|       // fix alignment
 | ||
|       // Compute the Raw value for the bytes up to the alignment boundary
 | ||
|       //    Raw(x) = Avg(x) + ((Raw(x-bpp) + Prior(x))/2)
 | ||
|    "avg_lp1:                       \n\t"
 | ||
|       "xorl %%eax, %%eax           \n\t"
 | ||
|       "movb (%4," PBX ",), %%cl    \n\t" // load cl with Prior(x)
 | ||
|       "movb (" PDX "," PBX ",), %%al \n\t" // load al with Raw(x-bpp)
 | ||
|       "addw %%cx, %%ax             \n\t"
 | ||
|       "incl %%ebx                  \n\t"
 | ||
|       "shrw %%ax                   \n\t" // divide by 2
 | ||
|       "addb -1(%5," PBX ",), %%al  \n\t" // add Avg(x); -1 to offset inc ebx
 | ||
|       "cmpl %%ebp, %%ebx           \n\t" // check if at alignment boundary
 | ||
|       "movb %%al, -1(%5," PBX ",)  \n\t" // write Raw(x); -1 to offset inc ebx
 | ||
|       "jb avg_lp1                  \n\t" // repeat until at alignment boundary
 | ||
| 
 | ||
|    "avg_go:                        \n\t"
 | ||
|       RESTORE_FullLength "%%eax    \n\t" // FullLength -> eax
 | ||
|       "movl %%eax, %%ecx           \n\t" // copy -> ecx
 | ||
|       "subl %%ebx, %%eax           \n\t" // subtract alignment fix
 | ||
|       "andl $0x00000007, %%eax     \n\t" // calc bytes over mult of 8
 | ||
|       "subl %%eax, %%ecx           \n\t" // sub over-bytes from original length
 | ||
| //out "movl %%ecx, MMXLength       \n\t"
 | ||
|       "movl %%ebp, %%eax           \n\t" // ebp = diff, but no reg constraint(?)
 | ||
|       RESTORE_ebp                        //  (could swap ebp and edx functions)
 | ||
|       RESTORE_r15
 | ||
|       RESTORE_GOT_ebx
 | ||
| 
 | ||
| // "There is no way for you to specify that an input operand is modified
 | ||
| // without also specifying it as an output operand."  [makes sense]
 | ||
| 
 | ||
| // "Unless an output operand has the `&' constraint modifier, GCC may
 | ||
| // allocate it in the same register as an unrelated input operand, on the
 | ||
| // assumption the inputs are consumed before the outputs are produced."
 | ||
| // [trying to _force_ this]
 | ||
| 
 | ||
| // "`='   Means that this operand is write-only for this instruction:
 | ||
| //        the previous value is discarded and replaced by output data."
 | ||
| //        [operand == variable name, presumably]
 | ||
| 
 | ||
|       // output regs
 | ||
|       // these are operands 0-1 (originally 0-3):
 | ||
|       : "=c" (MMXLength),      // %0 -> %0
 | ||
|         "=a" (diff)            // %3 -> %1
 | ||
| //      "=S" (dummy_value_S),  // %1 -> GONE
 | ||
| //      "=D" (dummy_value_D),  // %2 -> GONE
 | ||
| 
 | ||
|       // input regs
 | ||
|       // these are operands 2-5 (originally 4-7); two of their constraints say
 | ||
|       // they must go in same places as operands 0-1 (originally 0-3) above:
 | ||
|       : "0" (bpp),         // %4 -> %2 ecx
 | ||
|         "1" (FullLength),  // %7 -> %3 eax
 | ||
|         "S" (prev_row),    // %5 -> %4 esi/rsi
 | ||
|         "D" (row)          // %6 -> %5 edi/rdi
 | ||
| 
 | ||
|       : "%edx"                           // clobber list
 | ||
|         _CLOBBER_r15
 | ||
|         _CLOBBER_ebp
 | ||
|         _CLOBBER_GOT_ebx
 | ||
|    );
 | ||
| 
 | ||
|    // now do the math for the rest of the row
 | ||
|    switch (bpp)
 | ||
|    {
 | ||
|       case 3:
 | ||
|       {
 | ||
| //       _ShiftBpp = 24;    // == 3 * 8
 | ||
| //       _ShiftRem = 40;    // == 64 - 24
 | ||
| 
 | ||
|          __asm__ __volatile__ (
 | ||
|             // re-init address pointers and offset
 | ||
|             LOAD_GOT_rbp
 | ||
|             "movq " AMASK5_3_0 ", %%mm7    \n\t" // _amask5_3_0 -> mm7
 | ||
| // preload  "movl  diff, %%ecx             \n\t" // ecx:  x = offset to
 | ||
|                                                  //  alignment boundary
 | ||
|             "movq " LB_CARRY_MASK ", %%mm5 \n\t" // [interleave for parallel.?]
 | ||
| // preload  "movl  row, %1                 \n\t" // edi:  Avg(x)
 | ||
|             "movq " HB_CLEAR_MASK ", %%mm4 \n\t" // _HBClearMask -> mm4
 | ||
| // preload  "movl  prev_row, %0            \n\t" // esi:  Prior(x)
 | ||
|             RESTORE_rbp
 | ||
| 
 | ||
|             // prime the pump:  load the first Raw(x-bpp) data set
 | ||
|             "movq  -8(%1," PCX ",), %%mm2 \n\t"// load previous aligned 8 bytes
 | ||
|                                                // (correct pos. in loop below)
 | ||
|          "avg_3lp:                        \n\t"
 | ||
|             "movq  (%1," PCX ",), %%mm0   \n\t" // load mm0 with Avg(x)
 | ||
|             "movq  %%mm5, %%mm3           \n\t"
 | ||
|             "psrlq $40, %%mm2             \n\t" // correct position Raw(x-bpp)
 | ||
|                                                 // data
 | ||
|             "movq  (%0," PCX ",), %%mm1   \n\t" // load mm1 with Prior(x)
 | ||
|             "movq  %%mm7, %%mm6           \n\t"
 | ||
|             "pand  %%mm1, %%mm3           \n\t" // get lsb for each prevrow byte
 | ||
|             "psrlq $1, %%mm1              \n\t" // divide prev_row bytes by 2
 | ||
|             "pand  %%mm4, %%mm1           \n\t" // clear invalid bit 7 of each
 | ||
|                                                 // byte
 | ||
|             "paddb %%mm1, %%mm0           \n\t" // add (Prev_row/2) to Avg for
 | ||
|                                                 // each byte
 | ||
|             // add 1st active group (Raw(x-bpp)/2) to average with LBCarry
 | ||
|             "movq  %%mm3, %%mm1           \n\t" // now use mm1 for getting
 | ||
|                                                 // LBCarrys
 | ||
|             "pand  %%mm2, %%mm1           \n\t" // get LBCarrys for each byte
 | ||
|                                                 // where both lsb's were == 1
 | ||
|                                                 // (valid only for active group)
 | ||
|             "psrlq $1, %%mm2              \n\t" // divide raw bytes by 2
 | ||
|             "pand  %%mm4, %%mm2           \n\t" // clear invalid bit 7 of each
 | ||
|                                                 // byte
 | ||
|             "paddb %%mm1, %%mm2           \n\t" // add LBCarrys to Raw(x-bpp)/2
 | ||
|                                                 // for each byte
 | ||
|             "pand  %%mm6, %%mm2           \n\t" // leave only Active Group 1
 | ||
|                                                 // bytes to add to Avg
 | ||
|             "paddb %%mm2, %%mm0           \n\t" // add (Raw/2) + LBCarrys to
 | ||
|                                                 // Avg for each Active byte
 | ||
|             // add 2nd active group (Raw(x-bpp)/2) to average with _LBCarry
 | ||
|             "psllq $24, %%mm6             \n\t" // shift the mm6 mask to cover
 | ||
|                                                 // bytes 3-5
 | ||
|             "movq  %%mm0, %%mm2           \n\t" // mov updated Raws to mm2
 | ||
|             "psllq $24, %%mm2             \n\t" // shift data to pos. correctly
 | ||
|             "movq  %%mm3, %%mm1           \n\t" // now use mm1 for getting
 | ||
|                                                 // LBCarrys
 | ||
|             "pand  %%mm2, %%mm1           \n\t" // get LBCarrys for each byte
 | ||
|                                                 // where both lsb's were == 1
 | ||
|                                                 // (valid only for active group)
 | ||
|             "psrlq $1, %%mm2              \n\t" // divide raw bytes by 2
 | ||
|             "pand  %%mm4, %%mm2           \n\t" // clear invalid bit 7 of each
 | ||
|                                                 // byte
 | ||
|             "paddb %%mm1, %%mm2           \n\t" // add LBCarrys to Raw(x-bpp)/2
 | ||
|                                                 // for each byte
 | ||
|             "pand  %%mm6, %%mm2           \n\t" // leave only Active Group 2
 | ||
|                                                 // bytes to add to Avg
 | ||
|             "paddb %%mm2, %%mm0           \n\t" // add (Raw/2) + LBCarrys to
 | ||
|                                                 // Avg for each Active byte
 | ||
| 
 | ||
|             // add 3rd active group (Raw(x-bpp)/2) to average with _LBCarry
 | ||
|             "psllq $24, %%mm6             \n\t" // shift mm6 mask to cover last
 | ||
|                                                 // two bytes
 | ||
|             "movq  %%mm0, %%mm2           \n\t" // mov updated Raws to mm2
 | ||
|             "psllq $24, %%mm2             \n\t" // shift data to pos. correctly
 | ||
|                               // Data need be shifted only once here to
 | ||
|                               // get the correct x-bpp offset.
 | ||
|             "movq  %%mm3, %%mm1           \n\t" // now use mm1 for getting
 | ||
|                                                 // LBCarrys
 | ||
|             "pand  %%mm2, %%mm1           \n\t" // get LBCarrys for each byte
 | ||
|                                                 // where both
 | ||
|                               // lsb's were == 1 (only valid for active group)
 | ||
|             "psrlq $1, %%mm2              \n\t" // divide raw bytes by 2
 | ||
|             "pand  %%mm4, %%mm2           \n\t" // clear invalid bit 7 of each
 | ||
|                                                 // byte
 | ||
|             "paddb %%mm1, %%mm2           \n\t" // add LBCarrys to Raw(x-bpp)/2
 | ||
|                                                 // for each byte
 | ||
|             "pand  %%mm6, %%mm2           \n\t" // leave only Active Group 2
 | ||
|                                                 // bytes to add to Avg
 | ||
|             "addl  $8, %%ecx              \n\t"
 | ||
|             "paddb %%mm2, %%mm0           \n\t" // add (Raw/2) + LBCarrys to
 | ||
|                                                 // Avg for each Active byte
 | ||
|             // now ready to write back to memory
 | ||
|             "movq  %%mm0, -8(%1," PCX ",) \n\t"
 | ||
|             // move updated Raw(x) to use as Raw(x-bpp) for next loop
 | ||
|             "cmpl  %%eax, %%ecx           \n\t" // MMXLength
 | ||
|             "movq  %%mm0, %%mm2           \n\t" // mov updated Raw(x) to mm2
 | ||
|             "jb avg_3lp                   \n\t"
 | ||
| 
 | ||
|             : "=S" (dummy_value_S),            // output regs (dummy)
 | ||
|               "=D" (dummy_value_D),
 | ||
|               "=c" (dummy_value_c),
 | ||
|               "=a" (dummy_value_a)
 | ||
| 
 | ||
|             : "0" (prev_row),    // esi/rsi    // input regs
 | ||
|               "1" (row),         // edi/rdi
 | ||
|               "2" (diff),        // ecx
 | ||
|               "3" (MMXLength)    // eax
 | ||
| 
 | ||
| #if defined(CLOBBER_MMX_REGS_SUPPORTED)
 | ||
|             : "%mm0", "%mm1", "%mm2", "%mm3"   // clobber list
 | ||
|             , "%mm4", "%mm5", "%mm6", "%mm7"
 | ||
| #endif
 | ||
|          );
 | ||
|       }
 | ||
|       break;  // end 3 bpp
 | ||
| 
 | ||
|       case 4:   // formerly shared with 6 bpp case via _ShiftBpp and _ShiftRem,
 | ||
|       {         // but loop uses all 8 MMX regs, and psrlq/psllq require 64-bit
 | ||
|                 // mem (PIC/.so problems), MMX reg (none left), or immediate
 | ||
| //       _ShiftBpp = bpp << 3;        // 32 (psllq)
 | ||
| //       _ShiftRem = 64 - _ShiftBpp;  // 32 (psrlq)
 | ||
| 
 | ||
|          __asm__ __volatile__ (
 | ||
|             LOAD_GOT_rbp
 | ||
|             "movq " HB_CLEAR_MASK ", %%mm4 \n\t" // _HBClearMask -> mm4
 | ||
|             "movq " LB_CARRY_MASK ", %%mm5 \n\t" // _LBCarryMask -> mm5
 | ||
|             // re-init address pointers and offset
 | ||
| // preload  "movl  diff, %%ecx             \n\t" // ecx:  x = offset to
 | ||
|                                                  // alignment boundary
 | ||
|             "movq " AMASK0_8_0 ", %%mm7    \n\t" // _amask0_8_0 -> mm7
 | ||
|             RESTORE_rbp
 | ||
| 
 | ||
|             // ... and clear all bytes except for 1st active group
 | ||
| // preload  "movl  row, %1               \n\t" // edi:  Avg(x)
 | ||
|             "psrlq $32, %%mm7            \n\t" // was _ShiftRem
 | ||
| // preload  "movl  prev_row, %0          \n\t" // esi:  Prior(x)
 | ||
|             "movq  %%mm7, %%mm6          \n\t"
 | ||
|             "psllq $32, %%mm6            \n\t" // mask for 2nd active group
 | ||
| 
 | ||
|             // prime the pump:  load the first Raw(x-bpp) data set
 | ||
|             "movq -8(%1," PCX ",), %%mm2 \n\t" // load previous aligned 8 bytes
 | ||
|                                              // (we correct pos. in loop below)
 | ||
|          "avg_4lp:                       \n\t"
 | ||
|             "movq (%1," PCX ",), %%mm0   \n\t"
 | ||
|             "psrlq $32, %%mm2            \n\t" // shift data to pos. correctly
 | ||
|             "movq (%0," PCX ",), %%mm1   \n\t"
 | ||
|             // add (Prev_row/2) to average
 | ||
|             "movq %%mm5, %%mm3           \n\t"
 | ||
|             "pand %%mm1, %%mm3           \n\t" // get lsb for each prev_row byte
 | ||
|             "psrlq $1, %%mm1             \n\t" // divide prev_row bytes by 2
 | ||
|             "pand  %%mm4, %%mm1          \n\t" // clear invalid bit 7 of each
 | ||
|                                                // byte
 | ||
|             "paddb %%mm1, %%mm0          \n\t" // add (Prev_row/2) to Avg for
 | ||
|                                                // each byte
 | ||
|             // add 1st active group (Raw(x-bpp)/2) to average with _LBCarry
 | ||
|             "movq %%mm3, %%mm1           \n\t" // now use mm1 for getting
 | ||
|                                                // LBCarrys
 | ||
|             "pand %%mm2, %%mm1           \n\t" // get LBCarrys for each byte
 | ||
|                                                // where both
 | ||
|                               // lsb's were == 1 (only valid for active group)
 | ||
|             "psrlq $1, %%mm2             \n\t" // divide raw bytes by 2
 | ||
|             "pand  %%mm4, %%mm2          \n\t" // clear invalid bit 7 of each
 | ||
|                                                // byte
 | ||
|             "paddb %%mm1, %%mm2          \n\t" // add LBCarrys to (Raw(x-bpp)/2)
 | ||
|                                                // for each byte
 | ||
|             "pand %%mm7, %%mm2           \n\t" // leave only Active Group 1
 | ||
|                                                // bytes to add to Avg
 | ||
|             "paddb %%mm2, %%mm0          \n\t" // add (Raw/2) + LBCarrys to Avg
 | ||
|                                                // for each Active
 | ||
|                               // byte
 | ||
|             // add 2nd active group (Raw(x-bpp)/2) to average with _LBCarry
 | ||
|             "movq %%mm0, %%mm2           \n\t" // mov updated Raws to mm2
 | ||
|             "psllq $32, %%mm2            \n\t" // shift data to pos. correctly
 | ||
|             "addl $8, %%ecx              \n\t"
 | ||
|             "movq %%mm3, %%mm1           \n\t" // now use mm1 for getting
 | ||
|                                                // LBCarrys
 | ||
|             "pand %%mm2, %%mm1           \n\t" // get LBCarrys for each byte
 | ||
|                                                // where both
 | ||
|                               // lsb's were == 1 (only valid for active group)
 | ||
|             "psrlq $1, %%mm2             \n\t" // divide raw bytes by 2
 | ||
|             "pand  %%mm4, %%mm2          \n\t" // clear invalid bit 7 of each
 | ||
|                                                // byte
 | ||
|             "paddb %%mm1, %%mm2          \n\t" // add LBCarrys to (Raw(x-bpp)/2)
 | ||
|                                                // for each byte
 | ||
|             "pand %%mm6, %%mm2           \n\t" // leave only Active Group 2
 | ||
|                                                // bytes to add to Avg
 | ||
|             "paddb %%mm2, %%mm0          \n\t" // add (Raw/2) + LBCarrys to
 | ||
|                                                // Avg for each Active byte
 | ||
|             "cmpl %%eax, %%ecx           \n\t" // MMXLength
 | ||
|             // now ready to write back to memory
 | ||
|             "movq %%mm0, -8(%1," PCX ",) \n\t"
 | ||
|             // prep Raw(x-bpp) for next loop
 | ||
|             "movq %%mm0, %%mm2           \n\t" // mov updated Raws to mm2
 | ||
|             "jb avg_4lp                  \n\t"
 | ||
| 
 | ||
|             : "=S" (dummy_value_S),            // output regs (dummy)
 | ||
|               "=D" (dummy_value_D),
 | ||
|               "=c" (dummy_value_c),
 | ||
|               "=a" (dummy_value_a)
 | ||
| 
 | ||
|             : "0" (prev_row),    // esi/rsi    // input regs
 | ||
|               "1" (row),         // edi/rdi
 | ||
|               "2" (diff),        // ecx
 | ||
|               "3" (MMXLength)    // eax
 | ||
| 
 | ||
| #if defined(CLOBBER_MMX_REGS_SUPPORTED)
 | ||
|             : "%mm0", "%mm1", "%mm2", "%mm3"   // clobber list
 | ||
|             , "%mm4", "%mm5", "%mm6", "%mm7"
 | ||
| #endif
 | ||
|          );
 | ||
|       }
 | ||
|       break;  // end 4 bpp
 | ||
| 
 | ||
|       case 1:
 | ||
|       {
 | ||
|          __asm__ __volatile__ (
 | ||
|             // re-init address pointers and offset
 | ||
| // preload  "movl diff, %%ecx            \n\t" // ecx: x = offset to align. bdry
 | ||
| // preload  "movl row, %1                \n\t" // edi/rdi:  Avg(x)
 | ||
| // preload  "movl FullLength, %%eax      \n\t"
 | ||
|             "cmpl %%eax, %%ecx           \n\t" // test if offset at end of array
 | ||
|             "jnb avg_1end                \n\t"
 | ||
| 
 | ||
|             SAVE_ebp
 | ||
| 
 | ||
|             // do Avg decode for remaining bytes
 | ||
| // preload  "movl prev_row, %0           \n\t" // esi/rsi:  Prior(x)
 | ||
|             "mov  %1, " PBP "            \n\t" // copy of row pointer...
 | ||
|             "dec  " PBP "                \n\t" // ebp/rbp:  Raw(x-bpp)
 | ||
|             "xorl %%edx, %%edx           \n\t" // zero edx before using dl & dx
 | ||
|                                                //  in loop below
 | ||
|             SAVE_GOT_ebx
 | ||
| 
 | ||
|          "avg_1lp:                       \n\t"
 | ||
|             // Raw(x) = Avg(x) + ((Raw(x-bpp) + Prior(x))/2)
 | ||
|             "xorl %%ebx, %%ebx           \n\t"
 | ||
|             "movb (%0," PCX ",), %%dl    \n\t" // load dl with Prior(x)
 | ||
|             "movb (" PBP "," PCX ",), %%bl \n\t" // load bl with Raw(x-bpp)
 | ||
|             "addw %%dx, %%bx             \n\t"
 | ||
|             "incl %%ecx                  \n\t"
 | ||
|             "shrw %%bx                   \n\t" // divide by 2
 | ||
|             "addb -1(%1," PCX ",), %%bl  \n\t" // add Avg(x); -1 to offset
 | ||
|                                                // inc ecx
 | ||
|             "cmpl %%eax, %%ecx           \n\t" // check if at end of array
 | ||
|             "movb %%bl, -1(%1," PCX ",)  \n\t" // write back Raw(x);
 | ||
|                          // mov does not affect flags; -1 to offset inc ecx
 | ||
|             "jb avg_1lp                  \n\t"
 | ||
| 
 | ||
|             RESTORE_GOT_ebx
 | ||
|             RESTORE_ebp
 | ||
| 
 | ||
|          "avg_1end:                      \n\t"
 | ||
| 
 | ||
|             : "=S" (dummy_value_S),            // output regs (dummy)
 | ||
|               "=D" (dummy_value_D),
 | ||
|               "=c" (dummy_value_c),
 | ||
|               "=a" (dummy_value_a)
 | ||
| 
 | ||
|             : "0" (prev_row),    // esi/rsi    // input regs
 | ||
|               "1" (row),         // edi/rdi
 | ||
|               "2" (diff),        // ecx
 | ||
|               "3" (FullLength)   // eax
 | ||
| 
 | ||
|             : "%edx"                           // clobber list
 | ||
|               _CLOBBER_GOT_ebx
 | ||
|               _CLOBBER_ebp
 | ||
|          );
 | ||
|       }
 | ||
|       return;  // end 1 bpp
 | ||
| 
 | ||
|       case 2:
 | ||
|       {
 | ||
| //       _ShiftBpp = 16;   // == 2 * 8
 | ||
| //       _ShiftRem = 48;   // == 64 - _ShiftBpp
 | ||
| 
 | ||
|          __asm__ __volatile__ (
 | ||
|             LOAD_GOT_rbp
 | ||
|             // load (former) _ActiveMask
 | ||
|             "movq " AMASK6_2_0 ", %%mm7    \n\t" // _amask6_2_0 -> mm7
 | ||
|             // re-init address pointers and offset
 | ||
| // preload  "movl  diff, %%ecx             \n\t" // ecx:  x = offset to
 | ||
|                                                  // alignment boundary
 | ||
|             "movq " LB_CARRY_MASK ", %%mm5 \n\t" // _LBCarryMask -> mm5
 | ||
| // preload  "movl  row, %1                 \n\t" // edi:  Avg(x)
 | ||
|             "movq " HB_CLEAR_MASK ", %%mm4 \n\t" // _HBClearMask -> mm4
 | ||
| // preload  "movl  prev_row, %0            \n\t" // esi:  Prior(x)
 | ||
|             RESTORE_rbp
 | ||
| 
 | ||
|             // prime the pump:  load the first Raw(x-bpp) data set
 | ||
|             "movq -8(%1," PCX ",), %%mm2 \n\t" // load previous aligned 8 bytes
 | ||
|                                              // (we correct pos. in loop below)
 | ||
|          "avg_2lp:                       \n\t"
 | ||
|             "movq (%1," PCX ",), %%mm0   \n\t"
 | ||
|             "psrlq $48, %%mm2            \n\t" // shift data to pos. correctly
 | ||
|             "movq (%0," PCX ",), %%mm1   \n\t" //  (GRR BUGFIX:  was psllq)
 | ||
|             // add (Prev_row/2) to average
 | ||
|             "movq %%mm5, %%mm3           \n\t"
 | ||
|             "pand %%mm1, %%mm3           \n\t" // get lsb for each prev_row byte
 | ||
|             "psrlq $1, %%mm1             \n\t" // divide prev_row bytes by 2
 | ||
|             "pand  %%mm4, %%mm1          \n\t" // clear invalid bit 7 of each
 | ||
|                                                // byte
 | ||
|             "movq %%mm7, %%mm6           \n\t"
 | ||
|             "paddb %%mm1, %%mm0          \n\t" // add (Prev_row/2) to Avg for
 | ||
|                                                // each byte
 | ||
| 
 | ||
|             // add 1st active group (Raw(x-bpp)/2) to average with _LBCarry
 | ||
|             "movq %%mm3, %%mm1           \n\t" // now use mm1 for getting
 | ||
|                                                // LBCarrys
 | ||
|             "pand %%mm2, %%mm1           \n\t" // get LBCarrys for each byte
 | ||
|                                                // where both
 | ||
|                                                // lsb's were == 1 (only valid
 | ||
|                                                // for active group)
 | ||
|             "psrlq $1, %%mm2             \n\t" // divide raw bytes by 2
 | ||
|             "pand  %%mm4, %%mm2          \n\t" // clear invalid bit 7 of each
 | ||
|                                                // byte
 | ||
|             "paddb %%mm1, %%mm2          \n\t" // add LBCarrys to (Raw(x-bpp)/2)
 | ||
|                                                // for each byte
 | ||
|             "pand %%mm6, %%mm2           \n\t" // leave only Active Group 1
 | ||
|                                                // bytes to add to Avg
 | ||
|             "paddb %%mm2, %%mm0          \n\t" // add (Raw/2) + LBCarrys to Avg
 | ||
|                                                // for each Active byte
 | ||
| 
 | ||
|             // add 2nd active group (Raw(x-bpp)/2) to average with _LBCarry
 | ||
|             "psllq $16, %%mm6            \n\t" // shift the mm6 mask to cover
 | ||
|                                                // bytes 2 & 3
 | ||
|             "movq %%mm0, %%mm2           \n\t" // mov updated Raws to mm2
 | ||
|             "psllq $16, %%mm2            \n\t" // shift data to pos. correctly
 | ||
|             "movq %%mm3, %%mm1           \n\t" // now use mm1 for getting
 | ||
|                                                // LBCarrys
 | ||
|             "pand %%mm2, %%mm1           \n\t" // get LBCarrys for each byte
 | ||
|                                                // where both
 | ||
|                                                // lsb's were == 1 (only valid
 | ||
|                                                // for active group)
 | ||
|             "psrlq $1, %%mm2             \n\t" // divide raw bytes by 2
 | ||
|             "pand  %%mm4, %%mm2          \n\t" // clear invalid bit 7 of each
 | ||
|                                                // byte
 | ||
|             "paddb %%mm1, %%mm2          \n\t" // add LBCarrys to (Raw(x-bpp)/2)
 | ||
|                                                // for each byte
 | ||
|             "pand %%mm6, %%mm2           \n\t" // leave only Active Group 2
 | ||
|                                                // bytes to add to Avg
 | ||
|             "paddb %%mm2, %%mm0          \n\t" // add (Raw/2) + LBCarrys to
 | ||
|                                                // Avg for each Active byte
 | ||
| 
 | ||
|             // add 3rd active group (Raw(x-bpp)/2) to average with _LBCarry
 | ||
|             "psllq $16, %%mm6            \n\t" // shift the mm6 mask to cover
 | ||
|                                                // bytes 4 & 5
 | ||
|             "movq %%mm0, %%mm2           \n\t" // mov updated Raws to mm2
 | ||
|             "psllq $16, %%mm2            \n\t" // shift data to pos. correctly
 | ||
|             "movq %%mm3, %%mm1           \n\t" // now use mm1 for getting
 | ||
|                                                // LBCarrys
 | ||
|             "pand %%mm2, %%mm1           \n\t" // get LBCarrys for each byte
 | ||
|                                                // where both lsb's were == 1
 | ||
|                                                // (only valid for active group)
 | ||
|             "psrlq $1, %%mm2             \n\t" // divide raw bytes by 2
 | ||
|             "pand  %%mm4, %%mm2          \n\t" // clear invalid bit 7 of each
 | ||
|                                                // byte
 | ||
|             "paddb %%mm1, %%mm2          \n\t" // add LBCarrys to (Raw(x-bpp)/2)
 | ||
|                                                // for each byte
 | ||
|             "pand %%mm6, %%mm2           \n\t" // leave only Active Group 2
 | ||
|                                                // bytes to add to Avg
 | ||
|             "paddb %%mm2, %%mm0          \n\t" // add (Raw/2) + LBCarrys to
 | ||
|                                                // Avg for each Active byte
 | ||
| 
 | ||
|             // add 4th active group (Raw(x-bpp)/2) to average with _LBCarry
 | ||
|             "psllq $16, %%mm6            \n\t" // shift the mm6 mask to cover
 | ||
|                                                // bytes 6 & 7
 | ||
|             "movq %%mm0, %%mm2           \n\t" // mov updated Raws to mm2
 | ||
|             "psllq $16, %%mm2            \n\t" // shift data to pos. correctly
 | ||
|             "addl $8, %%ecx              \n\t"
 | ||
|             "movq %%mm3, %%mm1           \n\t" // now use mm1 for getting
 | ||
|                                                // LBCarrys
 | ||
|             "pand %%mm2, %%mm1           \n\t" // get LBCarrys for each byte
 | ||
|                                                // where both
 | ||
|                                                // lsb's were == 1 (only valid
 | ||
|                                                // for active group)
 | ||
|             "psrlq $1, %%mm2             \n\t" // divide raw bytes by 2
 | ||
|             "pand  %%mm4, %%mm2          \n\t" // clear invalid bit 7 of each
 | ||
|                                                // byte
 | ||
|             "paddb %%mm1, %%mm2          \n\t" // add LBCarrys to (Raw(x-bpp)/2)
 | ||
|                                                // for each byte
 | ||
|             "pand %%mm6, %%mm2           \n\t" // leave only Active Group 2
 | ||
|                                                // bytes to add to Avg
 | ||
|             "paddb %%mm2, %%mm0          \n\t" // add (Raw/2) + LBCarrys to
 | ||
|                                                // Avg for each Active byte
 | ||
|             "cmpl %%eax, %%ecx           \n\t" // MMXLength
 | ||
|             // now ready to write back to memory
 | ||
|             "movq %%mm0, -8(%1," PCX ",) \n\t"
 | ||
|             // prep Raw(x-bpp) for next loop
 | ||
|             "movq %%mm0, %%mm2           \n\t" // mov updated Raws to mm2
 | ||
|             "jb avg_2lp                  \n\t"
 | ||
| 
 | ||
|             : "=S" (dummy_value_S),            // output regs (dummy)
 | ||
|               "=D" (dummy_value_D),
 | ||
|               "=c" (dummy_value_c),
 | ||
|               "=a" (dummy_value_a)
 | ||
| 
 | ||
|             : "0" (prev_row),    // esi/rsi    // input regs
 | ||
|               "1" (row),         // edi/rdi
 | ||
|               "2" (diff),        // ecx
 | ||
|               "3" (MMXLength)    // eax
 | ||
| 
 | ||
| #if defined(CLOBBER_MMX_REGS_SUPPORTED)
 | ||
|             : "%mm0", "%mm1", "%mm2", "%mm3"   // clobber list
 | ||
|             , "%mm4", "%mm5", "%mm6", "%mm7"
 | ||
| #endif
 | ||
|          );
 | ||
|       }
 | ||
|       break;  // end 2 bpp
 | ||
| 
 | ||
|       case 6:   // formerly shared with 4 bpp case (see comments there)
 | ||
|       {
 | ||
| //       _ShiftBpp = bpp << 3;        // 48 (psllq)
 | ||
| //       _ShiftRem = 64 - _ShiftBpp;  // 16 (psrlq)
 | ||
| 
 | ||
|          __asm__ __volatile__ (
 | ||
|             LOAD_GOT_rbp
 | ||
|             "movq " HB_CLEAR_MASK ", %%mm4 \n\t" // _HBClearMask -> mm4
 | ||
|             "movq " LB_CARRY_MASK ", %%mm5 \n\t" // _LBCarryMask -> mm5
 | ||
|             // re-init address pointers and offset
 | ||
| // preload  "movl  diff, %%ecx             \n\t" // ecx:  x = offset to
 | ||
|                                                  // alignment boundary
 | ||
|             "movq " AMASK0_8_0 ", %%mm7    \n\t" // _amask0_8_0 -> mm7
 | ||
|             RESTORE_rbp
 | ||
| 
 | ||
|             // ... and clear all bytes except for 1st active group
 | ||
| // preload  "movl  row, %1               \n\t" // edi:  Avg(x)
 | ||
|             "psrlq $16, %%mm7            \n\t"
 | ||
| // preload  "movl  prev_row, %0          \n\t" // esi:  Prior(x)
 | ||
|             "movq  %%mm7, %%mm6          \n\t"
 | ||
|             "psllq $48, %%mm6            \n\t" // mask for 2nd active group
 | ||
| 
 | ||
|             // prime the pump:  load the first Raw(x-bpp) data set
 | ||
|             "movq -8(%1," PCX ",), %%mm2 \n\t" // load previous aligned 8 bytes
 | ||
|                                              // (we correct pos. in loop below)
 | ||
|          "avg_6lp:                       \n\t"
 | ||
|             "movq (%1," PCX ",), %%mm0   \n\t"
 | ||
|             "psrlq $16, %%mm2            \n\t" // shift data to pos. correctly
 | ||
|             "movq (%0," PCX ",), %%mm1   \n\t"
 | ||
|             // add (Prev_row/2) to average
 | ||
|             "movq %%mm5, %%mm3           \n\t"
 | ||
|             "pand %%mm1, %%mm3           \n\t" // get lsb for each prev_row byte
 | ||
|             "psrlq $1, %%mm1             \n\t" // divide prev_row bytes by 2
 | ||
|             "pand  %%mm4, %%mm1          \n\t" // clear invalid bit 7 of each
 | ||
|                                                // byte
 | ||
|             "paddb %%mm1, %%mm0          \n\t" // add (Prev_row/2) to Avg for
 | ||
|                                                // each byte
 | ||
|             // add 1st active group (Raw(x-bpp)/2) to average with _LBCarry
 | ||
|             "movq %%mm3, %%mm1           \n\t" // now use mm1 for getting
 | ||
|                                                // LBCarrys
 | ||
|             "pand %%mm2, %%mm1           \n\t" // get LBCarrys for each byte
 | ||
|                                                // where both
 | ||
|                               // lsb's were == 1 (only valid for active group)
 | ||
|             "psrlq $1, %%mm2             \n\t" // divide raw bytes by 2
 | ||
|             "pand  %%mm4, %%mm2          \n\t" // clear invalid bit 7 of each
 | ||
|                                                // byte
 | ||
|             "paddb %%mm1, %%mm2          \n\t" // add LBCarrys to (Raw(x-bpp)/2)
 | ||
|                                                // for each byte
 | ||
|             "pand %%mm7, %%mm2           \n\t" // leave only Active Group 1
 | ||
|                                                // bytes to add to Avg
 | ||
|             "paddb %%mm2, %%mm0          \n\t" // add (Raw/2) + LBCarrys to Avg
 | ||
|                                                // for each Active
 | ||
|                               // byte
 | ||
|             // add 2nd active group (Raw(x-bpp)/2) to average with _LBCarry
 | ||
|             "movq %%mm0, %%mm2           \n\t" // mov updated Raws to mm2
 | ||
|             "psllq $48, %%mm2            \n\t" // shift data to pos. correctly
 | ||
|             "addl $8, %%ecx              \n\t"
 | ||
|             "movq %%mm3, %%mm1           \n\t" // now use mm1 for getting
 | ||
|                                                // LBCarrys
 | ||
|             "pand %%mm2, %%mm1           \n\t" // get LBCarrys for each byte
 | ||
|                                                // where both
 | ||
|                               // lsb's were == 1 (only valid for active group)
 | ||
|             "psrlq $1, %%mm2             \n\t" // divide raw bytes by 2
 | ||
|             "pand  %%mm4, %%mm2          \n\t" // clear invalid bit 7 of each
 | ||
|                                                // byte
 | ||
|             "paddb %%mm1, %%mm2          \n\t" // add LBCarrys to (Raw(x-bpp)/2)
 | ||
|                                                // for each byte
 | ||
|             "pand %%mm6, %%mm2           \n\t" // leave only Active Group 2
 | ||
|                                                // bytes to add to Avg
 | ||
|             "paddb %%mm2, %%mm0          \n\t" // add (Raw/2) + LBCarrys to
 | ||
|                                                // Avg for each Active byte
 | ||
|             "cmpl %%eax, %%ecx           \n\t" // MMXLength
 | ||
|             // now ready to write back to memory
 | ||
|             "movq %%mm0, -8(%1," PCX ",) \n\t"
 | ||
|             // prep Raw(x-bpp) for next loop
 | ||
|             "movq %%mm0, %%mm2           \n\t" // mov updated Raws to mm2
 | ||
|             "jb avg_6lp                  \n\t"
 | ||
| 
 | ||
|             : "=S" (dummy_value_S),            // output regs (dummy)
 | ||
|               "=D" (dummy_value_D),
 | ||
|               "=c" (dummy_value_c),
 | ||
|               "=a" (dummy_value_a)
 | ||
| 
 | ||
|             : "0" (prev_row),    // esi/rsi    // input regs
 | ||
|               "1" (row),         // edi/rdi
 | ||
|               "2" (diff),        // ecx
 | ||
|               "3" (MMXLength)    // eax
 | ||
| 
 | ||
| #if defined(CLOBBER_MMX_REGS_SUPPORTED)
 | ||
|             : "%mm0", "%mm1", "%mm2", "%mm3"   // clobber list
 | ||
|             , "%mm4", "%mm5", "%mm6", "%mm7"
 | ||
| #endif
 | ||
|          );
 | ||
|       }
 | ||
|       break;  // end 6 bpp
 | ||
| 
 | ||
|       case 8:
 | ||
|       {
 | ||
|          __asm__ __volatile__ (
 | ||
|             // re-init address pointers and offset
 | ||
| // preload  "movl  diff, %%ecx             \n\t" // ecx:  x = offset to
 | ||
|                                                  // alignment boundary
 | ||
|             LOAD_GOT_rbp
 | ||
|             "movq " LB_CARRY_MASK ", %%mm5 \n\t" // [interleave for parallel.?]
 | ||
| // preload  "movl  row, %1                 \n\t" // edi:  Avg(x)
 | ||
|             "movq " HB_CLEAR_MASK ", %%mm4 \n\t" // _HBClearMask -> mm4
 | ||
| // preload  "movl  prev_row, %0            \n\t" // esi:  Prior(x)
 | ||
|             RESTORE_rbp
 | ||
| 
 | ||
|             // prime the pump:  load the first Raw(x-bpp) data set
 | ||
|             "movq -8(%1," PCX ",), %%mm2 \n\t" // load previous aligned 8 bytes
 | ||
|                                       // (NO NEED to correct pos. in loop below)
 | ||
| 
 | ||
|          "avg_8lp:                       \n\t"
 | ||
|             "movq (%1," PCX ",), %%mm0   \n\t"
 | ||
|             "movq %%mm5, %%mm3           \n\t"
 | ||
|             "movq (%0," PCX ",), %%mm1   \n\t"
 | ||
|             "addl $8, %%ecx              \n\t"
 | ||
|             "pand %%mm1, %%mm3           \n\t" // get lsb for each prev_row byte
 | ||
|             "psrlq $1, %%mm1             \n\t" // divide prev_row bytes by 2
 | ||
|             "pand %%mm2, %%mm3           \n\t" // get LBCarrys for each byte
 | ||
|                                                //  where both lsb's were == 1
 | ||
|             "psrlq $1, %%mm2             \n\t" // divide raw bytes by 2
 | ||
|             "pand  %%mm4, %%mm1          \n\t" // clear invalid bit 7, each byte
 | ||
|             "paddb %%mm3, %%mm0          \n\t" // add LBCarrys to Avg, each byte
 | ||
|             "pand  %%mm4, %%mm2          \n\t" // clear invalid bit 7, each byte
 | ||
|             "paddb %%mm1, %%mm0          \n\t" // add (Prev_row/2) to Avg, each
 | ||
|             "paddb %%mm2, %%mm0          \n\t" // add (Raw/2) to Avg for each
 | ||
|             "cmpl %%eax, %%ecx           \n\t" // MMXLength
 | ||
|             "movq %%mm0, -8(%1," PCX ",) \n\t"
 | ||
|             "movq %%mm0, %%mm2           \n\t" // reuse as Raw(x-bpp)
 | ||
|             "jb avg_8lp                  \n\t"
 | ||
| 
 | ||
|             : "=S" (dummy_value_S),            // output regs (dummy)
 | ||
|               "=D" (dummy_value_D),
 | ||
|               "=c" (dummy_value_c),
 | ||
|               "=a" (dummy_value_a)
 | ||
| 
 | ||
|             : "0" (prev_row),    // esi/rsi    // input regs
 | ||
|               "1" (row),         // edi/rdi
 | ||
|               "2" (diff),        // ecx
 | ||
|               "3" (MMXLength)    // eax
 | ||
| 
 | ||
| #if defined(CLOBBER_MMX_REGS_SUPPORTED)
 | ||
|             : "%mm0", "%mm1", "%mm2"           // clobber list
 | ||
|             , "%mm3", "%mm4", "%mm5"
 | ||
| #endif
 | ||
|          );
 | ||
|       }
 | ||
|       break;  // end 8 bpp
 | ||
| 
 | ||
|       default:                // bpp != 1,2,3,4,6,8:  doesn't exist
 | ||
|       {
 | ||
|          // ERROR:  SHOULD NEVER BE REACHED
 | ||
| #if defined(PNG_DEBUG)
 | ||
|          png_debug(1, "Internal libpng logic error (GCC "
 | ||
|            "png_read_filter_row_mmx_avg())\n");
 | ||
| #endif
 | ||
|       }
 | ||
|       break;
 | ||
| 
 | ||
|    } // end switch (bpp)
 | ||
| 
 | ||
|    __asm__ __volatile__ (
 | ||
|       // MMX acceleration complete; now do clean-up
 | ||
|       // check if any remaining bytes left to decode
 | ||
| //pre "movl FullLength, %%edx      \n\t"
 | ||
| //pre "movl MMXLength, %%eax       \n\t" // eax:  x == offset bytes after MMX
 | ||
| //pre "movl row, %2                \n\t" // edi:  Avg(x)
 | ||
|       "cmpl %%edx, %%eax           \n\t" // test if offset at end of array
 | ||
|       "jnb avg_end                 \n\t"
 | ||
| 
 | ||
|       SAVE_ebp
 | ||
| 
 | ||
|       // do Avg decode for remaining bytes
 | ||
| //pre "movl prev_row, %1           \n\t" // esi:  Prior(x)
 | ||
|       "mov  %2, " PBP "            \n\t" // copy of row pointer...
 | ||
| //pre "subl bpp, " PBP "           \n\t" // (bpp is preloaded into ecx)
 | ||
|       "sub  " PCX "," PBP "        \n\t" // ebp:  Raw(x-bpp)
 | ||
|       "xorl %%ecx, %%ecx           \n\t" // zero ecx before using cl & cx below
 | ||
| 
 | ||
|       SAVE_GOT_ebx
 | ||
| 
 | ||
|    "avg_lp2:                       \n\t"
 | ||
|       // Raw(x) = Avg(x) + ((Raw(x-bpp) + Prior(x))/2)
 | ||
|       "xorl %%ebx, %%ebx           \n\t"
 | ||
|       "movb (%1," PAX ",), %%cl    \n\t" // load cl with Prior(x)
 | ||
|       "movb (" PBP "," PAX ",), %%bl \n\t" // load bl with Raw(x-bpp)
 | ||
|       "addw %%cx, %%bx             \n\t"
 | ||
|       "incl %%eax                  \n\t"
 | ||
|       "shrw %%bx                   \n\t" // divide by 2
 | ||
|       "addb -1(%2," PAX ",), %%bl  \n\t" // add Avg(x); -1 to offset inc eax
 | ||
|       "cmpl %%edx, %%eax           \n\t" // check if at end of array
 | ||
|       "movb %%bl, -1(%2," PAX ",)  \n\t" // write back Raw(x) [mov does not
 | ||
|       "jb avg_lp2                  \n\t" //  affect flags; -1 to offset inc eax]
 | ||
| 
 | ||
|       RESTORE_GOT_ebx
 | ||
|       RESTORE_ebp
 | ||
| 
 | ||
|    "avg_end:                       \n\t"
 | ||
|       "EMMS                        \n\t" // end MMX; prep for poss. FP instrs.
 | ||
| 
 | ||
|       : "=c" (dummy_value_c),            // output regs (dummy)
 | ||
|         "=S" (dummy_value_S),
 | ||
|         "=D" (dummy_value_D),
 | ||
|         "=a" (dummy_value_a),
 | ||
|         "=d" (dummy_value_d)
 | ||
| 
 | ||
|       : "0" (bpp),         // ecx        // input regs
 | ||
|         "1" (prev_row),    // esi/rsi
 | ||
|         "2" (row),         // edi/rdi
 | ||
|         "3" (MMXLength),   // eax
 | ||
|         "4" (FullLength)   // edx
 | ||
| 
 | ||
|       CLOB_COLON_ebx_ebp                 // clobber list
 | ||
|         CLOBBER_GOT_ebx
 | ||
|         CLOB_COMMA_ebx_ebp
 | ||
|         CLOBBER_ebp
 | ||
|    );
 | ||
| 
 | ||
| } /* end png_read_filter_row_mmx_avg() */
 | ||
| 
 | ||
| #endif /* PNG_MMX_READ_FILTER_AVG_SUPPORTED */
 | ||
| 
 | ||
| 
 | ||
| 
 | ||
| #if defined(PNG_MMX_READ_FILTER_PAETH_SUPPORTED)
 | ||
| #if defined(PNG_x86_64_USE_GOTPCREL) || defined(PNG_THREAD_UNSAFE_OK)
 | ||
| 
 | ||
| //===========================================================================//
 | ||
| //                                                                           //
 | ||
| //         P N G _ R E A D _ F I L T E R _ R O W _ M M X _ P A E T H         //
 | ||
| //                                                                           //
 | ||
| //===========================================================================//
 | ||
| 
 | ||
| // Optimized code for PNG Paeth filter decoder
 | ||
| 
 | ||
| static void /* PRIVATE */
 | ||
| png_read_filter_row_mmx_paeth(png_row_infop row_info, png_bytep row,
 | ||
|                               png_bytep prev_row)
 | ||
| {
 | ||
|    unsigned FullLength, MMXLength;  // png_uint_32 is actually 64-bit on x86-64
 | ||
|    int bpp;
 | ||
|    int dummy_value_a;
 | ||
|    int dummy_value_c;   // fix 'forbidden register 2 (cx) was spilled' error
 | ||
|    int dummy_value_d;
 | ||
|    png_charp dummy_value_S;
 | ||
|    png_charp dummy_value_D;
 | ||
|    int diff; //     __attribute__((used));
 | ||
| 
 | ||
|    bpp = (row_info->pixel_depth + 7) >> 3;  // calc number of bytes per pixel
 | ||
|    FullLength = row_info->rowbytes;         // number of bytes to filter
 | ||
| 
 | ||
|    __asm__ __volatile__ (
 | ||
|       SAVE_GOT_ebx
 | ||
|       SAVE_r15
 | ||
|       SAVE_ebp
 | ||
| //pre "movl row, %2                \n\t" // edi/rdi
 | ||
|       "xorl %%ebx, %%ebx           \n\t" // ebx:  x offset
 | ||
| //pre "movl prev_row, %1           \n\t" // esi/rsi
 | ||
|       "xorl %%edx, %%edx           \n\t" // edx:  x-bpp offset
 | ||
| //pre "movl FullLength, %%eax      \n\t" // bring in via eax...
 | ||
|       SAVE_FullLength                    // ...but store for later use
 | ||
|       "xorl %%eax, %%eax           \n\t"
 | ||
| 
 | ||
|       // Compute the Raw value for the first bpp bytes
 | ||
|       // Note: the formula works out to be always
 | ||
|       //   Paeth(x) = Raw(x) + Prior(x)      where x < bpp
 | ||
|    "paeth_rlp:                     \n\t"
 | ||
|       "movb (%2," PBX ",), %%al    \n\t"
 | ||
|       "addb (%1," PBX ",), %%al    \n\t"
 | ||
|       "incl %%ebx                  \n\t"
 | ||
| //pre "cmpl bpp, %%ebx             \n\t" (bpp is preloaded into ecx)
 | ||
|       "cmpl %%ecx, %%ebx           \n\t"
 | ||
|       "movb %%al, -1(%2," PBX ",)  \n\t"
 | ||
|       "jb paeth_rlp                \n\t"
 | ||
| 
 | ||
|       // get # of bytes to alignment (note:  computing _delta_ of two pointers,
 | ||
|       // so hereafter %%ebp is sufficient even on 64-bit)
 | ||
|       "mov  %2, " PBP "            \n\t" // take start of row
 | ||
|       "add  " PBX "," PBP "        \n\t" // add bpp
 | ||
|       "add  $0xf, " PBP "          \n\t" // add 7+8 to incr past alignment bdry
 | ||
| //    "andl $0xfffffff8, %%ebp     \n\t" // mask to alignment boundary (32-bit!)
 | ||
|       CLEAR_BOTTOM_3_BITS  PBP    "\n\t" // mask to alignment boundary
 | ||
|       "sub  %2, " PBP "            \n\t" // subtract row ptr again => ebp =
 | ||
|       "jz paeth_go                 \n\t" //  target value of ebx at alignment
 | ||
| 
 | ||
|       "xorl %%ecx, %%ecx           \n\t"
 | ||
| 
 | ||
|       SAVE_r11_r12_r13
 | ||
| 
 | ||
|       // fix alignment
 | ||
|    "paeth_lp1:                     \n\t"
 | ||
|       "xorl %%eax, %%eax           \n\t"
 | ||
|       // pav = p - a = (a + b - c) - a = b - c
 | ||
|       "movb (%1," PBX ",), %%al    \n\t" // load Prior(x) into al
 | ||
|       "movb (%1," PDX ",), %%cl    \n\t" // load Prior(x-bpp) into cl
 | ||
|       "subl %%ecx, %%eax           \n\t" // subtract Prior(x-bpp)
 | ||
|       "movl %%eax, " pa_TEMP "     \n\t" // Save pav for later use
 | ||
|       "xorl %%eax, %%eax           \n\t"
 | ||
|       // pbv = p - b = (a + b - c) - b = a - c
 | ||
|       "movb (%2," PDX ",), %%al    \n\t" // load Raw(x-bpp) into al
 | ||
|       "subl %%ecx, %%eax           \n\t" // subtract Prior(x-bpp)
 | ||
|       "movl %%eax, %%ecx           \n\t"
 | ||
|       // pcv = p - c = (a + b - c) - c = (a - c) + (b - c) = pav + pbv
 | ||
|       "addl " pa_TEMP ", %%eax     \n\t" // pcv = pav + pbv
 | ||
|       // pc = abs(pcv)
 | ||
|       "testl $0x80000000, %%eax    \n\t"
 | ||
|       "jz paeth_pca                \n\t"
 | ||
|       "negl %%eax                  \n\t" // reverse sign of neg values
 | ||
| 
 | ||
|    "paeth_pca:                     \n\t"
 | ||
|       "movl %%eax, " pc_TEMP "     \n\t" // save pc for later use
 | ||
|       // pb = abs(pbv)
 | ||
|       "testl $0x80000000, %%ecx    \n\t"
 | ||
|       "jz paeth_pba                \n\t"
 | ||
|       "negl %%ecx                  \n\t" // reverse sign of neg values
 | ||
| 
 | ||
|    "paeth_pba:                     \n\t"
 | ||
|       "movl %%ecx, " pb_TEMP "     \n\t" // save pb for later use
 | ||
|       // pa = abs(pav)
 | ||
|       "movl " pa_TEMP ", %%eax     \n\t"
 | ||
|       "testl $0x80000000, %%eax    \n\t"
 | ||
|       "jz paeth_paa                \n\t"
 | ||
|       "negl %%eax                  \n\t" // reverse sign of neg values
 | ||
| 
 | ||
|    "paeth_paa:                     \n\t"
 | ||
|       "movl %%eax, " pa_TEMP "     \n\t" // save pa for later use
 | ||
|       // test if pa <= pb
 | ||
|       "cmpl %%ecx, %%eax           \n\t"
 | ||
|       "jna paeth_abb               \n\t"
 | ||
|       // pa > pb; now test if pb <= pc
 | ||
|       "cmpl " pc_TEMP ", %%ecx     \n\t"
 | ||
|       "jna paeth_bbc               \n\t"
 | ||
|       // pb > pc; Raw(x) = Paeth(x) + Prior(x-bpp)
 | ||
|       "movb (%1," PDX ",), %%cl    \n\t" // load Prior(x-bpp) into cl
 | ||
|       "jmp paeth_paeth             \n\t"
 | ||
| 
 | ||
|    "paeth_bbc:                     \n\t"
 | ||
|       // pb <= pc; Raw(x) = Paeth(x) + Prior(x)
 | ||
|       "movb (%1," PBX ",), %%cl    \n\t" // load Prior(x) into cl
 | ||
|       "jmp paeth_paeth             \n\t"
 | ||
| 
 | ||
|    "paeth_abb:                     \n\t"
 | ||
|       // pa <= pb; now test if pa <= pc
 | ||
|       "cmpl " pc_TEMP ", %%eax     \n\t"
 | ||
|       "jna paeth_abc               \n\t"
 | ||
|       // pa > pc; Raw(x) = Paeth(x) + Prior(x-bpp)
 | ||
|       "movb (%1," PDX ",), %%cl    \n\t" // load Prior(x-bpp) into cl
 | ||
|       "jmp paeth_paeth             \n\t"
 | ||
| 
 | ||
|    "paeth_abc:                     \n\t"
 | ||
|       // pa <= pc; Raw(x) = Paeth(x) + Raw(x-bpp)
 | ||
|       "movb (%2," PDX ",), %%cl    \n\t" // load Raw(x-bpp) into cl
 | ||
| 
 | ||
|    "paeth_paeth:                   \n\t"
 | ||
|       "incl %%ebx                  \n\t"
 | ||
|       "incl %%edx                  \n\t"
 | ||
|       // Raw(x) = (Paeth(x) + Paeth_Predictor( a, b, c )) mod 256
 | ||
|       "addb %%cl, -1(%2," PBX ",)  \n\t"
 | ||
|       "cmpl %%ebp, %%ebx           \n\t"
 | ||
|       "jb paeth_lp1                \n\t"
 | ||
| 
 | ||
|       RESTORE_r11_r12_r13
 | ||
| 
 | ||
|    "paeth_go:                      \n\t"
 | ||
|       RESTORE_FullLength "%%ecx    \n\t" // FullLength -> ecx
 | ||
|       "movl %%ecx, %%eax           \n\t"
 | ||
|       "subl %%ebx, %%eax           \n\t" // subtract alignment fix
 | ||
|       "andl $0x00000007, %%eax     \n\t" // calc bytes over mult of 8
 | ||
|       "subl %%eax, %%ecx           \n\t" // drop over bytes from original length
 | ||
| //out "movl %%ecx, MMXLength       \n\t"
 | ||
|       "movl %%ebp, %%eax           \n\t" // ebp = diff, but no reg constraint(?)
 | ||
|       RESTORE_ebp                        //  (could swap ebp and edx functions)
 | ||
|       RESTORE_r15
 | ||
|       RESTORE_GOT_ebx
 | ||
| 
 | ||
|       : "=c" (MMXLength),                // output regs
 | ||
|         "=S" (dummy_value_S),
 | ||
|         "=D" (dummy_value_D),
 | ||
|         "=a" (diff)
 | ||
| 
 | ||
|       : "0" (bpp),         // ecx        // input regs
 | ||
|         "1" (prev_row),    // esi/rsi
 | ||
|         "2" (row),         // edi/rdi
 | ||
|         "3" (FullLength)   // eax
 | ||
| 
 | ||
|       : "%edx"                           // clobber list
 | ||
|         _CLOBBER_r11_r12_r13
 | ||
|         _CLOBBER_r15
 | ||
|         _CLOBBER_ebp
 | ||
|         _CLOBBER_GOT_ebx
 | ||
|    );
 | ||
| 
 | ||
|    // now do the math for the rest of the row
 | ||
|    switch (bpp)
 | ||
|    {
 | ||
|       case 3:
 | ||
|       {
 | ||
| //       _ShiftBpp = 24;    // == bpp * 8
 | ||
| //       _ShiftRem = 40;    // == 64 - _ShiftBpp
 | ||
| 
 | ||
|          __asm__ __volatile__ (
 | ||
|             LOAD_GOT_rbp
 | ||
| // preload  "movl diff, %%ecx            \n\t"
 | ||
| // preload  "movl row, %1                \n\t" // edi/rdi
 | ||
| // preload  "movl prev_row, %0           \n\t" // esi/rsi
 | ||
|             "pxor %%mm0, %%mm0           \n\t"
 | ||
| 
 | ||
|             // prime the pump:  load the first Raw(x-bpp) data set
 | ||
|             "movq -8(%1," PCX ",), %%mm1 \n\t"
 | ||
|          "paeth_3lp:                     \n\t"
 | ||
|             "psrlq $40, %%mm1            \n\t" // shift last 3 bytes to 1st
 | ||
|                                                // 3 bytes
 | ||
|             "movq (%0," PCX ",), %%mm2   \n\t" // load b=Prior(x)
 | ||
|             "punpcklbw %%mm0, %%mm1      \n\t" // unpack High bytes of a
 | ||
|             "movq -8(%0," PCX ",), %%mm3 \n\t" // prep c=Prior(x-bpp) bytes
 | ||
|             "punpcklbw %%mm0, %%mm2      \n\t" // unpack High bytes of b
 | ||
|             "psrlq $40, %%mm3            \n\t" // shift last 3 bytes to 1st
 | ||
|                                                // 3 bytes
 | ||
|             // pav = p - a = (a + b - c) - a = b - c
 | ||
|             "movq %%mm2, %%mm4           \n\t"
 | ||
|             "punpcklbw %%mm0, %%mm3      \n\t" // unpack High bytes of c
 | ||
|             // pbv = p - b = (a + b - c) - b = a - c
 | ||
|             "movq %%mm1, %%mm5           \n\t"
 | ||
|             "psubw %%mm3, %%mm4          \n\t"
 | ||
|             "pxor %%mm7, %%mm7           \n\t"
 | ||
|             // pcv = p - c = (a + b - c) -c = (a - c) + (b - c) = pav + pbv
 | ||
|             "movq %%mm4, %%mm6           \n\t"
 | ||
|             "psubw %%mm3, %%mm5          \n\t"
 | ||
| 
 | ||
|             // pa = abs(p-a) = abs(pav)
 | ||
|             // pb = abs(p-b) = abs(pbv)
 | ||
|             // pc = abs(p-c) = abs(pcv)
 | ||
|             "pcmpgtw %%mm4, %%mm0        \n\t" // create mask pav bytes < 0
 | ||
|             "paddw %%mm5, %%mm6          \n\t"
 | ||
|             "pand %%mm4, %%mm0           \n\t" // only pav bytes < 0 in mm7
 | ||
|             "pcmpgtw %%mm5, %%mm7        \n\t" // create mask pbv bytes < 0
 | ||
|             "psubw %%mm0, %%mm4          \n\t"
 | ||
|             "pand %%mm5, %%mm7           \n\t" // only pbv bytes < 0 in mm0
 | ||
|             "psubw %%mm0, %%mm4          \n\t"
 | ||
|             "psubw %%mm7, %%mm5          \n\t"
 | ||
|             "pxor %%mm0, %%mm0           \n\t"
 | ||
|             "pcmpgtw %%mm6, %%mm0        \n\t" // create mask pcv bytes < 0
 | ||
|             "pand %%mm6, %%mm0           \n\t" // only pav bytes < 0 in mm7
 | ||
|             "psubw %%mm7, %%mm5          \n\t"
 | ||
|             "psubw %%mm0, %%mm6          \n\t"
 | ||
|             //  test pa <= pb
 | ||
|             "movq %%mm4, %%mm7           \n\t"
 | ||
|             "psubw %%mm0, %%mm6          \n\t"
 | ||
|             "pcmpgtw %%mm5, %%mm7        \n\t" // pa > pb?
 | ||
|             "movq %%mm7, %%mm0           \n\t"
 | ||
|             // use mm7 mask to merge pa & pb
 | ||
|             "pand %%mm7, %%mm5           \n\t"
 | ||
|             // use mm0 mask copy to merge a & b
 | ||
|             "pand %%mm0, %%mm2           \n\t"
 | ||
|             "pandn %%mm4, %%mm7          \n\t"
 | ||
|             "pandn %%mm1, %%mm0          \n\t"
 | ||
|             "paddw %%mm5, %%mm7          \n\t"
 | ||
|             "paddw %%mm2, %%mm0          \n\t"
 | ||
|             //  test  ((pa <= pb)? pa:pb) <= pc
 | ||
|             "pcmpgtw %%mm6, %%mm7        \n\t" // pab > pc?
 | ||
|             "pxor %%mm1, %%mm1           \n\t"
 | ||
|             "pand %%mm7, %%mm3           \n\t"
 | ||
|             "pandn %%mm0, %%mm7          \n\t"
 | ||
|             "paddw %%mm3, %%mm7          \n\t"
 | ||
|             "pxor %%mm0, %%mm0           \n\t"
 | ||
|             "packuswb %%mm1, %%mm7       \n\t"
 | ||
|             "movq (%0," PCX ",), %%mm3   \n\t" // load c=Prior(x-bpp)
 | ||
|             "pand " AMASK5_3_0 ", %%mm7  \n\t" // _amask5_3_0 (was _ActiveMask)
 | ||
|             "movq %%mm3, %%mm2           \n\t" // load b=Prior(x) step 1
 | ||
|             "paddb (%1," PCX ",), %%mm7  \n\t" // add Paeth predictor + Raw(x)
 | ||
|             "punpcklbw %%mm0, %%mm3      \n\t" // unpack High bytes of c
 | ||
|             "movq %%mm7, (%1," PCX ",)   \n\t" // write back updated value
 | ||
|             "movq %%mm7, %%mm1           \n\t" // now mm1 will be used as
 | ||
|                                                // Raw(x-bpp)
 | ||
|             // now do Paeth for 2nd set of bytes (3-5)
 | ||
|             "psrlq $24, %%mm2            \n\t" // load b=Prior(x) step 2
 | ||
|             "punpcklbw %%mm0, %%mm1      \n\t" // unpack High bytes of a
 | ||
|             "pxor %%mm7, %%mm7           \n\t"
 | ||
|             "punpcklbw %%mm0, %%mm2      \n\t" // unpack High bytes of b
 | ||
|             // pbv = p - b = (a + b - c) - b = a - c
 | ||
|             "movq %%mm1, %%mm5           \n\t"
 | ||
|             // pav = p - a = (a + b - c) - a = b - c
 | ||
|             "movq %%mm2, %%mm4           \n\t"
 | ||
|             "psubw %%mm3, %%mm5          \n\t"
 | ||
|             "psubw %%mm3, %%mm4          \n\t"
 | ||
|             // pcv = p - c = (a + b - c) -c = (a - c) + (b - c) =
 | ||
|             //       pav + pbv = pbv + pav
 | ||
|             "movq %%mm5, %%mm6           \n\t"
 | ||
|             "paddw %%mm4, %%mm6          \n\t"
 | ||
| 
 | ||
|             // pa = abs(p-a) = abs(pav)
 | ||
|             // pb = abs(p-b) = abs(pbv)
 | ||
|             // pc = abs(p-c) = abs(pcv)
 | ||
|             "pcmpgtw %%mm5, %%mm0        \n\t" // create mask pbv bytes < 0
 | ||
|             "pcmpgtw %%mm4, %%mm7        \n\t" // create mask pav bytes < 0
 | ||
|             "pand %%mm5, %%mm0           \n\t" // only pbv bytes < 0 in mm0
 | ||
|             "pand %%mm4, %%mm7           \n\t" // only pav bytes < 0 in mm7
 | ||
|             "psubw %%mm0, %%mm5          \n\t"
 | ||
|             "psubw %%mm7, %%mm4          \n\t"
 | ||
|             "psubw %%mm0, %%mm5          \n\t"
 | ||
|             "psubw %%mm7, %%mm4          \n\t"
 | ||
|             "pxor %%mm0, %%mm0           \n\t"
 | ||
|             "pcmpgtw %%mm6, %%mm0        \n\t" // create mask pcv bytes < 0
 | ||
|             "pand %%mm6, %%mm0           \n\t" // only pav bytes < 0 in mm7
 | ||
|             "psubw %%mm0, %%mm6          \n\t"
 | ||
|             //  test pa <= pb
 | ||
|             "movq %%mm4, %%mm7           \n\t"
 | ||
|             "psubw %%mm0, %%mm6          \n\t"
 | ||
|             "pcmpgtw %%mm5, %%mm7        \n\t" // pa > pb?
 | ||
|             "movq %%mm7, %%mm0           \n\t"
 | ||
|             // use mm7 mask to merge pa & pb
 | ||
|             "pand %%mm7, %%mm5           \n\t"
 | ||
|             // use mm0 mask copy to merge a & b
 | ||
|             "pand %%mm0, %%mm2           \n\t"
 | ||
|             "pandn %%mm4, %%mm7          \n\t"
 | ||
|             "pandn %%mm1, %%mm0          \n\t"
 | ||
|             "paddw %%mm5, %%mm7          \n\t"
 | ||
|             "paddw %%mm2, %%mm0          \n\t"
 | ||
|             //  test  ((pa <= pb)? pa:pb) <= pc
 | ||
|             "pcmpgtw %%mm6, %%mm7        \n\t" // pab > pc?
 | ||
|             "movq (%0," PCX ",), %%mm2   \n\t" // load b=Prior(x)
 | ||
|             "pand %%mm7, %%mm3           \n\t"
 | ||
|             "pandn %%mm0, %%mm7          \n\t"
 | ||
|             "pxor %%mm1, %%mm1           \n\t"
 | ||
|             "paddw %%mm3, %%mm7          \n\t"
 | ||
|             "pxor %%mm0, %%mm0           \n\t"
 | ||
|             "packuswb %%mm1, %%mm7       \n\t"
 | ||
|             "movq %%mm2, %%mm3           \n\t" // load c=Prior(x-bpp) step 1
 | ||
|             "pand " AMASK5_3_0 ", %%mm7  \n\t" // _amask5_3_0 (was _ActiveMask)
 | ||
|             "punpckhbw %%mm0, %%mm2      \n\t" // unpack High bytes of b
 | ||
|             "psllq $24, %%mm7            \n\t" // shift bytes to 2nd group of
 | ||
|                                                // 3 bytes
 | ||
|              // pav = p - a = (a + b - c) - a = b - c
 | ||
|             "movq %%mm2, %%mm4           \n\t"
 | ||
|             "paddb (%1," PCX ",), %%mm7  \n\t" // add Paeth predictor + Raw(x)
 | ||
|             "psllq $24, %%mm3            \n\t" // load c=Prior(x-bpp) step 2
 | ||
|             "movq %%mm7, (%1," PCX ",)   \n\t" // write back updated value
 | ||
|             "movq %%mm7, %%mm1           \n\t"
 | ||
|             "punpckhbw %%mm0, %%mm3      \n\t" // unpack High bytes of c
 | ||
|             "psllq $24, %%mm1            \n\t" // shift bytes (was _ShiftBpp)
 | ||
|                                     // now mm1 will be used as Raw(x-bpp)
 | ||
|             // now do Paeth for 3rd, and final, set of bytes (6-7)
 | ||
|             "pxor %%mm7, %%mm7           \n\t"
 | ||
|             "punpckhbw %%mm0, %%mm1      \n\t" // unpack High bytes of a
 | ||
|             "psubw %%mm3, %%mm4          \n\t"
 | ||
|             // pbv = p - b = (a + b - c) - b = a - c
 | ||
|             "movq %%mm1, %%mm5           \n\t"
 | ||
|             // pcv = p - c = (a + b - c) -c = (a - c) + (b - c) = pav + pbv
 | ||
|             "movq %%mm4, %%mm6           \n\t"
 | ||
|             "psubw %%mm3, %%mm5          \n\t"
 | ||
|             "pxor %%mm0, %%mm0           \n\t"
 | ||
|             "paddw %%mm5, %%mm6          \n\t"
 | ||
| 
 | ||
|             // pa = abs(p-a) = abs(pav)
 | ||
|             // pb = abs(p-b) = abs(pbv)
 | ||
|             // pc = abs(p-c) = abs(pcv)
 | ||
|             "pcmpgtw %%mm4, %%mm0        \n\t" // create mask pav bytes < 0
 | ||
|             "pcmpgtw %%mm5, %%mm7        \n\t" // create mask pbv bytes < 0
 | ||
|             "pand %%mm4, %%mm0           \n\t" // only pav bytes < 0 in mm7
 | ||
|             "pand %%mm5, %%mm7           \n\t" // only pbv bytes < 0 in mm0
 | ||
|             "psubw %%mm0, %%mm4          \n\t"
 | ||
|             "psubw %%mm7, %%mm5          \n\t"
 | ||
|             "psubw %%mm0, %%mm4          \n\t"
 | ||
|             "psubw %%mm7, %%mm5          \n\t"
 | ||
|             "pxor %%mm0, %%mm0           \n\t"
 | ||
|             "pcmpgtw %%mm6, %%mm0        \n\t" // create mask pcv bytes < 0
 | ||
|             "pand %%mm6, %%mm0           \n\t" // only pav bytes < 0 in mm7
 | ||
|             "psubw %%mm0, %%mm6          \n\t"
 | ||
|             //  test pa <= pb
 | ||
|             "movq %%mm4, %%mm7           \n\t"
 | ||
|             "psubw %%mm0, %%mm6          \n\t"
 | ||
|             "pcmpgtw %%mm5, %%mm7        \n\t" // pa > pb?
 | ||
|             "movq %%mm7, %%mm0           \n\t"
 | ||
|             // use mm0 mask copy to merge a & b
 | ||
|             "pand %%mm0, %%mm2           \n\t"
 | ||
|             // use mm7 mask to merge pa & pb
 | ||
|             "pand %%mm7, %%mm5           \n\t"
 | ||
|             "pandn %%mm1, %%mm0          \n\t"
 | ||
|             "pandn %%mm4, %%mm7          \n\t"
 | ||
|             "paddw %%mm2, %%mm0          \n\t"
 | ||
|             "paddw %%mm5, %%mm7          \n\t"
 | ||
|             //  test  ((pa <= pb)? pa:pb) <= pc
 | ||
|             "pcmpgtw %%mm6, %%mm7        \n\t" // pab > pc?
 | ||
|             "pand %%mm7, %%mm3           \n\t"
 | ||
|             "pandn %%mm0, %%mm7          \n\t"
 | ||
|             "paddw %%mm3, %%mm7          \n\t"
 | ||
|             "pxor %%mm1, %%mm1           \n\t"
 | ||
|             "packuswb %%mm7, %%mm1       \n\t"
 | ||
|             // step ecx to next set of 8 bytes and repeat loop til done
 | ||
|             "addl $8, %%ecx              \n\t"
 | ||
|             "pand " AMASK0_2_6 ", %%mm1  \n\t" // _amask0_2_6 (_ActiveMaskEnd)
 | ||
|             "paddb -8(%1," PCX ",), %%mm1 \n\t" // add Paeth predictor + Raw(x)
 | ||
|             "cmpl %%eax, %%ecx           \n\t" // MMXLength
 | ||
|             "pxor %%mm0, %%mm0           \n\t" // pxor does not affect flags
 | ||
|             "movq %%mm1, -8(%1," PCX ",) \n\t" // write back updated value
 | ||
|                                  // mm1 will be used as Raw(x-bpp) next loop
 | ||
|                            // mm3 ready to be used as Prior(x-bpp) next loop
 | ||
|             "jb paeth_3lp                \n\t"
 | ||
|             RESTORE_rbp
 | ||
| 
 | ||
|             : "=S" (dummy_value_S),            // output regs (dummy)
 | ||
|               "=D" (dummy_value_D),
 | ||
|               "=c" (dummy_value_c),
 | ||
|               "=a" (dummy_value_a)
 | ||
| 
 | ||
|             : "0" (prev_row),  // esi/rsi      // input regs
 | ||
|               "1" (row),       // edi/rdi
 | ||
|               "2" (diff),      // ecx
 | ||
|               "3" (MMXLength)  // eax
 | ||
| 
 | ||
| #if defined(CLOBBER_MMX_REGS_SUPPORTED)
 | ||
|             : "%mm0", "%mm1", "%mm2", "%mm3"   // clobber list
 | ||
|             , "%mm4", "%mm5", "%mm6", "%mm7"
 | ||
| #endif
 | ||
|          );
 | ||
|       }
 | ||
|       break;  // end 3 bpp
 | ||
| 
 | ||
|       case 4:
 | ||
|       {
 | ||
|          __asm__ __volatile__ (
 | ||
| // preload  "movl diff, %%ecx            \n\t"
 | ||
| // preload  "movl row, %1                \n\t" // edi/rdi
 | ||
| // preload  "movl prev_row, %0           \n\t" // esi/rsi
 | ||
|             "pxor %%mm0, %%mm0           \n\t"
 | ||
|             // prime the pump:  load the first Raw(x-bpp) data set
 | ||
|             "movq -8(%1," PCX ",), %%mm1 \n\t" // only time should need to read
 | ||
|                                                //  a=Raw(x-bpp) bytes
 | ||
|          "paeth_4lp:                     \n\t"
 | ||
|             // do first set of 4 bytes
 | ||
|             "movq -8(%0," PCX ",), %%mm3 \n\t" // read c=Prior(x-bpp) bytes
 | ||
|             "punpckhbw %%mm0, %%mm1      \n\t" // unpack Low bytes of a
 | ||
|             "movq (%0," PCX ",), %%mm2   \n\t" // load b=Prior(x)
 | ||
|             "punpcklbw %%mm0, %%mm2      \n\t" // unpack High bytes of b
 | ||
|             // pav = p - a = (a + b - c) - a = b - c
 | ||
|             "movq %%mm2, %%mm4           \n\t"
 | ||
|             "punpckhbw %%mm0, %%mm3      \n\t" // unpack High bytes of c
 | ||
|             // pbv = p - b = (a + b - c) - b = a - c
 | ||
|             "movq %%mm1, %%mm5           \n\t"
 | ||
|             "psubw %%mm3, %%mm4          \n\t"
 | ||
|             "pxor %%mm7, %%mm7           \n\t"
 | ||
|             // pcv = p - c = (a + b - c) -c = (a - c) + (b - c) = pav + pbv
 | ||
|             "movq %%mm4, %%mm6           \n\t"
 | ||
|             "psubw %%mm3, %%mm5          \n\t"
 | ||
|             // pa = abs(p-a) = abs(pav)
 | ||
|             // pb = abs(p-b) = abs(pbv)
 | ||
|             // pc = abs(p-c) = abs(pcv)
 | ||
|             "pcmpgtw %%mm4, %%mm0        \n\t" // create mask pav bytes < 0
 | ||
|             "paddw %%mm5, %%mm6          \n\t"
 | ||
|             "pand %%mm4, %%mm0           \n\t" // only pav bytes < 0 in mm7
 | ||
|             "pcmpgtw %%mm5, %%mm7        \n\t" // create mask pbv bytes < 0
 | ||
|             "psubw %%mm0, %%mm4          \n\t"
 | ||
|             "pand %%mm5, %%mm7           \n\t" // only pbv bytes < 0 in mm0
 | ||
|             "psubw %%mm0, %%mm4          \n\t"
 | ||
|             "psubw %%mm7, %%mm5          \n\t"
 | ||
|             "pxor %%mm0, %%mm0           \n\t"
 | ||
|             "pcmpgtw %%mm6, %%mm0        \n\t" // create mask pcv bytes < 0
 | ||
|             "pand %%mm6, %%mm0           \n\t" // only pav bytes < 0 in mm7
 | ||
|             "psubw %%mm7, %%mm5          \n\t"
 | ||
|             "psubw %%mm0, %%mm6          \n\t"
 | ||
|             //  test pa <= pb
 | ||
|             "movq %%mm4, %%mm7           \n\t"
 | ||
|             "psubw %%mm0, %%mm6          \n\t"
 | ||
|             "pcmpgtw %%mm5, %%mm7        \n\t" // pa > pb?
 | ||
|             "movq %%mm7, %%mm0           \n\t"
 | ||
|             // use mm7 mask to merge pa & pb
 | ||
|             "pand %%mm7, %%mm5           \n\t"
 | ||
|             // use mm0 mask copy to merge a & b
 | ||
|             "pand %%mm0, %%mm2           \n\t"
 | ||
|             "pandn %%mm4, %%mm7          \n\t"
 | ||
|             "pandn %%mm1, %%mm0          \n\t"
 | ||
|             "paddw %%mm5, %%mm7          \n\t"
 | ||
|             "paddw %%mm2, %%mm0          \n\t"
 | ||
|             //  test  ((pa <= pb)? pa:pb) <= pc
 | ||
|             "pcmpgtw %%mm6, %%mm7        \n\t" // pab > pc?
 | ||
|             "pxor %%mm1, %%mm1           \n\t"
 | ||
|             "pand %%mm7, %%mm3           \n\t"
 | ||
|             "pandn %%mm0, %%mm7          \n\t"
 | ||
|             "paddw %%mm3, %%mm7          \n\t"
 | ||
|             "pxor %%mm0, %%mm0           \n\t"
 | ||
|             "packuswb %%mm1, %%mm7       \n\t"
 | ||
|             "movq (%0," PCX ",), %%mm3   \n\t" // load c=Prior(x-bpp)
 | ||
|             LOAD_GOT_rbp
 | ||
|             "pand " AMASK4_4_0 ", %%mm7  \n\t" // _amask4_4_0 (was _ActiveMask)
 | ||
|             RESTORE_rbp
 | ||
|             "movq %%mm3, %%mm2           \n\t" // load b=Prior(x) step 1
 | ||
|             "paddb (%1," PCX ",), %%mm7  \n\t" // add Paeth predictor + Raw(x)
 | ||
|             "punpcklbw %%mm0, %%mm3      \n\t" // unpack High bytes of c
 | ||
|             "movq %%mm7, (%1," PCX ",)   \n\t" // write back updated value
 | ||
|             "movq %%mm7, %%mm1           \n\t" // now mm1 will be used as
 | ||
|                                                // Raw(x-bpp)
 | ||
|             // do second set of 4 bytes
 | ||
|             "punpckhbw %%mm0, %%mm2      \n\t" // unpack Low bytes of b
 | ||
|             "punpcklbw %%mm0, %%mm1      \n\t" // unpack Low bytes of a
 | ||
|             // pav = p - a = (a + b - c) - a = b - c
 | ||
|             "movq %%mm2, %%mm4           \n\t"
 | ||
|             // pbv = p - b = (a + b - c) - b = a - c
 | ||
|             "movq %%mm1, %%mm5           \n\t"
 | ||
|             "psubw %%mm3, %%mm4          \n\t"
 | ||
|             "pxor %%mm7, %%mm7           \n\t"
 | ||
|             // pcv = p - c = (a + b - c) -c = (a - c) + (b - c) = pav + pbv
 | ||
|             "movq %%mm4, %%mm6           \n\t"
 | ||
|             "psubw %%mm3, %%mm5          \n\t"
 | ||
|             // pa = abs(p-a) = abs(pav)
 | ||
|             // pb = abs(p-b) = abs(pbv)
 | ||
|             // pc = abs(p-c) = abs(pcv)
 | ||
|             "pcmpgtw %%mm4, %%mm0        \n\t" // create mask pav bytes < 0
 | ||
|             "paddw %%mm5, %%mm6          \n\t"
 | ||
|             "pand %%mm4, %%mm0           \n\t" // only pav bytes < 0 in mm7
 | ||
|             "pcmpgtw %%mm5, %%mm7        \n\t" // create mask pbv bytes < 0
 | ||
|             "psubw %%mm0, %%mm4          \n\t"
 | ||
|             "pand %%mm5, %%mm7           \n\t" // only pbv bytes < 0 in mm0
 | ||
|             "psubw %%mm0, %%mm4          \n\t"
 | ||
|             "psubw %%mm7, %%mm5          \n\t"
 | ||
|             "pxor %%mm0, %%mm0           \n\t"
 | ||
|             "pcmpgtw %%mm6, %%mm0        \n\t" // create mask pcv bytes < 0
 | ||
|             "pand %%mm6, %%mm0           \n\t" // only pav bytes < 0 in mm7
 | ||
|             "psubw %%mm7, %%mm5          \n\t"
 | ||
|             "psubw %%mm0, %%mm6          \n\t"
 | ||
|             //  test pa <= pb
 | ||
|             "movq %%mm4, %%mm7           \n\t"
 | ||
|             "psubw %%mm0, %%mm6          \n\t"
 | ||
|             "pcmpgtw %%mm5, %%mm7        \n\t" // pa > pb?
 | ||
|             "movq %%mm7, %%mm0           \n\t"
 | ||
|             // use mm7 mask to merge pa & pb
 | ||
|             "pand %%mm7, %%mm5           \n\t"
 | ||
|             // use mm0 mask copy to merge a & b
 | ||
|             "pand %%mm0, %%mm2           \n\t"
 | ||
|             "pandn %%mm4, %%mm7          \n\t"
 | ||
|             "pandn %%mm1, %%mm0          \n\t"
 | ||
|             "paddw %%mm5, %%mm7          \n\t"
 | ||
|             "paddw %%mm2, %%mm0          \n\t"
 | ||
|             //  test  ((pa <= pb)? pa:pb) <= pc
 | ||
|             "pcmpgtw %%mm6, %%mm7        \n\t" // pab > pc?
 | ||
|             "pxor %%mm1, %%mm1           \n\t"
 | ||
|             "pand %%mm7, %%mm3           \n\t"
 | ||
|             "pandn %%mm0, %%mm7          \n\t"
 | ||
|             "pxor %%mm1, %%mm1           \n\t"
 | ||
|             "paddw %%mm3, %%mm7          \n\t"
 | ||
|             "pxor %%mm0, %%mm0           \n\t"
 | ||
|             // step ecx to next set of 8 bytes and repeat loop til done
 | ||
|             "addl $8, %%ecx              \n\t"
 | ||
|             "packuswb %%mm7, %%mm1       \n\t"
 | ||
|             "paddb -8(%1," PCX ",), %%mm1 \n\t" // add predictor with Raw(x)
 | ||
|             "cmpl %%eax, %%ecx           \n\t" // MMXLength
 | ||
|             "movq %%mm1, -8(%1," PCX ",) \n\t" // write back updated value
 | ||
|                                  // mm1 will be used as Raw(x-bpp) next loop
 | ||
|             "jb paeth_4lp                \n\t"
 | ||
| 
 | ||
|             : "=S" (dummy_value_S),            // output regs (dummy)
 | ||
|               "=D" (dummy_value_D),
 | ||
|               "=c" (dummy_value_c),
 | ||
|               "=a" (dummy_value_a)
 | ||
| 
 | ||
|             : "0" (prev_row),  // esi/rsi      // input regs
 | ||
|               "1" (row),       // edi/rdi
 | ||
|               "2" (diff),      // ecx
 | ||
|               "3" (MMXLength)  // eax
 | ||
| 
 | ||
| #if defined(CLOBBER_MMX_REGS_SUPPORTED)
 | ||
|             : "%mm0", "%mm1", "%mm2", "%mm3"   // clobber list
 | ||
|             , "%mm4", "%mm5", "%mm6", "%mm7"
 | ||
| #endif
 | ||
|          );
 | ||
|       }
 | ||
|       break;  // end 4 bpp
 | ||
| 
 | ||
|       case 1:
 | ||
|       case 2:
 | ||
|       {
 | ||
|          __asm__ __volatile__ (
 | ||
| // preload  "movl diff, %%eax            \n\t" // eax: x = offset to align. bdry
 | ||
| // preload  "movl FullLength, %%edx      \n\t"
 | ||
|             "cmpl %%edx, %%eax           \n\t"
 | ||
|             "jnb paeth_dend              \n\t"
 | ||
| 
 | ||
|             SAVE_ebp
 | ||
| 
 | ||
| // preload  "movl row, %2                \n\t" // edi/rdi
 | ||
|             // do Paeth decode for remaining bytes
 | ||
| // preload  "movl prev_row, %1           \n\t" // esi/rsi
 | ||
|             "movl %%eax, %%ebp           \n\t"
 | ||
| // preload  "subl bpp, %%ebp             \n\t" // (bpp is preloaded into ecx)
 | ||
|             "subl %%ecx, %%ebp           \n\t" // ebp = eax - bpp
 | ||
|             "xorl %%ecx, %%ecx           \n\t" // zero ecx before using cl & cx
 | ||
| 
 | ||
|             SAVE_GOT_ebx
 | ||
|             SAVE_r11_r12_r13
 | ||
| 
 | ||
|          "paeth_dlp:                     \n\t"
 | ||
|             "xorl %%ebx, %%ebx           \n\t"
 | ||
|             // pav = p - a = (a + b - c) - a = b - c
 | ||
|             "movb (%1," PAX ",), %%bl    \n\t" // load Prior(x) into bl
 | ||
|             "movb (%1," PBP ",), %%cl    \n\t" // load Prior(x-bpp) into cl
 | ||
|             "subl %%ecx, %%ebx           \n\t" // subtract Prior(x-bpp)
 | ||
|             "movl %%ebx, " pa_TEMP "     \n\t" // Save pav for later use
 | ||
|             "xorl %%ebx, %%ebx           \n\t"
 | ||
|             // pbv = p - b = (a + b - c) - b = a - c
 | ||
|             "movb (%2," PBP ",), %%bl    \n\t" // load Raw(x-bpp) into bl
 | ||
|             "subl %%ecx, %%ebx           \n\t" // subtract Prior(x-bpp)
 | ||
|             "movl %%ebx, %%ecx           \n\t"
 | ||
|             // pcv = p - c = (a + b - c) -c = (a - c) + (b - c) = pav + pbv
 | ||
|             "addl " pa_TEMP ", %%ebx     \n\t" // pcv = pav + pbv
 | ||
|             // pc = abs(pcv)
 | ||
|             "testl $0x80000000, %%ebx    \n\t"
 | ||
|             "jz paeth_dpca               \n\t"
 | ||
|             "negl %%ebx                  \n\t" // reverse sign of neg values
 | ||
| 
 | ||
|          "paeth_dpca:                    \n\t"
 | ||
|             "movl %%ebx, " pc_TEMP "     \n\t" // save pc for later use
 | ||
|             // pb = abs(pbv)
 | ||
|             "testl $0x80000000, %%ecx    \n\t"
 | ||
|             "jz paeth_dpba               \n\t"
 | ||
|             "negl %%ecx                  \n\t" // reverse sign of neg values
 | ||
| 
 | ||
|          "paeth_dpba:                    \n\t"
 | ||
|             "movl %%ecx, " pb_TEMP "     \n\t" // save pb for later use
 | ||
|             // pa = abs(pav)
 | ||
|             "movl " pa_TEMP ", %%ebx     \n\t"
 | ||
|             "testl $0x80000000, %%ebx    \n\t"
 | ||
|             "jz paeth_dpaa               \n\t"
 | ||
|             "negl %%ebx                  \n\t" // reverse sign of neg values
 | ||
| 
 | ||
|          "paeth_dpaa:                    \n\t"
 | ||
|             "movl %%ebx, " pa_TEMP "     \n\t" // save pa for later use
 | ||
|             // test if pa <= pb
 | ||
|             "cmpl %%ecx, %%ebx           \n\t"
 | ||
|             "jna paeth_dabb              \n\t"
 | ||
|             // pa > pb; now test if pb <= pc
 | ||
|             "cmpl " pc_TEMP ", %%ecx     \n\t"
 | ||
|             "jna paeth_dbbc              \n\t"
 | ||
|             // pb > pc; Raw(x) = Paeth(x) + Prior(x-bpp)
 | ||
|             "movb (%1," PBP ",), %%cl    \n\t" // load Prior(x-bpp) into cl
 | ||
|             "jmp paeth_dpaeth            \n\t"
 | ||
| 
 | ||
|          "paeth_dbbc:                    \n\t"
 | ||
|             // pb <= pc; Raw(x) = Paeth(x) + Prior(x)
 | ||
|             "movb (%1," PAX ",), %%cl    \n\t" // load Prior(x) into cl
 | ||
|             "jmp paeth_dpaeth            \n\t"
 | ||
| 
 | ||
|          "paeth_dabb:                    \n\t"
 | ||
|             // pa <= pb; now test if pa <= pc
 | ||
|             "cmpl " pc_TEMP ", %%ebx     \n\t"
 | ||
|             "jna paeth_dabc              \n\t"
 | ||
|             // pa > pc; Raw(x) = Paeth(x) + Prior(x-bpp)
 | ||
|             "movb (%1," PBP ",), %%cl   \n\t" // load Prior(x-bpp) into cl
 | ||
|             "jmp paeth_dpaeth            \n\t"
 | ||
| 
 | ||
|          "paeth_dabc:                    \n\t"
 | ||
|             // pa <= pc; Raw(x) = Paeth(x) + Raw(x-bpp)
 | ||
|             "movb (%2," PBP ",), %%cl    \n\t" // load Raw(x-bpp) into cl
 | ||
| 
 | ||
|          "paeth_dpaeth:                  \n\t"
 | ||
|             "incl %%eax                  \n\t"
 | ||
|             "incl %%ebp                  \n\t"
 | ||
|             // Raw(x) = (Paeth(x) + Paeth_Predictor( a, b, c )) mod 256
 | ||
|             "addb %%cl, -1(%2," PAX ",)  \n\t"
 | ||
|             "cmpl %%edx, %%eax           \n\t" // check against FullLength
 | ||
|             "jb paeth_dlp                \n\t"
 | ||
| 
 | ||
|             RESTORE_r11_r12_r13
 | ||
|             RESTORE_GOT_ebx
 | ||
|             RESTORE_ebp
 | ||
| 
 | ||
|          "paeth_dend:                    \n\t"
 | ||
| 
 | ||
|             : "=c" (dummy_value_c),            // output regs (dummy)
 | ||
|               "=S" (dummy_value_S),
 | ||
|               "=D" (dummy_value_D),
 | ||
|               "=a" (dummy_value_a),
 | ||
|               "=d" (dummy_value_d)
 | ||
| 
 | ||
|             : "0" (bpp),         // ecx        // input regs
 | ||
|               "1" (prev_row),    // esi/rsi
 | ||
|               "2" (row),         // edi/rdi
 | ||
|               "3" (diff),        // eax
 | ||
|               "4" (FullLength)   // edx
 | ||
| 
 | ||
|             CLOB_COLON_ebx_ebp_r1X             // clobber list
 | ||
|               CLOBBER_GOT_ebx
 | ||
|               CLOB_COMMA_ebx_ebp
 | ||
|               CLOBBER_ebp
 | ||
|               CLOB_COMMA_ebX_r1X
 | ||
|               CLOBBER_r11_r12_r13
 | ||
|          );
 | ||
|       }
 | ||
|       return; // end 1 or 2 bpp (no need to go further with this one)
 | ||
| 
 | ||
|       case 6:
 | ||
|       {
 | ||
| //       _ActiveMask2 = 0xffffffff00000000LL;  // NOT USED ("_amask_0_4_4")
 | ||
| //       _ShiftBpp = 48;       // bpp << 3 == bpp * 8
 | ||
| //       _ShiftRem = 16;       // 64 - _ShiftBpp
 | ||
| 
 | ||
|          __asm__ __volatile__ (
 | ||
| // preload  "movl diff, %%ecx            \n\t"
 | ||
| // preload  "movl row, %1                \n\t" // edi/rdi
 | ||
| // preload  "movl prev_row, %0           \n\t" // esi/rsi
 | ||
|             // prime the pump:  load the first Raw(x-bpp) data set
 | ||
|             "movq -8(%1," PCX ",), %%mm1 \n\t"
 | ||
|             "pxor %%mm0, %%mm0           \n\t"
 | ||
| 
 | ||
|          "paeth_6lp:                     \n\t"
 | ||
|             // must shift to position Raw(x-bpp) data
 | ||
|             "psrlq $16, %%mm1            \n\t" // was _ShiftRem
 | ||
|             // do first set of 4 bytes
 | ||
|             "movq -8(%0," PCX ",), %%mm3 \n\t" // read c=Prior(x-bpp) bytes
 | ||
|             "punpcklbw %%mm0, %%mm1      \n\t" // unpack Low bytes of a
 | ||
|             "movq (%0," PCX ",), %%mm2   \n\t" // load b=Prior(x)
 | ||
|             "punpcklbw %%mm0, %%mm2      \n\t" // unpack Low bytes of b
 | ||
|             // must shift to position Prior(x-bpp) data
 | ||
|             "psrlq $16, %%mm3            \n\t" // was _ShiftRem
 | ||
|             // pav = p - a = (a + b - c) - a = b - c
 | ||
|             "movq %%mm2, %%mm4           \n\t"
 | ||
|             "punpcklbw %%mm0, %%mm3      \n\t" // unpack Low bytes of c
 | ||
|             // pbv = p - b = (a + b - c) - b = a - c
 | ||
|             "movq %%mm1, %%mm5           \n\t"
 | ||
|             "psubw %%mm3, %%mm4          \n\t"
 | ||
|             "pxor %%mm7, %%mm7           \n\t"
 | ||
|             // pcv = p - c = (a + b - c) -c = (a - c) + (b - c) = pav + pbv
 | ||
|             "movq %%mm4, %%mm6           \n\t"
 | ||
|             "psubw %%mm3, %%mm5          \n\t"
 | ||
|             // pa = abs(p-a) = abs(pav)
 | ||
|             // pb = abs(p-b) = abs(pbv)
 | ||
|             // pc = abs(p-c) = abs(pcv)
 | ||
|             "pcmpgtw %%mm4, %%mm0        \n\t" // create mask pav bytes < 0
 | ||
|             "paddw %%mm5, %%mm6          \n\t"
 | ||
|             "pand %%mm4, %%mm0           \n\t" // only pav bytes < 0 in mm7
 | ||
|             "pcmpgtw %%mm5, %%mm7        \n\t" // create mask pbv bytes < 0
 | ||
|             "psubw %%mm0, %%mm4          \n\t"
 | ||
|             "pand %%mm5, %%mm7           \n\t" // only pbv bytes < 0 in mm0
 | ||
|             "psubw %%mm0, %%mm4          \n\t"
 | ||
|             "psubw %%mm7, %%mm5          \n\t"
 | ||
|             "pxor %%mm0, %%mm0           \n\t"
 | ||
|             "pcmpgtw %%mm6, %%mm0        \n\t" // create mask pcv bytes < 0
 | ||
|             "pand %%mm6, %%mm0           \n\t" // only pav bytes < 0 in mm7
 | ||
|             "psubw %%mm7, %%mm5          \n\t"
 | ||
|             "psubw %%mm0, %%mm6          \n\t"
 | ||
|             //  test pa <= pb
 | ||
|             "movq %%mm4, %%mm7           \n\t"
 | ||
|             "psubw %%mm0, %%mm6          \n\t"
 | ||
|             "pcmpgtw %%mm5, %%mm7        \n\t" // pa > pb?
 | ||
|             "movq %%mm7, %%mm0           \n\t"
 | ||
|             // use mm7 mask to merge pa & pb
 | ||
|             "pand %%mm7, %%mm5           \n\t"
 | ||
|             // use mm0 mask copy to merge a & b
 | ||
|             "pand %%mm0, %%mm2           \n\t"
 | ||
|             "pandn %%mm4, %%mm7          \n\t"
 | ||
|             "pandn %%mm1, %%mm0          \n\t"
 | ||
|             "paddw %%mm5, %%mm7          \n\t"
 | ||
|             "paddw %%mm2, %%mm0          \n\t"
 | ||
|             //  test  ((pa <= pb)? pa:pb) <= pc
 | ||
|             "pcmpgtw %%mm6, %%mm7        \n\t" // pab > pc?
 | ||
|             "pxor %%mm1, %%mm1           \n\t"
 | ||
|             "pand %%mm7, %%mm3           \n\t"
 | ||
|             "pandn %%mm0, %%mm7          \n\t"
 | ||
|             "paddw %%mm3, %%mm7          \n\t"
 | ||
|             "pxor %%mm0, %%mm0           \n\t"
 | ||
|             "packuswb %%mm1, %%mm7       \n\t"
 | ||
|             "movq -8(%0," PCX ",), %%mm3 \n\t" // load c=Prior(x-bpp)
 | ||
|             LOAD_GOT_rbp
 | ||
|             "pand " AMASK4_4_0 ", %%mm7  \n\t" // _amask4_4_0 (was _ActiveMask)
 | ||
|             RESTORE_rbp
 | ||
|             "psrlq $16, %%mm3            \n\t"
 | ||
|             "movq (%0," PCX ",), %%mm2   \n\t" // load b=Prior(x) step 1
 | ||
|             "paddb (%1," PCX ",), %%mm7  \n\t" // add Paeth predictor + Raw(x)
 | ||
|             "movq %%mm2, %%mm6           \n\t"
 | ||
|             "movq %%mm7, (%1," PCX ",)   \n\t" // write back updated value
 | ||
|             "movq -8(%1," PCX ",), %%mm1 \n\t"
 | ||
|             "psllq $48, %%mm6            \n\t" // bpp * 8 = bits per pixel
 | ||
|             "movq %%mm7, %%mm5           \n\t"
 | ||
|             "psrlq $16, %%mm1            \n\t" // 64 - (bpp * 8) = remainder
 | ||
|             "por %%mm6, %%mm3            \n\t"
 | ||
|             "psllq $48, %%mm5            \n\t" // was _ShiftBpp
 | ||
|             "punpckhbw %%mm0, %%mm3      \n\t" // unpack High bytes of c
 | ||
|             "por %%mm5, %%mm1            \n\t"
 | ||
|             // do second set of 4 bytes
 | ||
|             "punpckhbw %%mm0, %%mm2      \n\t" // unpack High bytes of b
 | ||
|             "punpckhbw %%mm0, %%mm1      \n\t" // unpack High bytes of a
 | ||
|             // pav = p - a = (a + b - c) - a = b - c
 | ||
|             "movq %%mm2, %%mm4           \n\t"
 | ||
|             // pbv = p - b = (a + b - c) - b = a - c
 | ||
|             "movq %%mm1, %%mm5           \n\t"
 | ||
|             "psubw %%mm3, %%mm4          \n\t"
 | ||
|             "pxor %%mm7, %%mm7           \n\t"
 | ||
|             // pcv = p - c = (a + b - c) -c = (a - c) + (b - c) = pav + pbv
 | ||
|             "movq %%mm4, %%mm6           \n\t"
 | ||
|             "psubw %%mm3, %%mm5          \n\t"
 | ||
|             // pa = abs(p-a) = abs(pav)
 | ||
|             // pb = abs(p-b) = abs(pbv)
 | ||
|             // pc = abs(p-c) = abs(pcv)
 | ||
|             "pcmpgtw %%mm4, %%mm0        \n\t" // create mask pav bytes < 0
 | ||
|             "paddw %%mm5, %%mm6          \n\t"
 | ||
|             "pand %%mm4, %%mm0           \n\t" // only pav bytes < 0 in mm7
 | ||
|             "pcmpgtw %%mm5, %%mm7        \n\t" // create mask pbv bytes < 0
 | ||
|             "psubw %%mm0, %%mm4          \n\t"
 | ||
|             "pand %%mm5, %%mm7           \n\t" // only pbv bytes < 0 in mm0
 | ||
|             "psubw %%mm0, %%mm4          \n\t"
 | ||
|             "psubw %%mm7, %%mm5          \n\t"
 | ||
|             "pxor %%mm0, %%mm0           \n\t"
 | ||
|             "pcmpgtw %%mm6, %%mm0        \n\t" // create mask pcv bytes < 0
 | ||
|             "pand %%mm6, %%mm0           \n\t" // only pav bytes < 0 in mm7
 | ||
|             "psubw %%mm7, %%mm5          \n\t"
 | ||
|             "psubw %%mm0, %%mm6          \n\t"
 | ||
|             //  test pa <= pb
 | ||
|             "movq %%mm4, %%mm7           \n\t"
 | ||
|             "psubw %%mm0, %%mm6          \n\t"
 | ||
|             "pcmpgtw %%mm5, %%mm7        \n\t" // pa > pb?
 | ||
|             "movq %%mm7, %%mm0           \n\t"
 | ||
|             // use mm7 mask to merge pa & pb
 | ||
|             "pand %%mm7, %%mm5           \n\t"
 | ||
|             // use mm0 mask copy to merge a & b
 | ||
|             "pand %%mm0, %%mm2           \n\t"
 | ||
|             "pandn %%mm4, %%mm7          \n\t"
 | ||
|             "pandn %%mm1, %%mm0          \n\t"
 | ||
|             "paddw %%mm5, %%mm7          \n\t"
 | ||
|             "paddw %%mm2, %%mm0          \n\t"
 | ||
|             //  test  ((pa <= pb)? pa:pb) <= pc
 | ||
|             "pcmpgtw %%mm6, %%mm7        \n\t" // pab > pc?
 | ||
|             "pxor %%mm1, %%mm1           \n\t"
 | ||
|             "pand %%mm7, %%mm3           \n\t"
 | ||
|             "pandn %%mm0, %%mm7          \n\t"
 | ||
|             "pxor %%mm1, %%mm1           \n\t"
 | ||
|             "paddw %%mm3, %%mm7          \n\t"
 | ||
|             "pxor %%mm0, %%mm0           \n\t"
 | ||
|             // step ecx to next set of 8 bytes and repeat loop til done
 | ||
|             "addl $8, %%ecx              \n\t"
 | ||
|             "packuswb %%mm7, %%mm1       \n\t"
 | ||
|             "paddb -8(%1," PCX ",), %%mm1 \n\t" // add Paeth predictor + Raw(x)
 | ||
|             "cmpl %%eax, %%ecx           \n\t" // MMXLength
 | ||
|             "movq %%mm1, -8(%1," PCX ",) \n\t" // write back updated value
 | ||
|                                  // mm1 will be used as Raw(x-bpp) next loop
 | ||
|             "jb paeth_6lp                \n\t"
 | ||
| 
 | ||
|             : "=S" (dummy_value_S),            // output regs (dummy)
 | ||
|               "=D" (dummy_value_D),
 | ||
|               "=c" (dummy_value_c),
 | ||
|               "=a" (dummy_value_a)
 | ||
| 
 | ||
|             : "0" (prev_row),  // esi/rsi      // input regs
 | ||
|               "1" (row),       // edi/rdi
 | ||
|               "2" (diff),      // ecx
 | ||
|               "3" (MMXLength)  // eax
 | ||
| 
 | ||
| #if defined(CLOBBER_MMX_REGS_SUPPORTED)
 | ||
|             : "%mm0", "%mm1", "%mm2", "%mm3"   // clobber list
 | ||
|             , "%mm4", "%mm5", "%mm6", "%mm7"
 | ||
| #endif
 | ||
|          );
 | ||
|       }
 | ||
|       break;  // end 6 bpp
 | ||
| 
 | ||
|       case 8:                          // bpp == 8
 | ||
|       {
 | ||
|          __asm__ __volatile__ (
 | ||
| // preload  "movl diff, %%ecx            \n\t"
 | ||
| // preload  "movl row, %1                \n\t" // edi/rdi
 | ||
| // preload  "movl prev_row, %0           \n\t" // esi/rsi
 | ||
|             "pxor %%mm0, %%mm0           \n\t"
 | ||
|             // prime the pump:  load the first Raw(x-bpp) data set
 | ||
|             "movq -8(%1," PCX ",), %%mm1 \n\t" // only time should need to read
 | ||
|                                                //  a=Raw(x-bpp) bytes
 | ||
|          "paeth_8lp:                     \n\t"
 | ||
|             // do first set of 4 bytes
 | ||
|             "movq -8(%0," PCX ",), %%mm3 \n\t" // read c=Prior(x-bpp) bytes
 | ||
|             "punpcklbw %%mm0, %%mm1      \n\t" // unpack Low bytes of a
 | ||
|             "movq (%0," PCX ",), %%mm2   \n\t" // load b=Prior(x)
 | ||
|             "punpcklbw %%mm0, %%mm2      \n\t" // unpack Low bytes of b
 | ||
|             // pav = p - a = (a + b - c) - a = b - c
 | ||
|             "movq %%mm2, %%mm4           \n\t"
 | ||
|             "punpcklbw %%mm0, %%mm3      \n\t" // unpack Low bytes of c
 | ||
|             // pbv = p - b = (a + b - c) - b = a - c
 | ||
|             "movq %%mm1, %%mm5           \n\t"
 | ||
|             "psubw %%mm3, %%mm4          \n\t"
 | ||
|             "pxor %%mm7, %%mm7           \n\t"
 | ||
|             // pcv = p - c = (a + b - c) -c = (a - c) + (b - c) = pav + pbv
 | ||
|             "movq %%mm4, %%mm6           \n\t"
 | ||
|             "psubw %%mm3, %%mm5          \n\t"
 | ||
|             // pa = abs(p-a) = abs(pav)
 | ||
|             // pb = abs(p-b) = abs(pbv)
 | ||
|             // pc = abs(p-c) = abs(pcv)
 | ||
|             "pcmpgtw %%mm4, %%mm0        \n\t" // create mask pav bytes < 0
 | ||
|             "paddw %%mm5, %%mm6          \n\t"
 | ||
|             "pand %%mm4, %%mm0           \n\t" // only pav bytes < 0 in mm7
 | ||
|             "pcmpgtw %%mm5, %%mm7        \n\t" // create mask pbv bytes < 0
 | ||
|             "psubw %%mm0, %%mm4          \n\t"
 | ||
|             "pand %%mm5, %%mm7           \n\t" // only pbv bytes < 0 in mm0
 | ||
|             "psubw %%mm0, %%mm4          \n\t"
 | ||
|             "psubw %%mm7, %%mm5          \n\t"
 | ||
|             "pxor %%mm0, %%mm0           \n\t"
 | ||
|             "pcmpgtw %%mm6, %%mm0        \n\t" // create mask pcv bytes < 0
 | ||
|             "pand %%mm6, %%mm0           \n\t" // only pav bytes < 0 in mm7
 | ||
|             "psubw %%mm7, %%mm5          \n\t"
 | ||
|             "psubw %%mm0, %%mm6          \n\t"
 | ||
|             //  test pa <= pb
 | ||
|             "movq %%mm4, %%mm7           \n\t"
 | ||
|             "psubw %%mm0, %%mm6          \n\t"
 | ||
|             "pcmpgtw %%mm5, %%mm7        \n\t" // pa > pb?
 | ||
|             "movq %%mm7, %%mm0           \n\t"
 | ||
|             // use mm7 mask to merge pa & pb
 | ||
|             "pand %%mm7, %%mm5           \n\t"
 | ||
|             // use mm0 mask copy to merge a & b
 | ||
|             "pand %%mm0, %%mm2           \n\t"
 | ||
|             "pandn %%mm4, %%mm7          \n\t"
 | ||
|             "pandn %%mm1, %%mm0          \n\t"
 | ||
|             "paddw %%mm5, %%mm7          \n\t"
 | ||
|             "paddw %%mm2, %%mm0          \n\t"
 | ||
|             //  test  ((pa <= pb)? pa:pb) <= pc
 | ||
|             "pcmpgtw %%mm6, %%mm7        \n\t" // pab > pc?
 | ||
|             "pxor %%mm1, %%mm1           \n\t"
 | ||
|             "pand %%mm7, %%mm3           \n\t"
 | ||
|             "pandn %%mm0, %%mm7          \n\t"
 | ||
|             "paddw %%mm3, %%mm7          \n\t"
 | ||
|             "pxor %%mm0, %%mm0           \n\t"
 | ||
|             "packuswb %%mm1, %%mm7       \n\t"
 | ||
|             "movq -8(%0," PCX ",), %%mm3 \n\t" // read c=Prior(x-bpp) bytes
 | ||
|             LOAD_GOT_rbp
 | ||
|             "pand " AMASK4_4_0 ", %%mm7  \n\t" // _amask4_4_0 (was _ActiveMask)
 | ||
|             RESTORE_rbp
 | ||
|             "movq (%0," PCX ",), %%mm2   \n\t" // load b=Prior(x)
 | ||
|             "paddb (%1," PCX ",), %%mm7  \n\t" // add Paeth predictor + Raw(x)
 | ||
|             "punpckhbw %%mm0, %%mm3      \n\t" // unpack High bytes of c
 | ||
|             "movq %%mm7, (%1," PCX ",)   \n\t" // write back updated value
 | ||
|             "movq -8(%1," PCX ",), %%mm1 \n\t" // read a=Raw(x-bpp) bytes
 | ||
| 
 | ||
|             // do second set of 4 bytes
 | ||
|             "punpckhbw %%mm0, %%mm2      \n\t" // unpack High bytes of b
 | ||
|             "punpckhbw %%mm0, %%mm1      \n\t" // unpack High bytes of a
 | ||
|             // pav = p - a = (a + b - c) - a = b - c
 | ||
|             "movq %%mm2, %%mm4           \n\t"
 | ||
|             // pbv = p - b = (a + b - c) - b = a - c
 | ||
|             "movq %%mm1, %%mm5           \n\t"
 | ||
|             "psubw %%mm3, %%mm4          \n\t"
 | ||
|             "pxor %%mm7, %%mm7           \n\t"
 | ||
|             // pcv = p - c = (a + b - c) -c = (a - c) + (b - c) = pav + pbv
 | ||
|             "movq %%mm4, %%mm6           \n\t"
 | ||
|             "psubw %%mm3, %%mm5          \n\t"
 | ||
|             // pa = abs(p-a) = abs(pav)
 | ||
|             // pb = abs(p-b) = abs(pbv)
 | ||
|             // pc = abs(p-c) = abs(pcv)
 | ||
|             "pcmpgtw %%mm4, %%mm0        \n\t" // create mask pav bytes < 0
 | ||
|             "paddw %%mm5, %%mm6          \n\t"
 | ||
|             "pand %%mm4, %%mm0           \n\t" // only pav bytes < 0 in mm7
 | ||
|             "pcmpgtw %%mm5, %%mm7        \n\t" // create mask pbv bytes < 0
 | ||
|             "psubw %%mm0, %%mm4          \n\t"
 | ||
|             "pand %%mm5, %%mm7           \n\t" // only pbv bytes < 0 in mm0
 | ||
|             "psubw %%mm0, %%mm4          \n\t"
 | ||
|             "psubw %%mm7, %%mm5          \n\t"
 | ||
|             "pxor %%mm0, %%mm0           \n\t"
 | ||
|             "pcmpgtw %%mm6, %%mm0        \n\t" // create mask pcv bytes < 0
 | ||
|             "pand %%mm6, %%mm0           \n\t" // only pav bytes < 0 in mm7
 | ||
|             "psubw %%mm7, %%mm5          \n\t"
 | ||
|             "psubw %%mm0, %%mm6          \n\t"
 | ||
|             //  test pa <= pb
 | ||
|             "movq %%mm4, %%mm7           \n\t"
 | ||
|             "psubw %%mm0, %%mm6          \n\t"
 | ||
|             "pcmpgtw %%mm5, %%mm7        \n\t" // pa > pb?
 | ||
|             "movq %%mm7, %%mm0           \n\t"
 | ||
|             // use mm7 mask to merge pa & pb
 | ||
|             "pand %%mm7, %%mm5           \n\t"
 | ||
|             // use mm0 mask copy to merge a & b
 | ||
|             "pand %%mm0, %%mm2           \n\t"
 | ||
|             "pandn %%mm4, %%mm7          \n\t"
 | ||
|             "pandn %%mm1, %%mm0          \n\t"
 | ||
|             "paddw %%mm5, %%mm7          \n\t"
 | ||
|             "paddw %%mm2, %%mm0          \n\t"
 | ||
|             //  test  ((pa <= pb)? pa:pb) <= pc
 | ||
|             "pcmpgtw %%mm6, %%mm7        \n\t" // pab > pc?
 | ||
|             "pxor %%mm1, %%mm1           \n\t"
 | ||
|             "pand %%mm7, %%mm3           \n\t"
 | ||
|             "pandn %%mm0, %%mm7          \n\t"
 | ||
|             "pxor %%mm1, %%mm1           \n\t"
 | ||
|             "paddw %%mm3, %%mm7          \n\t"
 | ||
|             "pxor %%mm0, %%mm0           \n\t"
 | ||
|             // step ecx to next set of 8 bytes and repeat loop til done
 | ||
|             "addl $8, %%ecx              \n\t"
 | ||
|             "packuswb %%mm7, %%mm1       \n\t"
 | ||
|             "paddb -8(%1," PCX ",), %%mm1 \n\t" // add Paeth predictor + Raw(x)
 | ||
|             "cmpl %%eax, %%ecx           \n\t" // MMXLength
 | ||
|             "movq %%mm1, -8(%1," PCX ",) \n\t" // write back updated value
 | ||
|                                  // mm1 will be used as Raw(x-bpp) next loop
 | ||
|             "jb paeth_8lp                \n\t"
 | ||
| 
 | ||
|             : "=S" (dummy_value_S),            // output regs (dummy)
 | ||
|               "=D" (dummy_value_D),
 | ||
|               "=c" (dummy_value_c),
 | ||
|               "=a" (dummy_value_a)
 | ||
| 
 | ||
|             : "0" (prev_row),  // esi/rsi      // input regs
 | ||
|               "1" (row),       // edi/rdi
 | ||
|               "2" (diff),      // ecx
 | ||
|               "3" (MMXLength)  // eax
 | ||
| 
 | ||
| #if defined(CLOBBER_MMX_REGS_SUPPORTED)
 | ||
|             : "%mm0", "%mm1", "%mm2", "%mm3"   // clobber list
 | ||
|             , "%mm4", "%mm5", "%mm6", "%mm7"
 | ||
| #endif
 | ||
|          );
 | ||
|       }
 | ||
|       break;  // end 8 bpp
 | ||
| 
 | ||
|       default:                // bpp != 1,2,3,4,6,8:  doesn't exist
 | ||
|       {
 | ||
|          // ERROR:  SHOULD NEVER BE REACHED
 | ||
| #if defined(PNG_DEBUG)
 | ||
|          png_debug(1, "Internal libpng logic error (GCC "
 | ||
|            "png_read_filter_row_mmx_paeth())\n");
 | ||
| #endif
 | ||
|       }
 | ||
|       break;
 | ||
| 
 | ||
|    } // end switch (bpp)
 | ||
| 
 | ||
|    __asm__ __volatile__ (
 | ||
|       // MMX acceleration complete; now do clean-up
 | ||
|       // check if any remaining bytes left to decode
 | ||
| //pre "movl FullLength, %%edx      \n\t"
 | ||
| //pre "movl MMXLength, %%eax       \n\t"
 | ||
|       "cmpl %%edx, %%eax           \n\t"
 | ||
|       "jnb paeth_end               \n\t"
 | ||
| 
 | ||
|       SAVE_ebp
 | ||
| 
 | ||
| //pre "movl row, %2                \n\t" // edi/rdi
 | ||
| //pre "movl prev_row, %1           \n\t" // esi/rsi
 | ||
|       // do Paeth decode for remaining bytes
 | ||
|       "movl %%eax, %%ebp           \n\t"
 | ||
| //pre "subl bpp, %%ebp             \n\t" // (bpp is preloaded into ecx)
 | ||
|       "subl %%ecx, %%ebp           \n\t" // ebp = eax - bpp
 | ||
|       "xorl %%ecx, %%ecx           \n\t" // zero ecx before using cl & cx below
 | ||
| 
 | ||
|       SAVE_GOT_ebx
 | ||
|       SAVE_r11_r12_r13
 | ||
| 
 | ||
|    "paeth_lp2:                     \n\t"
 | ||
|       "xorl %%ebx, %%ebx           \n\t"
 | ||
|       // pav = p - a = (a + b - c) - a = b - c
 | ||
|       "movb (%1," PAX ",), %%bl    \n\t" // load Prior(x) into bl
 | ||
|       "movb (%1," PBP ",), %%cl    \n\t" // load Prior(x-bpp) into cl
 | ||
|       "subl %%ecx, %%ebx           \n\t" // subtract Prior(x-bpp)
 | ||
|       "movl %%ebx, " pa_TEMP "     \n\t" // Save pav for later use
 | ||
|       "xorl %%ebx, %%ebx           \n\t"
 | ||
|       // pbv = p - b = (a + b - c) - b = a - c
 | ||
|       "movb (%2," PBP ",), %%bl    \n\t" // load Raw(x-bpp) into bl
 | ||
|       "subl %%ecx, %%ebx           \n\t" // subtract Prior(x-bpp)
 | ||
|       "movl %%ebx, %%ecx           \n\t"
 | ||
|       // pcv = p - c = (a + b - c) - c = (a - c) + (b - c) = pav + pbv
 | ||
|       "addl " pa_TEMP ", %%ebx     \n\t" // pcv = pav + pbv
 | ||
|       // pc = abs(pcv)
 | ||
|       "testl $0x80000000, %%ebx    \n\t"
 | ||
|       "jz paeth_pca2               \n\t"
 | ||
|       "negl %%ebx                  \n\t" // reverse sign of neg values
 | ||
| 
 | ||
|    "paeth_pca2:                    \n\t"
 | ||
|       "movl %%ebx, " pc_TEMP "     \n\t" // save pc for later use
 | ||
|       // pb = abs(pbv)
 | ||
|       "testl $0x80000000, %%ecx    \n\t"
 | ||
|       "jz paeth_pba2               \n\t"
 | ||
|       "negl %%ecx                  \n\t" // reverse sign of neg values
 | ||
| 
 | ||
|    "paeth_pba2:                    \n\t"
 | ||
|       "movl %%ecx, " pb_TEMP "     \n\t" // save pb for later use
 | ||
|       // pa = abs(pav)
 | ||
|       "movl " pa_TEMP ", %%ebx     \n\t"
 | ||
|       "testl $0x80000000, %%ebx    \n\t"
 | ||
|       "jz paeth_paa2               \n\t"
 | ||
|       "negl %%ebx                  \n\t" // reverse sign of neg values
 | ||
| 
 | ||
|    "paeth_paa2:                    \n\t"
 | ||
|       "movl %%ebx, " pa_TEMP "     \n\t" // save pa for later use
 | ||
|       // test if pa <= pb
 | ||
|       "cmpl %%ecx, %%ebx           \n\t"
 | ||
|       "jna paeth_abb2              \n\t"
 | ||
|       // pa > pb; now test if pb <= pc
 | ||
|       "cmpl " pc_TEMP ", %%ecx     \n\t"
 | ||
|       "jna paeth_bbc2              \n\t"
 | ||
|       // pb > pc; Raw(x) = Paeth(x) + Prior(x-bpp)
 | ||
|       "movb (%1," PBP ",), %%cl    \n\t" // load Prior(x-bpp) into cl
 | ||
|       "jmp paeth_paeth2            \n\t"
 | ||
| 
 | ||
|    "paeth_bbc2:                    \n\t"
 | ||
|       // pb <= pc; Raw(x) = Paeth(x) + Prior(x)
 | ||
|       "movb (%1," PAX ",), %%cl    \n\t" // load Prior(x) into cl
 | ||
|       "jmp paeth_paeth2            \n\t"
 | ||
| 
 | ||
|    "paeth_abb2:                    \n\t"
 | ||
|       // pa <= pb; now test if pa <= pc
 | ||
|       "cmpl " pc_TEMP ", %%ebx     \n\t"
 | ||
|       "jna paeth_abc2              \n\t"
 | ||
|       // pa > pc; Raw(x) = Paeth(x) + Prior(x-bpp)
 | ||
|       "movb (%1," PBP ",), %%cl    \n\t" // load Prior(x-bpp) into cl
 | ||
|       "jmp paeth_paeth2            \n\t"
 | ||
| 
 | ||
|    "paeth_abc2:                    \n\t"
 | ||
|       // pa <= pc; Raw(x) = Paeth(x) + Raw(x-bpp)
 | ||
|       "movb (%2," PBP ",), %%cl    \n\t" // load Raw(x-bpp) into cl
 | ||
| 
 | ||
|    "paeth_paeth2:                  \n\t"
 | ||
|       "incl %%eax                  \n\t"
 | ||
|       "incl %%ebp                  \n\t"
 | ||
|       // Raw(x) = (Paeth(x) + Paeth_Predictor( a, b, c )) mod 256
 | ||
|       "addb %%cl, -1(%2," PAX ",)  \n\t"
 | ||
|       "cmpl %%edx, %%eax           \n\t" // check against FullLength
 | ||
|       "jb paeth_lp2                \n\t"
 | ||
| 
 | ||
|       RESTORE_r11_r12_r13
 | ||
|       RESTORE_GOT_ebx
 | ||
|       RESTORE_ebp
 | ||
| 
 | ||
|    "paeth_end:                     \n\t"
 | ||
|       "EMMS                        \n\t" // end MMX; prep for poss. FP instrs.
 | ||
| 
 | ||
|       : "=c" (dummy_value_c),            // output regs (dummy)
 | ||
|         "=S" (dummy_value_S),
 | ||
|         "=D" (dummy_value_D),
 | ||
|         "=a" (dummy_value_a),
 | ||
|         "=d" (dummy_value_d)
 | ||
| 
 | ||
|       : "0" (bpp),         // ecx        // input regs
 | ||
|         "1" (prev_row),    // esi/rsi
 | ||
|         "2" (row),         // edi/rdi
 | ||
|         "3" (MMXLength),   // eax
 | ||
|         "4" (FullLength)   // edx
 | ||
| 
 | ||
|       CLOB_COLON_ebx_ebp_r1X             // clobber list
 | ||
|         CLOBBER_GOT_ebx
 | ||
|         CLOB_COMMA_ebx_ebp
 | ||
|         CLOBBER_ebp
 | ||
|         CLOB_COMMA_ebX_r1X
 | ||
|         CLOBBER_r11_r12_r13
 | ||
|    );
 | ||
| 
 | ||
| } /* end png_read_filter_row_mmx_paeth() */
 | ||
| 
 | ||
| #endif // PNG_x86_64_USE_GOTPCREL || PNG_THREAD_UNSAFE_OK
 | ||
| #endif /* PNG_MMX_READ_FILTER_PAETH_SUPPORTED */
 | ||
| 
 | ||
| 
 | ||
| 
 | ||
| 
 | ||
| #if defined(PNG_MMX_READ_FILTER_SUB_SUPPORTED)
 | ||
| 
 | ||
| //===========================================================================//
 | ||
| //                                                                           //
 | ||
| //           P N G _ R E A D _ F I L T E R _ R O W _ M M X _ S U B           //
 | ||
| //                                                                           //
 | ||
| //===========================================================================//
 | ||
| 
 | ||
| // Optimized code for PNG Sub filter decoder
 | ||
| 
 | ||
| static void /* PRIVATE */
 | ||
| png_read_filter_row_mmx_sub(png_row_infop row_info, png_bytep row)
 | ||
| {
 | ||
|    unsigned FullLength, MMXLength;  // png_uint_32 is actually 64-bit on x86-64
 | ||
|    int bpp;
 | ||
|    int dummy_value_a;
 | ||
|    int dummy_value_c;
 | ||
|    int dummy_value_d;
 | ||
|    png_bytep dummy_value_D;
 | ||
|    int diff; //     __attribute__((used));
 | ||
| 
 | ||
|    bpp = (row_info->pixel_depth + 7) >> 3;  // calc number of bytes per pixel
 | ||
|    FullLength = row_info->rowbytes - bpp;   // number of bytes to filter
 | ||
|      // (why do we subtract off bpp?  not so in avg or paeth...)
 | ||
| 
 | ||
|    __asm__ __volatile__ (
 | ||
|       SAVE_r15
 | ||
|       SAVE_ebp
 | ||
| //pre "movl row, %1                \n\t" // edi/rdi
 | ||
|       "mov  %1, " PSI "            \n\t" // lp = row
 | ||
| //pre "movl bpp, %%ecx             \n\t"
 | ||
|       "add  " PCX ", %1            \n\t" // rp = row + bpp
 | ||
| //pre "movl FullLength, %%eax      \n\t" // bring in via eax...
 | ||
|       SAVE_FullLength                    // ...but store for later use
 | ||
| 
 | ||
|       "xorl %%eax, %%eax           \n\t"
 | ||
| 
 | ||
|       // get # of bytes to alignment (note:  computing _delta_ of two pointers,
 | ||
|       // so hereafter %%ebp is sufficient even on 64-bit)
 | ||
|       "mov  %1, " PBP "            \n\t" // take start of row
 | ||
|       "add  $0xf, " PBP "          \n\t" // add 7+8 to incr past alignment bdry
 | ||
| //    "andl $0xfffffff8, %%ebp     \n\t" // mask to alignment boundary (32-bit!)
 | ||
|       CLEAR_BOTTOM_3_BITS  PBP    "\n\t" // mask to alignment boundary
 | ||
|       "sub  %1, " PBP "            \n\t" // subtract row ptr again => ebp =
 | ||
|       "jz sub_go                   \n\t" //  target value of eax at alignment
 | ||
| 
 | ||
|    "sub_lp1:                       \n\t" // fix alignment
 | ||
|       "movb (" PSI "," PAX ",), %%cl \n\t"
 | ||
|       "addb %%cl, (%1," PAX ",)    \n\t"
 | ||
|       "incl %%eax                  \n\t"
 | ||
|       "cmpl %%ebp, %%eax           \n\t"
 | ||
|       "jb sub_lp1                  \n\t"
 | ||
| 
 | ||
|    "sub_go:                        \n\t"
 | ||
|       RESTORE_FullLength "%%ecx    \n\t" // FullLength -> ecx
 | ||
|       "movl %%ecx, %%edx           \n\t"
 | ||
|       "subl %%eax, %%edx           \n\t" // subtract alignment fix
 | ||
|       "andl $0x00000007, %%edx     \n\t" // calc bytes over mult of 8
 | ||
|       "subl %%edx, %%ecx           \n\t" // drop over bytes from length
 | ||
| //out "movl %%ecx, MMXLength       \n\t"
 | ||
|       "movl %%ebp, %%eax           \n\t" // ebp = diff, but no reg constraint(?)
 | ||
|       RESTORE_ebp                        //  (could swap ebp and ecx functions,
 | ||
|       RESTORE_r15                        //  but %%cl issues...)
 | ||
| 
 | ||
|       : "=c" (MMXLength),       // 0     // output regs
 | ||
|         "=D" (dummy_value_D),   // 1
 | ||
|         "=a" (diff)             // 2
 | ||
| 
 | ||
|       : "0" (bpp),              // ecx   // input regs
 | ||
|         "1" (row),              // edi
 | ||
|         "2" (FullLength)        // eax
 | ||
| 
 | ||
|       : "%esi", "%edx"                   // clobber list
 | ||
|         _CLOBBER_r15
 | ||
|         _CLOBBER_ebp
 | ||
|    );
 | ||
| 
 | ||
|    // now do the math for the rest of the row
 | ||
|    switch (bpp)
 | ||
|    {
 | ||
|       case 3:
 | ||
|       {
 | ||
| //       _ShiftBpp = 24;       // == 3 * 8
 | ||
| //       _ShiftRem  = 40;      // == 64 - 24
 | ||
| 
 | ||
|          __asm__ __volatile__ (
 | ||
| // preload  "mov  row, %1                 \n\t" // edi/rdi
 | ||
|             LOAD_GOT_rbp
 | ||
|             // load (former) _ActiveMask for 2nd active byte group
 | ||
|             "movq " AMASK2_3_3 ", %%mm7   \n\t" // _amask2_3_3
 | ||
|             RESTORE_rbp
 | ||
| 
 | ||
| // notused  "mov  %1, " PSI "             \n\t" // lp = row
 | ||
| // preload  "movl bpp, %%ecx              \n\t"
 | ||
|             "add  " PCX ", %1             \n\t" // rp = row + bpp
 | ||
|             "movq %%mm7, %%mm6            \n\t"
 | ||
| // preload  "movl diff, %%edx             \n\t"
 | ||
|             "psllq $24, %%mm6             \n\t" // move mask in mm6 to cover
 | ||
|                                                 //  3rd active byte group
 | ||
|             // prime the pump:  load the first Raw(x-bpp) data set
 | ||
|             "movq -8(%1," PDX ",), %%mm1  \n\t"
 | ||
| 
 | ||
|          "sub_3lp:                        \n\t" // shift data for adding first
 | ||
|             "psrlq $40, %%mm1             \n\t" //  bpp bytes (no need for mask;
 | ||
|                                                 //  shift clears inactive bytes)
 | ||
|             // add 1st active group
 | ||
|             "movq (%1," PDX ",), %%mm0    \n\t"
 | ||
|             "paddb %%mm1, %%mm0           \n\t"
 | ||
| 
 | ||
|             // add 2nd active group
 | ||
|             "movq %%mm0, %%mm1            \n\t" // mov updated Raws to mm1
 | ||
|             "psllq $24, %%mm1             \n\t" // shift data to pos. correctly
 | ||
|             "pand %%mm7, %%mm1            \n\t" // mask to use 2nd active group
 | ||
|             "paddb %%mm1, %%mm0           \n\t"
 | ||
| 
 | ||
|             // add 3rd active group
 | ||
|             "movq %%mm0, %%mm1            \n\t" // mov updated Raws to mm1
 | ||
|             "psllq $24, %%mm1             \n\t" // shift data to pos. correctly
 | ||
|             "pand %%mm6, %%mm1            \n\t" // mask to use 3rd active group
 | ||
|             "addl $8, %%edx               \n\t"
 | ||
|             "paddb %%mm1, %%mm0           \n\t"
 | ||
| 
 | ||
|             "cmpl %%eax, %%edx            \n\t" // MMXLength
 | ||
|             "movq %%mm0, -8(%1," PDX ",)  \n\t" // write updated Raws to array
 | ||
|             "movq %%mm0, %%mm1            \n\t" // prep 1st add at top of loop
 | ||
|             "jb sub_3lp                   \n\t"
 | ||
| 
 | ||
|             : "=c" (dummy_value_c),   // 0      // output regs (dummy)
 | ||
|               "=D" (dummy_value_D),   // 1
 | ||
|               "=d" (dummy_value_d),   // 2
 | ||
|               "=a" (dummy_value_a)    // 3
 | ||
| 
 | ||
|             : "0" (bpp),              // ecx    // input regs
 | ||
|               "1" (row),              // edi
 | ||
|               "2" (diff),             // edx
 | ||
|               "3" (MMXLength)         // eax
 | ||
| 
 | ||
| #if defined(CLOBBER_MMX_REGS_SUPPORTED)
 | ||
|             : "%mm0", "%mm1", "%mm6", "%mm7"    // clobber list
 | ||
| #endif
 | ||
|          );
 | ||
|       }
 | ||
|       break;  // end 3 bpp
 | ||
| 
 | ||
|       case 4:   // formerly shared with 6 bpp case via _ShiftBpp and _ShiftRem,
 | ||
|       {         // but 64-bit PIC/.so problems (could still share, moving vars
 | ||
|                 // into unused MMX regs via ecx/edx, but kludgy)
 | ||
| //       _ShiftBpp = bpp << 3;        // 32 (psllq)
 | ||
| //       _ShiftRem = 64 - _ShiftBpp;  // 32 (psrlq)
 | ||
| 
 | ||
|          __asm__ __volatile__ (
 | ||
| // preload  "mov  row, %1                 \n\t" // edi/rdi
 | ||
| // preload  "movl diff, %%edx             \n\t"
 | ||
| // notused  "mov  %1, " PSI "             \n\t" // lp = row
 | ||
| // preload  "movl bpp, %%ecx              \n\t"
 | ||
|             "add  " PCX ", %1             \n\t" // rp = row + bpp
 | ||
| 
 | ||
|             // prime the pump:  load the first Raw(x-bpp) data set
 | ||
|             "movq -8(%1," PDX ",), %%mm1  \n\t"
 | ||
| 
 | ||
|          "sub_4lp:                        \n\t" // shift data for adding first
 | ||
|             "psrlq $32, %%mm1             \n\t" //  bpp bytes (no need for mask;
 | ||
|                                                 //  shift clears inactive bytes)
 | ||
|             "movq (%1," PDX ",), %%mm0    \n\t"
 | ||
|             "paddb %%mm1, %%mm0           \n\t"
 | ||
| 
 | ||
|             // add 2nd active group
 | ||
|             "movq %%mm0, %%mm1            \n\t" // mov updated Raws to mm1
 | ||
|             "psllq $32, %%mm1             \n\t" // shift data to pos. correctly
 | ||
|             "addl $8, %%edx               \n\t"
 | ||
|             "paddb %%mm1, %%mm0           \n\t"
 | ||
| 
 | ||
|             "cmpl %%eax, %%edx            \n\t" // MMXLength
 | ||
|             "movq %%mm0, -8(%1," PDX ",)  \n\t" // write updated Raws to array
 | ||
|             "movq %%mm0, %%mm1            \n\t" // prep 1st add at top of loop
 | ||
|             "jb sub_4lp                   \n\t"
 | ||
| 
 | ||
|             : "=c" (dummy_value_c),   // 0      // output regs (dummy)
 | ||
|               "=D" (dummy_value_D),   // 1
 | ||
|               "=d" (dummy_value_d),   // 2
 | ||
|               "=a" (dummy_value_a)    // 3
 | ||
| 
 | ||
|             : "0" (bpp),              // ecx    // input regs
 | ||
|               "1" (row),              // edi
 | ||
|               "2" (diff),             // edx
 | ||
|               "3" (MMXLength)         // eax
 | ||
| 
 | ||
| #if defined(CLOBBER_MMX_REGS_SUPPORTED)
 | ||
|             : "%mm0", "%mm1"                    // clobber list
 | ||
| #endif
 | ||
|          );
 | ||
|       }
 | ||
|       break;  // end 4 bpp
 | ||
| 
 | ||
|       case 1:
 | ||
|       {
 | ||
|          __asm__ __volatile__ (
 | ||
| // preload  "movl diff, %%edx              \n\t"
 | ||
| // preload  "mov  row, %1                  \n\t" // edi/rdi
 | ||
| // preload  "cmpl FullLength, %%edx        \n\t"
 | ||
|             "cmpl %%eax, %%edx             \n\t"
 | ||
|             "jnb sub_1end                  \n\t"
 | ||
|             "mov  %1, " PSI "              \n\t" // lp = row
 | ||
| // irrel.   "xorl %%ecx, %%ecx             \n\t" // (actually bug with preload)
 | ||
| // preload  "movl bpp, %%ecx               \n\t"
 | ||
|             "add  " PCX ", %1              \n\t" // rp = row + bpp
 | ||
| 
 | ||
|          "sub_1lp:                         \n\t"
 | ||
|             "movb (" PSI "," PDX ",), %%cl \n\t"
 | ||
|             "addb %%cl, (%1," PDX ",)      \n\t"
 | ||
|             "incl %%edx                    \n\t"
 | ||
|             "cmpl %%eax, %%edx             \n\t" // compare with FullLength
 | ||
|             "jb sub_1lp                    \n\t"
 | ||
| 
 | ||
|          "sub_1end:                        \n\t"
 | ||
| 
 | ||
|             : "=c" (dummy_value_c),   // 0      // output regs (dummy)
 | ||
|               "=D" (dummy_value_D),   // 1
 | ||
|               "=d" (dummy_value_d),   // 2
 | ||
|               "=a" (dummy_value_a)    // 3
 | ||
| 
 | ||
|             : "0" (bpp),              // ecx    // input regs
 | ||
|               "1" (row),              // edi
 | ||
|               "2" (diff),             // edx
 | ||
|               "3" (FullLength)        // eax
 | ||
| 
 | ||
|             : "%esi"                            // clobber list
 | ||
|          );
 | ||
|       }
 | ||
|       return;  // end 1 bpp (bypassing cleanup block!)
 | ||
| 
 | ||
|       case 2:
 | ||
|       {
 | ||
| //       _ShiftBpp = 16;       // == 2 * 8
 | ||
| //       _ShiftRem = 48;       // == 64 - 16
 | ||
| 
 | ||
|          __asm__ __volatile__ (
 | ||
|             LOAD_GOT_rbp
 | ||
|             // load (former) _ActiveMask for 2nd active byte group
 | ||
|             "movq " AMASK4_2_2 ", %%mm7   \n\t" // _amask4_2_2
 | ||
|             RESTORE_rbp
 | ||
| // preload  "movl diff, %%edx             \n\t"
 | ||
|             "movq %%mm7, %%mm6            \n\t"
 | ||
| // preload  "mov  row, %1                 \n\t" // edi/rdi
 | ||
|             "psllq $16, %%mm6             \n\t" // move mask in mm6 to cover
 | ||
|                                                 //  3rd active byte group
 | ||
| // notused  "mov  %1, " PSI "             \n\t" // lp = row
 | ||
|             "movq %%mm6, %%mm5            \n\t"
 | ||
| // preload  "movl bpp, %%ecx              \n\t"
 | ||
|             "add  " PCX ", %1             \n\t" // rp = row + bpp
 | ||
|             "psllq $16, %%mm5             \n\t" // move mask in mm5 to cover
 | ||
|                                                 //  4th active byte group
 | ||
|             // prime the pump:  load the first Raw(x-bpp) data set
 | ||
|             "movq -8(%1," PDX ",), %%mm1  \n\t"
 | ||
| 
 | ||
|          "sub_2lp:                        \n\t" // shift data for adding first
 | ||
|             "psrlq $48, %%mm1             \n\t" //  bpp bytes (no need for mask;
 | ||
|                                                 //  shift clears inactive bytes)
 | ||
|             // add 1st active group
 | ||
|             "movq (%1," PDX ",), %%mm0    \n\t"
 | ||
|             "paddb %%mm1, %%mm0           \n\t"
 | ||
| 
 | ||
|             // add 2nd active group
 | ||
|             "movq %%mm0, %%mm1            \n\t" // mov updated Raws to mm1
 | ||
|             "psllq $16, %%mm1             \n\t" // shift data to pos. correctly
 | ||
|             "pand %%mm7, %%mm1            \n\t" // mask to use 2nd active group
 | ||
|             "paddb %%mm1, %%mm0           \n\t"
 | ||
| 
 | ||
|             // add 3rd active group
 | ||
|             "movq %%mm0, %%mm1            \n\t" // mov updated Raws to mm1
 | ||
|             "psllq $16, %%mm1             \n\t" // shift data to pos. correctly
 | ||
|             "pand %%mm6, %%mm1            \n\t" // mask to use 3rd active group
 | ||
|             "paddb %%mm1, %%mm0           \n\t"
 | ||
| 
 | ||
|             // add 4th active group
 | ||
|             "movq %%mm0, %%mm1            \n\t" // mov updated Raws to mm1
 | ||
|             "psllq $16, %%mm1             \n\t" // shift data to pos. correctly
 | ||
|             "pand %%mm5, %%mm1            \n\t" // mask to use 4th active group
 | ||
|             "addl $8, %%edx               \n\t"
 | ||
|             "paddb %%mm1, %%mm0           \n\t"
 | ||
|             "cmpl %%eax, %%edx            \n\t" // MMXLength
 | ||
|             "movq %%mm0, -8(%1," PDX ",)  \n\t" // write updated Raws to array
 | ||
|             "movq %%mm0, %%mm1            \n\t" // prep 1st add at top of loop
 | ||
|             "jb sub_2lp                   \n\t"
 | ||
| 
 | ||
|             : "=c" (dummy_value_c),   // 0      // output regs (dummy)
 | ||
|               "=D" (dummy_value_D),   // 1
 | ||
|               "=d" (dummy_value_d),   // 2
 | ||
|               "=a" (dummy_value_a)    // 3
 | ||
| 
 | ||
|             : "0" (bpp),              // ecx    // input regs
 | ||
|               "1" (row),              // edi
 | ||
|               "2" (diff),             // edx
 | ||
|               "3" (MMXLength)         // eax
 | ||
| 
 | ||
| #if defined(CLOBBER_MMX_REGS_SUPPORTED)
 | ||
|             : "%mm0", "%mm1", "%mm5", "%mm6"    // clobber list
 | ||
|             , "%mm7"
 | ||
| #endif
 | ||
|          );
 | ||
|       }
 | ||
|       break;  // end 2 bpp
 | ||
| 
 | ||
|       case 6:   // formerly shared with 4 bpp case (see comments there)
 | ||
|       {
 | ||
| //       _ShiftBpp = bpp << 3;        // 48 (psllq)
 | ||
| //       _ShiftRem = 64 - _ShiftBpp;  // 16 (psrlq)
 | ||
| 
 | ||
|          __asm__ __volatile__ (
 | ||
| // preload  "mov  row, %1                 \n\t" // edi/rdi
 | ||
| // preload  "movl diff, %%edx             \n\t"
 | ||
| // notused  "mov  %1, " PSI "             \n\t" // lp = row
 | ||
| // preload  "movl bpp, %%ecx              \n\t"
 | ||
|             "add  " PCX ", %1             \n\t" // rp = row + bpp
 | ||
| 
 | ||
|             // prime the pump:  load the first Raw(x-bpp) data set
 | ||
|             "movq -8(%1," PDX ",), %%mm1  \n\t"
 | ||
| 
 | ||
|          "sub_6lp:                        \n\t" // shift data for adding first
 | ||
|             "psrlq $16, %%mm1             \n\t" //  bpp bytes (no need for mask;
 | ||
|                                                 //  shift clears inactive bytes)
 | ||
|             "movq (%1," PDX ",), %%mm0    \n\t"
 | ||
|             "paddb %%mm1, %%mm0           \n\t"
 | ||
| 
 | ||
|             // add 2nd active group
 | ||
|             "movq %%mm0, %%mm1            \n\t" // mov updated Raws to mm1
 | ||
|             "psllq $48, %%mm1             \n\t" // shift data to pos. correctly
 | ||
|             "addl $8, %%edx               \n\t"
 | ||
|             "paddb %%mm1, %%mm0           \n\t"
 | ||
| 
 | ||
|             "cmpl %%eax, %%edx            \n\t" // MMXLength
 | ||
|             "movq %%mm0, -8(%1," PDX ",)  \n\t" // write updated Raws to array
 | ||
|             "movq %%mm0, %%mm1            \n\t" // prep 1st add at top of loop
 | ||
|             "jb sub_6lp                   \n\t"
 | ||
| 
 | ||
|             : "=c" (dummy_value_c),   // 0      // output regs (dummy)
 | ||
|               "=D" (dummy_value_D),   // 1
 | ||
|               "=d" (dummy_value_d),   // 2
 | ||
|               "=a" (dummy_value_a)    // 3
 | ||
| 
 | ||
|             : "0" (bpp),              // ecx    // input regs
 | ||
|               "1" (row),              // edi
 | ||
|               "2" (diff),             // edx
 | ||
|               "3" (MMXLength)         // eax
 | ||
| 
 | ||
| #if defined(CLOBBER_MMX_REGS_SUPPORTED)
 | ||
|             : "%mm0", "%mm1"                    // clobber list
 | ||
| #endif
 | ||
|          );
 | ||
|       }
 | ||
|       break;  // end 6 bpp
 | ||
| 
 | ||
|       case 8:
 | ||
|       {
 | ||
|          __asm__ __volatile__ (
 | ||
| // preload  "mov  row, %1                 \n\t" // edi/rdi
 | ||
| // preload  "movl diff, %%edx             \n\t"
 | ||
| // notused  "mov  %1, " PSI "             \n\t" // lp = row
 | ||
| // preload  "movl bpp, %%ecx              \n\t"
 | ||
|             "add  " PCX ", %1             \n\t" // rp = row + bpp
 | ||
| // preload  "movl MMXLength, %%eax        \n\t"
 | ||
| 
 | ||
|             // prime the pump:  load the first Raw(x-bpp) data set
 | ||
|             "movq -8(%1," PDX ",), %%mm7  \n\t"
 | ||
|             "movl %%eax, %%esi            \n\t" // copy of MMXLength -> esi
 | ||
|             "andl $0x0000003f, %%esi      \n\t" // calc bytes over mult of 64
 | ||
| 
 | ||
|          "sub_8lp:                        \n\t"
 | ||
|             "movq (%1," PDX ",), %%mm0    \n\t" // load Sub(x) for 1st 8 bytes
 | ||
|             "paddb %%mm7, %%mm0           \n\t"
 | ||
|             "movq 8(%1," PDX ",), %%mm1   \n\t" // load Sub(x) for 2nd 8 bytes
 | ||
|             "movq %%mm0, (%1," PDX ",)    \n\t" // write Raw(x) for 1st 8 bytes
 | ||
| 
 | ||
|             // Now mm0 will be used as Raw(x-bpp) for the 2nd group of 8 bytes.
 | ||
|             // This will be repeated for each group of 8 bytes with the 8th
 | ||
|             // group being used as the Raw(x-bpp) for the 1st group of the
 | ||
|             // next loop.
 | ||
| 
 | ||
|             "paddb %%mm0, %%mm1           \n\t"
 | ||
|             "movq 16(%1," PDX ",), %%mm2  \n\t" // load Sub(x) for 3rd 8 bytes
 | ||
|             "movq %%mm1, 8(%1," PDX ",)   \n\t" // write Raw(x) for 2nd 8 bytes
 | ||
|             "paddb %%mm1, %%mm2           \n\t"
 | ||
|             "movq 24(%1," PDX ",), %%mm3  \n\t" // load Sub(x) for 4th 8 bytes
 | ||
|             "movq %%mm2, 16(%1," PDX ",)  \n\t" // write Raw(x) for 3rd 8 bytes
 | ||
|             "paddb %%mm2, %%mm3           \n\t"
 | ||
|             "movq 32(%1," PDX ",), %%mm4  \n\t" // load Sub(x) for 5th 8 bytes
 | ||
|             "movq %%mm3, 24(%1," PDX ",)  \n\t" // write Raw(x) for 4th 8 bytes
 | ||
|             "paddb %%mm3, %%mm4           \n\t"
 | ||
|             "movq 40(%1," PDX ",), %%mm5  \n\t" // load Sub(x) for 6th 8 bytes
 | ||
|             "movq %%mm4, 32(%1," PDX ",)  \n\t" // write Raw(x) for 5th 8 bytes
 | ||
|             "paddb %%mm4, %%mm5           \n\t"
 | ||
|             "movq 48(%1," PDX ",), %%mm6  \n\t" // load Sub(x) for 7th 8 bytes
 | ||
|             "movq %%mm5, 40(%1," PDX ",)  \n\t" // write Raw(x) for 6th 8 bytes
 | ||
|             "paddb %%mm5, %%mm6           \n\t"
 | ||
|             "movq 56(%1," PDX ",), %%mm7  \n\t" // load Sub(x) for 8th 8 bytes
 | ||
|             "movq %%mm6, 48(%1," PDX ",)  \n\t" // write Raw(x) for 7th 8 bytes
 | ||
|             "addl $64, %%edx              \n\t"
 | ||
|             "paddb %%mm6, %%mm7           \n\t"
 | ||
|             "cmpl %%esi, %%edx            \n\t" // cmp to bytes over mult of 64
 | ||
|             "movq %%mm7, -8(%1," PDX ",)  \n\t" // write Raw(x) for 8th 8 bytes
 | ||
|             "jb sub_8lp                   \n\t"
 | ||
| 
 | ||
|             "cmpl %%eax, %%edx            \n\t" // compare to MMXLength
 | ||
|             "jnb sub_8lt8                 \n\t"
 | ||
| 
 | ||
|          "sub_8lpA:                       \n\t"
 | ||
|             "movq (%1," PDX ",), %%mm0    \n\t"
 | ||
|             "addl $8, %%edx               \n\t"
 | ||
|             "paddb %%mm7, %%mm0           \n\t"
 | ||
|             "cmpl %%eax, %%edx            \n\t" // compare to MMXLength
 | ||
|             "movq %%mm0, -8(%1," PDX ",)  \n\t" // -8 to offset early addl edx
 | ||
|             "movq %%mm0, %%mm7            \n\t" // move calculated Raw(x) data
 | ||
|             "jb sub_8lpA                  \n\t" //  to mm7 to be new Raw(x-bpp)
 | ||
|                                                 //  for next loop
 | ||
|          "sub_8lt8:                       \n\t"
 | ||
| 
 | ||
|             : "=c" (dummy_value_c),   // 0      // output regs (dummy)
 | ||
|               "=D" (dummy_value_D),   // 1
 | ||
|               "=d" (dummy_value_d),   // 2
 | ||
|               "=a" (dummy_value_a)    // 3
 | ||
| 
 | ||
|             : "0" (bpp),              // ecx    // input regs
 | ||
|               "1" (row),              // edi
 | ||
|               "2" (diff),             // edx
 | ||
|               "3" (MMXLength)         // eax
 | ||
| 
 | ||
|             : "%esi"                            // clobber list
 | ||
| #if defined(CLOBBER_MMX_REGS_SUPPORTED)
 | ||
|             , "%mm0", "%mm1", "%mm2", "%mm3"
 | ||
|             , "%mm4", "%mm5", "%mm6", "%mm7"
 | ||
| #endif
 | ||
|          );
 | ||
|       }
 | ||
|       break;  // end 8 bpp
 | ||
| 
 | ||
|       default:                // bpp != 1,2,3,4,6,8:  doesn't exist
 | ||
|       {
 | ||
|          // ERROR:  SHOULD NEVER BE REACHED
 | ||
| #if defined(PNG_DEBUG)
 | ||
|          png_debug(1, "Internal libpng logic error (GCC "
 | ||
|            "png_read_filter_row_mmx_sub())\n");
 | ||
| #endif
 | ||
|       }
 | ||
|       break;
 | ||
| 
 | ||
|    } // end switch (bpp)
 | ||
| 
 | ||
|    __asm__ __volatile__ (
 | ||
| //pre "movl MMXLength, %%eax         \n\t"
 | ||
| //pre "mov  row, %1                  \n\t" // edi/rdi
 | ||
| //pre "cmpl FullLength, %%eax        \n\t"
 | ||
|       "cmpl %%edx, %%eax             \n\t"
 | ||
|       "jnb sub_end                   \n\t"
 | ||
| 
 | ||
|       "mov  %1, " PSI "              \n\t" // lp = row
 | ||
| //pre "movl bpp, %%ecx               \n\t"
 | ||
|       "add  " PCX ", %1              \n\t" // rp = row + bpp
 | ||
|       "xorl %%ecx, %%ecx             \n\t"
 | ||
| 
 | ||
|    "sub_lp2:                         \n\t"
 | ||
|       "movb (" PSI "," PAX ",), %%cl \n\t"
 | ||
|       "addb %%cl, (%1," PAX ",)      \n\t"
 | ||
|       "incl %%eax                    \n\t"
 | ||
|       "cmpl %%edx, %%eax             \n\t" // FullLength
 | ||
|       "jb sub_lp2                    \n\t"
 | ||
| 
 | ||
|    "sub_end:                         \n\t"
 | ||
|       "EMMS                          \n\t" // end MMX instructions
 | ||
| 
 | ||
|       : "=c" (dummy_value_c),   // 0      // output regs (dummy)
 | ||
|         "=D" (dummy_value_D),   // 1
 | ||
|         "=a" (dummy_value_a),   // 2
 | ||
|         "=d" (dummy_value_d)    // 3
 | ||
| 
 | ||
|       : "0" (bpp),              // ecx    // input regs
 | ||
|         "1" (row),              // edi
 | ||
|         "2" (MMXLength),        // eax
 | ||
|         "3" (FullLength)        // edx
 | ||
| 
 | ||
|       : "%esi"                            // clobber list
 | ||
|    );
 | ||
| 
 | ||
| } // end of png_read_filter_row_mmx_sub()
 | ||
| 
 | ||
| #endif /* PNG_MMX_READ_FILTER_SUB_SUPPORTED */
 | ||
| 
 | ||
| 
 | ||
| 
 | ||
| 
 | ||
| #if defined(PNG_MMX_READ_FILTER_UP_SUPPORTED)
 | ||
| 
 | ||
| //===========================================================================//
 | ||
| //                                                                           //
 | ||
| //            P N G _ R E A D _ F I L T E R _ R O W _ M M X _ U P            //
 | ||
| //                                                                           //
 | ||
| //===========================================================================//
 | ||
| 
 | ||
| // Optimized code for PNG Up filter decoder
 | ||
| 
 | ||
| static void /* PRIVATE */
 | ||
| png_read_filter_row_mmx_up(png_row_infop row_info, png_bytep row,
 | ||
|                            png_bytep prev_row)
 | ||
| {
 | ||
|    unsigned len;        // png_uint_32 is actually 64-bit on x86-64
 | ||
|    int dummy_value_d;   // fix 'forbidden register 3 (dx) was spilled' error
 | ||
|    png_bytep dummy_value_S;
 | ||
|    png_bytep dummy_value_D;
 | ||
| 
 | ||
|    len = row_info->rowbytes;              // number of bytes to filter
 | ||
| 
 | ||
|    __asm__ __volatile__ (
 | ||
|       SAVE_GOT_ebx
 | ||
| //pre "mov  prev_row, %1           \n\t" // esi/rsi
 | ||
| //pre "movl row, %2                \n\t" // edi/rdi
 | ||
| 
 | ||
|       "xorl %%ebx, %%ebx           \n\t"
 | ||
|       "xorl %%eax, %%eax           \n\t"
 | ||
| 
 | ||
|       // get # of bytes to alignment (note:  computing _delta_ of two pointers,
 | ||
|       // so hereafter %%ecx is sufficient even on 64-bit)
 | ||
|       "mov  %2, " PCX "            \n\t" // take start of row
 | ||
|       "add  $0x7, " PCX "          \n\t" // add 7 to incr past alignment bdry
 | ||
| //    "andl $0xfffffff8, %%ecx     \n\t" // mask to alignment boundary (32-bit!)
 | ||
|       CLEAR_BOTTOM_3_BITS  PCX    "\n\t" // mask to alignment boundary
 | ||
|       "sub  %2, " PCX "            \n\t" // subtract row ptr again => ebp =
 | ||
|       "jz up_go                    \n\t" //  target value of ecx at alignment
 | ||
| 
 | ||
|    "up_lp1:                        \n\t" // fix alignment
 | ||
|       "movb (%2," PBX ",), %%al    \n\t"
 | ||
|       "addb (%1," PBX ",), %%al    \n\t"
 | ||
|       "incl %%ebx                  \n\t"
 | ||
|       "cmpl %%ecx, %%ebx           \n\t"
 | ||
|       "movb %%al, -1(%2," PBX ",)  \n\t" // mov does not affect flags; -1 to
 | ||
|       "jb up_lp1                   \n\t" //  offset incl ebx
 | ||
| 
 | ||
|    "up_go:                         \n\t"
 | ||
| //pre "movl len, %%edx             \n\t"
 | ||
|       "movl %%edx, %%ecx           \n\t"
 | ||
|       "subl %%ebx, %%edx           \n\t" // subtract alignment fix
 | ||
|       "andl $0x0000003f, %%edx     \n\t" // calc bytes over mult of 64
 | ||
|       "subl %%edx, %%ecx           \n\t" // sub over-bytes from original length
 | ||
| 
 | ||
|       // unrolled loop - use all MMX registers and interleave to reduce
 | ||
|       // number of branch instructions (loops) and reduce partial stalls
 | ||
|    "up_loop:                       \n\t"
 | ||
|       "movq (%1," PBX ",), %%mm1   \n\t"
 | ||
|       "movq (%2," PBX ",), %%mm0   \n\t"
 | ||
|       "movq 8(%1," PBX ",), %%mm3  \n\t"
 | ||
|       "paddb %%mm1, %%mm0          \n\t"
 | ||
|       "movq 8(%2," PBX ",), %%mm2  \n\t"
 | ||
|       "movq %%mm0, (%2," PBX ",)   \n\t"
 | ||
|       "paddb %%mm3, %%mm2          \n\t"
 | ||
|       "movq 16(%1," PBX ",), %%mm5 \n\t"
 | ||
|       "movq %%mm2, 8(%2," PBX ",)  \n\t"
 | ||
|       "movq 16(%2," PBX ",), %%mm4 \n\t"
 | ||
|       "movq 24(%1," PBX ",), %%mm7 \n\t"
 | ||
|       "paddb %%mm5, %%mm4          \n\t"
 | ||
|       "movq 24(%2," PBX ",), %%mm6 \n\t"
 | ||
|       "movq %%mm4, 16(%2," PBX ",) \n\t"
 | ||
|       "paddb %%mm7, %%mm6          \n\t"
 | ||
|       "movq 32(%1," PBX ",), %%mm1 \n\t"
 | ||
|       "movq %%mm6, 24(%2," PBX ",) \n\t"
 | ||
|       "movq 32(%2," PBX ",), %%mm0 \n\t"
 | ||
|       "movq 40(%1," PBX ",), %%mm3 \n\t"
 | ||
|       "paddb %%mm1, %%mm0          \n\t"
 | ||
|       "movq 40(%2," PBX ",), %%mm2 \n\t"
 | ||
|       "movq %%mm0, 32(%2," PBX ",) \n\t"
 | ||
|       "paddb %%mm3, %%mm2          \n\t"
 | ||
|       "movq 48(%1," PBX ",), %%mm5 \n\t"
 | ||
|       "movq %%mm2, 40(%2," PBX ",) \n\t"
 | ||
|       "movq 48(%2," PBX ",), %%mm4 \n\t"
 | ||
|       "movq 56(%1," PBX ",), %%mm7 \n\t"
 | ||
|       "paddb %%mm5, %%mm4          \n\t"
 | ||
|       "movq 56(%2," PBX ",), %%mm6 \n\t"
 | ||
|       "movq %%mm4, 48(%2," PBX ",) \n\t"
 | ||
|       "addl $64, %%ebx             \n\t"
 | ||
|       "paddb %%mm7, %%mm6          \n\t"
 | ||
|       "cmpl %%ecx, %%ebx           \n\t"
 | ||
|       "movq %%mm6, -8(%2," PBX ",) \n\t" // (+56)movq does not affect flags;
 | ||
|       "jb up_loop                  \n\t" //  -8 to offset addl ebx
 | ||
| 
 | ||
|       "cmpl $0, %%edx              \n\t" // test for bytes over mult of 64
 | ||
|       "jz up_end                   \n\t"
 | ||
| 
 | ||
|       "cmpl $8, %%edx              \n\t" // test for less than 8 bytes
 | ||
|       "jb up_lt8                   \n\t" //  [added by lcreeve at netins.net]
 | ||
| 
 | ||
|       "addl %%edx, %%ecx           \n\t"
 | ||
|       "andl $0x00000007, %%edx     \n\t" // calc bytes over mult of 8
 | ||
|       "subl %%edx, %%ecx           \n\t" // drop over-bytes from length
 | ||
|       "jz up_lt8                   \n\t"
 | ||
| 
 | ||
|    "up_lpA:                        \n\t" // use MMX regs to update 8 bytes sim.
 | ||
|       "movq (%1," PBX ",), %%mm1   \n\t"
 | ||
|       "movq (%2," PBX ",), %%mm0   \n\t"
 | ||
|       "addl $8, %%ebx              \n\t"
 | ||
|       "paddb %%mm1, %%mm0          \n\t"
 | ||
|       "cmpl %%ecx, %%ebx           \n\t"
 | ||
|       "movq %%mm0, -8(%2," PBX ",) \n\t" // movq does not affect flags; -8 to
 | ||
|       "jb up_lpA                   \n\t" //  offset add ebx
 | ||
|       "cmpl $0, %%edx              \n\t" // test for bytes over mult of 8
 | ||
|       "jz up_end                   \n\t"
 | ||
| 
 | ||
|    "up_lt8:                        \n\t"
 | ||
|       "xorl %%eax, %%eax           \n\t"
 | ||
|       "addl %%edx, %%ecx           \n\t" // move over byte count into counter
 | ||
| 
 | ||
|    "up_lp2:                        \n\t" // use x86 regs for remaining bytes
 | ||
|       "movb (%2," PBX ",), %%al    \n\t"
 | ||
|       "addb (%1," PBX ",), %%al    \n\t"
 | ||
|       "incl %%ebx                  \n\t"
 | ||
|       "cmpl %%ecx, %%ebx           \n\t"
 | ||
|       "movb %%al, -1(%2," PBX ",)  \n\t" // mov does not affect flags; -1 to
 | ||
|       "jb up_lp2                   \n\t" //  offset inc ebx
 | ||
| 
 | ||
|    "up_end:                        \n\t"
 | ||
|       "EMMS                        \n\t" // conversion of filtered row complete
 | ||
|       RESTORE_GOT_ebx
 | ||
| 
 | ||
|       : "=d" (dummy_value_d),   // 0     // output regs (dummy)
 | ||
|         "=S" (dummy_value_S),   // 1
 | ||
|         "=D" (dummy_value_D)    // 2
 | ||
| 
 | ||
|       : "0" (len),              // edx   // input regs
 | ||
|         "1" (prev_row),         // esi
 | ||
|         "2" (row)               // edi
 | ||
| 
 | ||
|       : "%eax", "%ecx"                   // clobber list (no input regs!)
 | ||
|         _CLOBBER_GOT_ebx
 | ||
| #if defined(PNG_CLOBBER_MMX_REGS_SUPPORTED)
 | ||
|       , "%mm0", "%mm1", "%mm2", "%mm3"
 | ||
|       , "%mm4", "%mm5", "%mm6", "%mm7"
 | ||
| #endif
 | ||
|    );
 | ||
| 
 | ||
| } // end of png_read_filter_row_mmx_up()
 | ||
| 
 | ||
| #endif /* PNG_MMX_READ_FILTER_UP_SUPPORTED */
 | ||
| 
 | ||
| 
 | ||
| 
 | ||
| 
 | ||
| /*===========================================================================*/
 | ||
| /*                                                                           */
 | ||
| /*                   P N G _ R E A D _ F I L T E R _ R O W                   */
 | ||
| /*                                                                           */
 | ||
| /*===========================================================================*/
 | ||
| 
 | ||
| /* Optimized png_read_filter_row routines */
 | ||
| 
 | ||
| void /* PRIVATE */
 | ||
| png_read_filter_row(png_structp png_ptr, png_row_infop row_info, png_bytep
 | ||
|    row, png_bytep prev_row, int filter)
 | ||
| {
 | ||
| #if defined(PNG_DEBUG)
 | ||
|    char filtname[10];
 | ||
| #endif
 | ||
| 
 | ||
|    if (_mmx_supported == 2) {
 | ||
| #if !defined(PNG_1_0_X)
 | ||
|        /* this should have happened in png_init_mmx_flags() already */
 | ||
|        png_warning(png_ptr, "asm_flags may not have been initialized");
 | ||
| #endif
 | ||
|        png_mmx_support();
 | ||
|    }
 | ||
| 
 | ||
| #if defined(PNG_DEBUG)
 | ||
|    png_debug(1, "in png_read_filter_row (pnggccrd.c)\n");
 | ||
|    switch (filter)
 | ||
|    {
 | ||
|       case 0:
 | ||
|          png_snprintf(filtname, 10, "none");
 | ||
|          break;
 | ||
| 
 | ||
|       case 1:
 | ||
|          png_snprintf(filtname, 10, "sub-%s",
 | ||
| #ifdef PNG_MMX_READ_FILTER_SUB_SUPPORTED
 | ||
| #if !defined(PNG_1_0_X)
 | ||
|            ((png_ptr->asm_flags & PNG_ASM_FLAG_MMX_READ_FILTER_SUB) &&
 | ||
|             (row_info->pixel_depth >= png_ptr->mmx_bitdepth_threshold) &&
 | ||
|             (row_info->rowbytes >= png_ptr->mmx_rowbytes_threshold)) ? "MMX" : 
 | ||
| #else
 | ||
|            _mmx_supported
 | ||
| #endif
 | ||
| #endif
 | ||
|            "C");
 | ||
|          break;
 | ||
| 
 | ||
|       case 2:
 | ||
|          png_snprintf(filtname, 10, "up-%s",
 | ||
| #ifdef PNG_MMX_READ_FILTER_UP_SUPPORTED
 | ||
| #if !defined(PNG_1_0_X)
 | ||
|            ((png_ptr->asm_flags & PNG_ASM_FLAG_MMX_READ_FILTER_UP) &&
 | ||
|             (row_info->pixel_depth >= png_ptr->mmx_bitdepth_threshold) &&
 | ||
|             (row_info->rowbytes >= png_ptr->mmx_rowbytes_threshold)) ? "MMX" :
 | ||
| #else
 | ||
|            _mmx_supported
 | ||
| #endif
 | ||
| #endif
 | ||
|            "C");
 | ||
|          break;
 | ||
| 
 | ||
|       case 3:
 | ||
|          png_snprintf(filtname, 10, "avg-%s",
 | ||
| #ifdef PNG_MMX_READ_FILTER_AVG_SUPPORTED
 | ||
| #if !defined(PNG_1_0_X)
 | ||
|            ((png_ptr->asm_flags & PNG_ASM_FLAG_MMX_READ_FILTER_AVG) &&
 | ||
|             (row_info->pixel_depth >= png_ptr->mmx_bitdepth_threshold) &&
 | ||
|             (row_info->rowbytes >= png_ptr->mmx_rowbytes_threshold)) ? "MMX" : 
 | ||
| #else
 | ||
|            _mmx_supported
 | ||
| #endif
 | ||
| #endif
 | ||
|            "C");
 | ||
|          break;
 | ||
| 
 | ||
|       case 4:
 | ||
|          png_snprintf(filtname, 10, "paeth-%s",
 | ||
| #ifdef PNG_MMX_READ_FILTER_PAETH_SUPPORTED
 | ||
| #if defined(PNG_x86_64_USE_GOTPCREL) || defined(PNG_THREAD_UNSAFE_OK)
 | ||
| #if !defined(PNG_1_0_X)
 | ||
|            ((png_ptr->asm_flags & PNG_ASM_FLAG_MMX_READ_FILTER_PAETH) &&
 | ||
|             (row_info->pixel_depth >= png_ptr->mmx_bitdepth_threshold) &&
 | ||
|             (row_info->rowbytes >= png_ptr->mmx_rowbytes_threshold))
 | ||
| #else
 | ||
|            _mmx_supported
 | ||
| #endif
 | ||
|            ? "MMX" :
 | ||
| #endif /* PNG_x86_64_USE_GOTPCREL || PNG_THREAD_UNSAFE_OK */
 | ||
| #endif
 | ||
|              "C");
 | ||
|          break;
 | ||
| 
 | ||
|       default:
 | ||
|          png_snprintf(filtname, 10, "unknown");
 | ||
|          break;
 | ||
|    }
 | ||
|    png_debug2(2, "row_number=%ld, %s, ", png_ptr->row_number, filtname);
 | ||
|    //png_debug1(0, "png_ptr=%10p, ", png_ptr);
 | ||
|    //png_debug1(0, "asm_flags=0x%08lx, ", png_ptr->asm_flags);
 | ||
|    png_debug1(0, "row=%10p, ", row);
 | ||
|    png_debug2(0, "pixdepth=%d, bytes=%d, ", (int)row_info->pixel_depth,
 | ||
|       (int)((row_info->pixel_depth + 7) >> 3));
 | ||
|    png_debug1(0, "rowbytes=%ld\n", row_info->rowbytes);
 | ||
| #endif /* PNG_DEBUG */
 | ||
| 
 | ||
|    switch (filter)
 | ||
|    {
 | ||
|       case PNG_FILTER_VALUE_NONE:
 | ||
|          break;
 | ||
| 
 | ||
|       case PNG_FILTER_VALUE_SUB:
 | ||
| #ifdef PNG_MMX_READ_FILTER_SUB_SUPPORTED
 | ||
| #if !defined(PNG_1_0_X)
 | ||
|          if ((png_ptr->asm_flags & PNG_ASM_FLAG_MMX_READ_FILTER_SUB) &&
 | ||
|              (row_info->pixel_depth >= png_ptr->mmx_bitdepth_threshold) &&
 | ||
|              (row_info->rowbytes >= png_ptr->mmx_rowbytes_threshold))
 | ||
| #else
 | ||
|          if (_mmx_supported)
 | ||
| #endif
 | ||
|          {
 | ||
|             png_read_filter_row_mmx_sub(row_info, row);
 | ||
|          }
 | ||
|          else
 | ||
| #endif
 | ||
|          {
 | ||
|             png_uint_32 i;
 | ||
|             png_uint_32 istop = row_info->rowbytes;
 | ||
|             png_uint_32 bpp = (row_info->pixel_depth + 7) >> 3;
 | ||
|             png_bytep rp = row + bpp;
 | ||
|             png_bytep lp = row;
 | ||
| 
 | ||
|             for (i = bpp; i < istop; i++)
 | ||
|             {
 | ||
|                *rp = (png_byte)(((int)(*rp) + (int)(*lp++)) & 0xff);
 | ||
|                rp++;
 | ||
|             }
 | ||
|          }  /* end !UseMMX_sub */
 | ||
|          break;
 | ||
| 
 | ||
|       case PNG_FILTER_VALUE_UP:
 | ||
| #ifdef PNG_MMX_READ_FILTER_UP_SUPPORTED
 | ||
| #if !defined(PNG_1_0_X)
 | ||
|          if ((png_ptr->asm_flags & PNG_ASM_FLAG_MMX_READ_FILTER_UP) &&
 | ||
|              (row_info->pixel_depth >= png_ptr->mmx_bitdepth_threshold) &&
 | ||
|              (row_info->rowbytes >= png_ptr->mmx_rowbytes_threshold))
 | ||
| #else
 | ||
|          if (_mmx_supported)
 | ||
| #endif
 | ||
|          {
 | ||
|             png_read_filter_row_mmx_up(row_info, row, prev_row);
 | ||
|          }
 | ||
|           else
 | ||
| #endif
 | ||
|          {
 | ||
|             png_uint_32 i;
 | ||
|             png_uint_32 istop = row_info->rowbytes;
 | ||
|             png_bytep rp = row;
 | ||
|             png_bytep pp = prev_row;
 | ||
| 
 | ||
|             for (i = 0; i < istop; ++i)
 | ||
|             {
 | ||
|                *rp = (png_byte)(((int)(*rp) + (int)(*pp++)) & 0xff);
 | ||
|                rp++;
 | ||
|             }
 | ||
|          }  /* end !UseMMX_up */
 | ||
|          break;
 | ||
| 
 | ||
|       case PNG_FILTER_VALUE_AVG:
 | ||
| #ifdef PNG_MMX_READ_FILTER_AVG_SUPPORTED
 | ||
| #if !defined(PNG_1_0_X)
 | ||
|          if ((png_ptr->asm_flags & PNG_ASM_FLAG_MMX_READ_FILTER_AVG) &&
 | ||
|              (row_info->pixel_depth >= png_ptr->mmx_bitdepth_threshold) &&
 | ||
|              (row_info->rowbytes >= png_ptr->mmx_rowbytes_threshold))
 | ||
| #else
 | ||
|          if (_mmx_supported)
 | ||
| #endif
 | ||
|          {
 | ||
|             png_read_filter_row_mmx_avg(row_info, row, prev_row);
 | ||
|          }
 | ||
|          else
 | ||
| #endif
 | ||
|          {
 | ||
|             png_uint_32 i;
 | ||
|             png_bytep rp = row;
 | ||
|             png_bytep pp = prev_row;
 | ||
|             png_bytep lp = row;
 | ||
|             png_uint_32 bpp = (row_info->pixel_depth + 7) >> 3;
 | ||
|             png_uint_32 istop = row_info->rowbytes - bpp;
 | ||
| 
 | ||
|             for (i = 0; i < bpp; i++)
 | ||
|             {
 | ||
|                *rp = (png_byte)(((int)(*rp) +
 | ||
|                   ((int)(*pp++) >> 1)) & 0xff);
 | ||
|                rp++;
 | ||
|             }
 | ||
| 
 | ||
|             for (i = 0; i < istop; i++)
 | ||
|             {
 | ||
|                *rp = (png_byte)(((int)(*rp) +
 | ||
|                   ((int)(*pp++ + *lp++) >> 1)) & 0xff);
 | ||
|                rp++;
 | ||
|             }
 | ||
|          }  /* end !UseMMX_avg */
 | ||
|          break;
 | ||
| 
 | ||
|       case PNG_FILTER_VALUE_PAETH:
 | ||
| #ifdef PNG_MMX_READ_FILTER_PAETH_SUPPORTED
 | ||
| #if defined(PNG_x86_64_USE_GOTPCREL) || defined(PNG_THREAD_UNSAFE_OK)
 | ||
| #if !defined(PNG_1_0_X)
 | ||
|          if ((png_ptr->asm_flags & PNG_ASM_FLAG_MMX_READ_FILTER_PAETH) &&
 | ||
|              (row_info->pixel_depth >= png_ptr->mmx_bitdepth_threshold) &&
 | ||
|              (row_info->rowbytes >= png_ptr->mmx_rowbytes_threshold))
 | ||
| #else
 | ||
|          if (_mmx_supported)
 | ||
| #endif
 | ||
|          {
 | ||
|             png_read_filter_row_mmx_paeth(row_info, row, prev_row);
 | ||
|          }
 | ||
|          else
 | ||
| #endif /* PNG_x86_64_USE_GOTPCREL || PNG_THREAD_UNSAFE_OK */
 | ||
| #endif
 | ||
|          {
 | ||
|             png_uint_32 i;
 | ||
|             png_bytep rp = row;
 | ||
|             png_bytep pp = prev_row;
 | ||
|             png_bytep lp = row;
 | ||
|             png_bytep cp = prev_row;
 | ||
|             png_uint_32 bpp = (row_info->pixel_depth + 7) >> 3;
 | ||
|             png_uint_32 istop = row_info->rowbytes - bpp;
 | ||
| 
 | ||
|             for (i = 0; i < bpp; i++)
 | ||
|             {
 | ||
|                *rp = (png_byte)(((int)(*rp) + (int)(*pp++)) & 0xff);
 | ||
|                rp++;
 | ||
|             }
 | ||
| 
 | ||
|             for (i = 0; i < istop; i++)   /* use leftover rp,pp */
 | ||
|             {
 | ||
|                int a, b, c, pa, pb, pc, p;
 | ||
| 
 | ||
|                a = *lp++;
 | ||
|                b = *pp++;
 | ||
|                c = *cp++;
 | ||
| 
 | ||
|                p = b - c;
 | ||
|                pc = a - c;
 | ||
| 
 | ||
| #if defined(PNG_USE_ABS)
 | ||
|                pa = abs(p);
 | ||
|                pb = abs(pc);
 | ||
|                pc = abs(p + pc);
 | ||
| #else
 | ||
|                pa = p < 0 ? -p : p;
 | ||
|                pb = pc < 0 ? -pc : pc;
 | ||
|                pc = (p + pc) < 0 ? -(p + pc) : p + pc;
 | ||
| #endif
 | ||
| 
 | ||
|                /*
 | ||
|                   if (pa <= pb && pa <= pc)
 | ||
|                      p = a;
 | ||
|                   else if (pb <= pc)
 | ||
|                      p = b;
 | ||
|                   else
 | ||
|                      p = c;
 | ||
|                 */
 | ||
| 
 | ||
|                p = (pa <= pb && pa <= pc) ? a : (pb <= pc) ? b : c;
 | ||
| 
 | ||
|                *rp = (png_byte)(((int)(*rp) + p) & 0xff);
 | ||
|                rp++;
 | ||
|             }
 | ||
|          }  /* end !UseMMX_paeth */
 | ||
|          break;
 | ||
| 
 | ||
|       default:
 | ||
|          png_warning(png_ptr, "Ignoring bad row-filter type");
 | ||
|          *row=0;
 | ||
|          break;
 | ||
|    }
 | ||
| }
 | ||
| 
 | ||
| #endif /* PNG_HAVE_MMX_READ_FILTER_ROW */
 | ||
| 
 | ||
| 
 | ||
| #endif /* PNG_ASSEMBLER_CODE_SUPPORTED && PNG_USE_PNGGCCRD */
 | ||
| #endif /* __GNUC__ */
 | 
