2362N/A * Copyright (c) 2000, 2003, Oracle and/or its affiliates. All rights reserved. 0N/A * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. 0N/A * This code is free software; you can redistribute it and/or modify it 0N/A * under the terms of the GNU General Public License version 2 only, as 2362N/A * published by the Free Software Foundation. Oracle designates this 0N/A * particular file as subject to the "Classpath" exception as provided 2362N/A * by Oracle in the LICENSE file that accompanied this code. 0N/A * This code is distributed in the hope that it will be useful, but WITHOUT 0N/A * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or 0N/A * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License 0N/A * version 2 for more details (a copy is included in the LICENSE file that 0N/A * accompanied this code). 0N/A * You should have received a copy of the GNU General Public License version 0N/A * 2 along with this work; if not, write to the Free Software Foundation, 0N/A * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. 2362N/A * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA 0N/A * Internal functions for mlib_ImageConv* on U8/S16/U16 types and 0N/A * MLIB_EDGE_DST_NO_WRITE mask 0N/A This define switches between functions of different data types 0N/A/***************************************************************/ 0N/A#
endif /* IMG_TYPE == 1 */ 0N/A/***************************************************************/ 0N/A/***************************************************************/ 0N/A#
endif /* MLIB_USE_FTOI_CLAMPING */ 0N/A/***************************************************************/ 0N/A/***************************************************************/ 0N/A#
endif /* _LITTLE_ENDIAN */ 0N/A/***************************************************************/ 0N/A#
else /* _NO_LONGLONG */ 0N/A#
else /* _LITTLE_ENDIAN */ 0N/A#
endif /* _LITTLE_ENDIAN */ 0N/A#
endif /* _NO_LONGLONG */ 0N/A/***************************************************************/ 0N/A/***************************************************************/ 0N/A/***************************************************************/ 0N/A/***************************************************************/ 0N/A /* keep kernel in regs */ \
0N/A/***************************************************************/ 0N/A/***************************************************************/ 0N/A/***************************************************************/ 0N/A/* Test for the presence of any "1" bit in bits 0N/A 8 to 31 of val. If present, then val is either 0N/A are uncommon, then this technique can be a win, 0N/A since only a single test, rather than two, is 0N/A necessary to determine if clamping is needed. 0N/A it adds an extra test. 0N/A#
endif /* IMG_TYPE == 1 */ 0N/A/***************************************************************/ 0N/A#
endif /* __SUNPRO_C */ 0N/A#
endif /* __SUNPRO_C */ 0N/A for (i = 0; i <= (
wid -
2); i +=
2) {
0N/A#
else /* _NO_LONGLONG */ 0N/A#
endif /* _NO_LONGLONG */ 0N/A#
endif /* IMG_TYPE != 1 */ 0N/A#
else /* _NO_LONGLONG */ 0N/A#
endif /* IMG_TYPE != 1 */ 0N/A#
endif /* _NO_LONGLONG */ 0N/A#
endif /* IMG_TYPE != 1 */ 0N/A#
endif /* IMG_TYPE == 1 */ 0N/A/***************************************************************/ 0N/A#
ifndef __sparc /* for x86, using integer multiplies is faster */ 0N/A#
endif /* IMG_TYPE != 1 */ 0N/A /* keep kernel in regs */ 0N/A#
endif /* __SUNPRO_C */ 0N/A for (i = 0; i <= (
wid -
2); i +=
2) {
0N/A#
endif /* __sparc ( for x86, using integer multiplies is faster ) */ 0N/A/***************************************************************/ 0N/A#
endif /* __SUNPRO_C */ 0N/A * First loop on two first lines of kernel 0N/A#
endif /* __SUNPRO_C */ 0N/A for (i = 0; i <= (
wid -
2); i +=
2) {
0N/A * Second loop on two last lines of kernel 0N/A#
endif /* __SUNPRO_C */ 0N/A for (i = 0; i <= (
wid -
2); i +=
2) {
0N/A/***************************************************************/ 0N/A#
endif /* __SUNPRO_C */ 0N/A#
endif /* __SUNPRO_C */ 0N/A for (i = 0; i <= (
wid -
2); i +=
2) {
0N/A#
endif /* __SUNPRO_C */ 0N/A for (i = 0; i <= (
wid -
2); i +=
2) {
0N/A#
endif /* __SUNPRO_C */ 0N/A for (i = 0; i <= (
wid -
2); i +=
2) {
0N/A/***************************************************************/ 0N/A#
ifndef __sparc /* for x86, using integer multiplies is faster */ 0N/A#
endif /* IMG_TYPE != 1 */ 0N/A#
endif /* __SUNPRO_C */ 0N/A for (i = 0; i <= (
wid -
2); i +=
2) {
0N/A#
endif /* __SUNPRO_C */ 0N/A for (i = 0; i <= (
wid -
2); i +=
2) {
0N/A#
endif /* __SUNPRO_C */ 0N/A for (i = 0; i <= (
wid -
2); i +=
2) {
0N/A#
endif /* __sparc ( for x86, using integer multiplies is faster ) */ 0N/A/***************************************************************/ 0N/A#
endif /* __SUNPRO_C */ 0N/A#
endif /* __SUNPRO_C */ 0N/A#
endif /* __SUNPRO_C */ 0N/A for (i = 0; i <= (
wid -
2); i +=
2) {
0N/A#
endif /* __SUNPRO_C */ 0N/A for (i = 0; i <= (
wid -
2); i +=
2) {
0N/A#
endif /* IMG_TYPE == 1 */ 0N/A/***************************************************************/ 0N/A#
endif /* __SUNPRO_C */ 0N/A#
endif /* __SUNPRO_C */ 0N/A#
endif /* __SUNPRO_C */ 0N/A#
endif /* __SUNPRO_C */ 0N/A#
endif /* __SUNPRO_C */ 0N/A }
else /* if (kh == 1) */ {
0N/A#
endif /* __SUNPRO_C */ 0N/A/***************************************************************/ 0N/A for (i = 0; i <
mn; i++) {
0N/A for (l = 0; l < n; l++) {
0N/A#
endif /* __SUNPRO_C */ 0N/A for (i = 0; i <
wid + (m -
1); i++) {
0N/A#
endif /* __SUNPRO_C */ 0N/A for (l = 0; l < n; l++) {
0N/A#
endif /* __SUNPRO_C */ 0N/A for (i = 0; i <= (
wid -
2); i +=
2) {
0N/A#
endif /* __SUNPRO_C */ 0N/A for (i = 0; i <= (
wid -
2); i +=
2) {
0N/A#
endif /* __SUNPRO_C */ 0N/A for (i = 0; i <= (
wid -
2); i +=
2) {
0N/A#
endif /* __SUNPRO_C */ 0N/A for (i = 0; i <= (
wid -
2); i +=
2) {
0N/A#
endif /* __SUNPRO_C */ 0N/A for (i = 0; i <= (
wid -
2); i +=
2) {
0N/A#
endif /* __SUNPRO_C */ 0N/A for (i = 0; i <= (
wid -
2); i +=
2) {
0N/A#
endif /* __SUNPRO_C */ 0N/A for (i = 0; i <= (
wid -
2); i +=
2) {
0N/A#
endif /* __SUNPRO_C */ 0N/A for (i = 0; i <= (
wid -
2); i +=
2) {
0N/A#
endif /* __SUNPRO_C */ 0N/A for (i = 0; i <= (
wid -
2); i +=
2) {
0N/A#
endif /* __SUNPRO_C */ 0N/A for (i = 0; i <= (
wid -
2); i +=
2) {
0N/A }
else /*if (kw == 2)*/ {
0N/A#
endif /* __SUNPRO_C */ 0N/A for (i = 0; i <= (
wid -
2); i +=
2) {
0N/A#
endif /* __SUNPRO_C */ 0N/A for (i = 0; i <= (
wid -
2); i +=
2) {
0N/A for (l = 0; l < n; l++) {
0N/A for (x = 0; x < m; x++) s +=
buff[x] * (*
pk++);
0N/A/***************************************************************/ 0N/A#
ifndef __sparc /* for x86, using integer multiplies is faster */ 0N/A#
endif /* IMG_TYPE != 1 */ 0N/A for (i = 0; i < m*n; i++) {
0N/A#
endif /* __SUNPRO_C */ 0N/A for (l = 0; l < n; l++) {
0N/A#
endif /* __SUNPRO_C */ 0N/A for (i = 0; i <= (
wid -
2); i +=
2) {
0N/A#
endif /* __SUNPRO_C */ 0N/A for (i = 0; i <= (
wid -
2); i +=
2) {
0N/A#
endif /* __SUNPRO_C */ 0N/A for (i = 0; i <= (
wid -
2); i +=
2) {
0N/A#
endif /* __SUNPRO_C */ 0N/A for (i = 0; i <= (
wid -
2); i +=
2) {
0N/A#
endif /* __SUNPRO_C */ 0N/A for (i = 0; i <= (
wid -
2); i +=
2) {
0N/A#
endif /* __SUNPRO_C */ 0N/A for (i = 0; i <= (
wid -
2); i +=
2) {
0N/A#
endif /* __SUNPRO_C */ 0N/A for (i = 0; i <= (
wid -
2); i +=
2) {
0N/A#
endif /* __SUNPRO_C */ 0N/A for (i = 0; i <= (
wid -
2); i +=
2) {
0N/A#
endif /* __SUNPRO_C */ 0N/A for (i = 0; i <= (
wid -
2); i +=
2) {
0N/A#
endif /* __SUNPRO_C */ 0N/A for (i = 0; i <= (
wid -
2); i +=
2) {
0N/A#
endif /* __SUNPRO_C */ 0N/A for (i = 0; i <= (
wid -
2); i +=
2) {
0N/A#
endif /* __SUNPRO_C */ 0N/A for (i = 0; i <= (
wid -
2); i +=
2) {
0N/A }
else /*if (kw == 1)*/ {
0N/A#
endif /* __SUNPRO_C */ 0N/A for (i = 0; i <= (
wid -
2); i +=
2) {
0N/A#
endif /* __SUNPRO_C */ 0N/A for (i = 0; i <= (
wid -
2); i +=
2) {
0N/A for (l = 0; l < n; l++) {
0N/A for (x = 0; x < m; x++) {
0N/A/***************************************************************/ 0N/A#
endif /* __sparc ( for x86, using integer multiplies is faster ) */ 0N/A/***************************************************************/