diff options
author | Peter S. Mazinger <ps.m@gmx.net> | 2005-11-11 22:24:53 +0000 |
---|---|---|
committer | Peter S. Mazinger <ps.m@gmx.net> | 2005-11-11 22:24:53 +0000 |
commit | d0c2fc99e181270291f0f09da87d29609ce6d328 (patch) | |
tree | 085c270f1013331d790d3933a8bcd5ddb1103684 /libc | |
parent | 5ab11964a80debf091ff65ad15938b628ab9f597 (diff) |
Split up MSRC file, bzero left out
Diffstat (limited to 'libc')
-rw-r--r-- | libc/string/powerpc/Makefile.arch | 21 | ||||
-rw-r--r-- | libc/string/powerpc/memcpy.c | 80 | ||||
-rw-r--r-- | libc/string/powerpc/memmove.c | 76 | ||||
-rw-r--r-- | libc/string/powerpc/memset.c | 81 | ||||
-rw-r--r-- | libc/string/powerpc/string.c | 213 |
5 files changed, 241 insertions, 230 deletions
diff --git a/libc/string/powerpc/Makefile.arch b/libc/string/powerpc/Makefile.arch index 5d268093a..47a8dac1d 100644 --- a/libc/string/powerpc/Makefile.arch +++ b/libc/string/powerpc/Makefile.arch @@ -5,31 +5,18 @@ # Licensed under the LGPL v2.1, see the file COPYING.LIB in this tarball. # -MSRC:=string.c -MOBJ:=memcpy.o memmove.o memset.o bzero.o - STRING_ARCH_DIR:=$(top_srcdir)libc/string/powerpc STRING_ARCH_OUT:=$(top_builddir)libc/string/powerpc -STRING_ARCH_MSRC:=$(patsubst %.c,$(STRING_ARCH_DIR)/%.c,$(MSRC)) -STRING_ARCH_MOBJ:=$(patsubst %.o,$(STRING_ARCH_OUT)/%.o,$(MOBJ)) - -STRING_ARCH_OBJS:=$(STRING_ARCH_MOBJ) - -#STRING_ARCH_DEF:=$(patsubst %,-DL_%,$(subst .o,,$(notdir $(STRING_ARCH_MOBJ)))) - -$(STRING_ARCH_MOBJ): $(STRING_ARCH_MSRC) - $(compile.m) +STRING_ARCH_CSRC:=$(wildcard $(STRING_ARCH_DIR)/*.c) +STRING_ARCH_COBJ:=$(patsubst $(STRING_ARCH_DIR)/%.c,$(STRING_ARCH_OUT)/%.o,$(STRING_ARCH_CSRC)) -$(STRING_ARCH_MOBJ:.o=.os): $(STRING_ARCH_MSRC) - $(compile.m) +STRING_ARCH_OBJS:=$(STRING_ARCH_COBJ) libc-a-$(UCLIBC_HAS_STRING_ARCH_OPT)+=$(STRING_ARCH_OBJS) libc-so-$(UCLIBC_HAS_STRING_ARCH_OPT)+=$(STRING_ARCH_OBJS:.o=.os) -#CFLAGS-multi-$(UCLIBC_HAS_STRING_ARCH_OPT)+=$(STRING_ARCH_DEF) -#libc-multi-$(UCLIBC_HAS_STRING_ARCH_OPT)+=$(STRING_ARCH_MSRC) -libc-nomulti-$(UCLIBC_HAS_STRING_ARCH_OPT)+=$(STRING_ARCH_OBJS) +libc-multi-$(UCLIBC_HAS_STRING_ARCH_OPT)+=$(STRING_ARCH_CSRC) objclean-y+=string_arch_objclean diff --git a/libc/string/powerpc/memcpy.c b/libc/string/powerpc/memcpy.c new file mode 100644 index 000000000..e731f3578 --- /dev/null +++ b/libc/string/powerpc/memcpy.c @@ -0,0 +1,80 @@ +/* + * Copyright (C) 2004 Joakim Tjernlund + * Copyright (C) 2000-2005 Erik Andersen <andersen@uclibc.org> + * + * Licensed under the LGPL v2.1, see the file COPYING.LIB in this tarball. + */ + +/* These are carefully optimized mem*() functions for PPC written in C. + * Don't muck around with these function without checking the generated + * assmbler code. + * It is possible to optimize these significantly more by using specific + * data cache instructions(mainly dcbz). However that requires knownledge + * about the CPU's cache line size. + * + * BUG ALERT! + * The cache instructions on MPC8xx CPU's are buggy(they don't update + * the DAR register when causing a DTLB Miss/Error) and cannot be + * used on 8xx CPU's without a kernel patch to work around this + * problem. + */ + +#include <string.h> + +void attribute_hidden *__memcpy(void *to, const void *from, size_t n) +/* PPC can do pre increment and load/store, but not post increment and load/store. + Therefore use *++ptr instead of *ptr++. */ +{ + unsigned long rem, chunks, tmp1, tmp2; + unsigned char *tmp_to; + unsigned char *tmp_from = (unsigned char *)from; + + chunks = n / 8; + tmp_from -= 4; + tmp_to = to - 4; + if (!chunks) + goto lessthan8; + rem = (unsigned long )tmp_to % 4; + if (rem) + goto align; + copy_chunks: + do { + /* make gcc to load all data, then store it */ + tmp1 = *(unsigned long *)(tmp_from+4); + tmp_from += 8; + tmp2 = *(unsigned long *)tmp_from; + *(unsigned long *)(tmp_to+4) = tmp1; + tmp_to += 8; + *(unsigned long *)tmp_to = tmp2; + } while (--chunks); + lessthan8: + n = n % 8; + if (n >= 4) { + *(unsigned long *)(tmp_to+4) = *(unsigned long *)(tmp_from+4); + tmp_from += 4; + tmp_to += 4; + n = n-4; + } + if (!n ) return to; + tmp_from += 3; + tmp_to += 3; + do { + *++tmp_to = *++tmp_from; + } while (--n); + + return to; + align: + rem = 4 - rem; + n = n - rem; + do { + *(tmp_to+4) = *(tmp_from+4); + ++tmp_from; + ++tmp_to; + } while (--rem); + chunks = n / 8; + if (chunks) + goto copy_chunks; + goto lessthan8; +} + +strong_alias(__memcpy, memcpy) diff --git a/libc/string/powerpc/memmove.c b/libc/string/powerpc/memmove.c new file mode 100644 index 000000000..28b1da270 --- /dev/null +++ b/libc/string/powerpc/memmove.c @@ -0,0 +1,76 @@ +/* + * Copyright (C) 2004 Joakim Tjernlund + * Copyright (C) 2000-2005 Erik Andersen <andersen@uclibc.org> + * + * Licensed under the LGPL v2.1, see the file COPYING.LIB in this tarball. + */ + +/* These are carefully optimized mem*() functions for PPC written in C. + * Don't muck around with these function without checking the generated + * assmbler code. + * It is possible to optimize these significantly more by using specific + * data cache instructions(mainly dcbz). However that requires knownledge + * about the CPU's cache line size. + * + * BUG ALERT! + * The cache instructions on MPC8xx CPU's are buggy(they don't update + * the DAR register when causing a DTLB Miss/Error) and cannot be + * used on 8xx CPU's without a kernel patch to work around this + * problem. + */ + +#include <string.h> + +void attribute_hidden *__memmove(void *to, const void *from, size_t n) +{ + unsigned long rem, chunks, tmp1, tmp2; + unsigned char *tmp_to; + unsigned char *tmp_from = (unsigned char *)from; + + if (tmp_from >= (unsigned char *)to) + return __memcpy(to, from, n); + chunks = n / 8; + tmp_from += n; + tmp_to = to + n; + if (!chunks) + goto lessthan8; + rem = (unsigned long )tmp_to % 4; + if (rem) + goto align; + copy_chunks: + do { + /* make gcc to load all data, then store it */ + tmp1 = *(unsigned long *)(tmp_from-4); + tmp_from -= 8; + tmp2 = *(unsigned long *)tmp_from; + *(unsigned long *)(tmp_to-4) = tmp1; + tmp_to -= 8; + *(unsigned long *)tmp_to = tmp2; + } while (--chunks); + lessthan8: + n = n % 8; + if (n >= 4) { + *(unsigned long *)(tmp_to-4) = *(unsigned long *)(tmp_from-4); + tmp_from -= 4; + tmp_to -= 4; + n = n-4; + } + if (!n ) return to; + do { + *--tmp_to = *--tmp_from; + } while (--n); + + return to; + align: + rem = 4 - rem; + n = n - rem; + do { + *--tmp_to = *--tmp_from; + } while (--rem); + chunks = n / 8; + if (chunks) + goto copy_chunks; + goto lessthan8; +} + +strong_alias(__memmove, memmove) diff --git a/libc/string/powerpc/memset.c b/libc/string/powerpc/memset.c new file mode 100644 index 000000000..c5903afa6 --- /dev/null +++ b/libc/string/powerpc/memset.c @@ -0,0 +1,81 @@ +/* + * Copyright (C) 2004 Joakim Tjernlund + * Copyright (C) 2000-2005 Erik Andersen <andersen@uclibc.org> + * + * Licensed under the LGPL v2.1, see the file COPYING.LIB in this tarball. + */ + +/* These are carefully optimized mem*() functions for PPC written in C. + * Don't muck around with these function without checking the generated + * assmbler code. + * It is possible to optimize these significantly more by using specific + * data cache instructions(mainly dcbz). However that requires knownledge + * about the CPU's cache line size. + * + * BUG ALERT! + * The cache instructions on MPC8xx CPU's are buggy(they don't update + * the DAR register when causing a DTLB Miss/Error) and cannot be + * used on 8xx CPU's without a kernel patch to work around this + * problem. + */ + +#include <string.h> + +static inline int expand_byte_word(int c){ + /* this does: + c = c << 8 | c; + c = c << 16 | c ; + */ + asm("rlwimi %0,%0,8,16,23\n" + "\trlwimi %0,%0,16,0,15\n" + : "=r" (c) : "0" (c)); + return c; +} +void attribute_hidden *__memset(void *to, int c, size_t n) +{ + unsigned long rem, chunks; + unsigned char *tmp_to; + + chunks = n / 8; + tmp_to = to - 4; + c = expand_byte_word(c); + if (!chunks) + goto lessthan8; + rem = (unsigned long )tmp_to % 4; + if (rem) + goto align; + copy_chunks: + do { + *(unsigned long *)(tmp_to+4) = c; + tmp_to += 4; + *(unsigned long *)(tmp_to+4) = c; + tmp_to += 4; + } while (--chunks); + lessthan8: + n = n % 8; + if (n >= 4) { + *(unsigned long *)(tmp_to+4) = c; + tmp_to += 4; + n = n-4; + } + if (!n ) return to; + tmp_to += 3; + do { + *++tmp_to = c; + } while (--n); + + return to; + align: + rem = 4 - rem; + n = n-rem; + do { + *(tmp_to+4) = c; + ++tmp_to; + } while (--rem); + chunks = n / 8; + if (chunks) + goto copy_chunks; + goto lessthan8; +} + +strong_alias(__memset, memset) diff --git a/libc/string/powerpc/string.c b/libc/string/powerpc/string.c deleted file mode 100644 index d02deda08..000000000 --- a/libc/string/powerpc/string.c +++ /dev/null @@ -1,213 +0,0 @@ -/* - * This program is free software; you can redistribute it and/or modify it under - * the terms of the GNU Library General Public License as published by the Free - * Software Foundation; either version 2 of the License, or (at your option) any - * later version. - * - * This program is distributed in the hope that it will be useful, but WITHOUT - * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS - * FOR A PARTICULAR PURPOSE. See the GNU Library General Public License for more - * details. - */ - -/* These are carefully optimized mem*() functions for PPC written in C. - * Don't muck around with these function without checking the generated - * assmbler code. - * It is possible to optimize these significantly more by using specific - * data cache instructions(mainly dcbz). However that requires knownledge - * about the CPU's cache line size. - * - * BUG ALERT! - * The cache instructions on MPC8xx CPU's are buggy(they don't update - * the DAR register when causing a DTLB Miss/Error) and cannot be - * used on 8xx CPU's without a kernel patch to work around this - * problem. - * - * Copyright (C) 2004 Joakim Tjernlund - */ - -#define _GNU_SOURCE -#include <string.h> -#include <locale.h> /* for __LOCALE_C_ONLY */ - -#ifdef L_memcpy -void attribute_hidden *__memcpy(void *to, const void *from, size_t n) -/* PPC can do pre increment and load/store, but not post increment and load/store. - Therefore use *++ptr instead of *ptr++. */ -{ - unsigned long rem, chunks, tmp1, tmp2; - unsigned char *tmp_to; - unsigned char *tmp_from = (unsigned char *)from; - - chunks = n / 8; - tmp_from -= 4; - tmp_to = to - 4; - if (!chunks) - goto lessthan8; - rem = (unsigned long )tmp_to % 4; - if (rem) - goto align; - copy_chunks: - do { - /* make gcc to load all data, then store it */ - tmp1 = *(unsigned long *)(tmp_from+4); - tmp_from += 8; - tmp2 = *(unsigned long *)tmp_from; - *(unsigned long *)(tmp_to+4) = tmp1; - tmp_to += 8; - *(unsigned long *)tmp_to = tmp2; - } while (--chunks); - lessthan8: - n = n % 8; - if (n >= 4) { - *(unsigned long *)(tmp_to+4) = *(unsigned long *)(tmp_from+4); - tmp_from += 4; - tmp_to += 4; - n = n-4; - } - if (!n ) return to; - tmp_from += 3; - tmp_to += 3; - do { - *++tmp_to = *++tmp_from; - } while (--n); - - return to; - align: - rem = 4 - rem; - n = n - rem; - do { - *(tmp_to+4) = *(tmp_from+4); - ++tmp_from; - ++tmp_to; - } while (--rem); - chunks = n / 8; - if (chunks) - goto copy_chunks; - goto lessthan8; -} -strong_alias(__memcpy, memcpy); -#endif - -#ifdef L_memmove -void attribute_hidden *__memmove(void *to, const void *from, size_t n) -{ - unsigned long rem, chunks, tmp1, tmp2; - unsigned char *tmp_to; - unsigned char *tmp_from = (unsigned char *)from; - - if (tmp_from >= (unsigned char *)to) - return memcpy(to, from, n); - chunks = n / 8; - tmp_from += n; - tmp_to = to + n; - if (!chunks) - goto lessthan8; - rem = (unsigned long )tmp_to % 4; - if (rem) - goto align; - copy_chunks: - do { - /* make gcc to load all data, then store it */ - tmp1 = *(unsigned long *)(tmp_from-4); - tmp_from -= 8; - tmp2 = *(unsigned long *)tmp_from; - *(unsigned long *)(tmp_to-4) = tmp1; - tmp_to -= 8; - *(unsigned long *)tmp_to = tmp2; - } while (--chunks); - lessthan8: - n = n % 8; - if (n >= 4) { - *(unsigned long *)(tmp_to-4) = *(unsigned long *)(tmp_from-4); - tmp_from -= 4; - tmp_to -= 4; - n = n-4; - } - if (!n ) return to; - do { - *--tmp_to = *--tmp_from; - } while (--n); - - return to; - align: - rem = 4 - rem; - n = n - rem; - do { - *--tmp_to = *--tmp_from; - } while (--rem); - chunks = n / 8; - if (chunks) - goto copy_chunks; - goto lessthan8; -} -strong_alias(__memmove, memmove); -#endif - -#ifdef L_memset -static inline int expand_byte_word(int c){ - /* this does: - c = c << 8 | c; - c = c << 16 | c ; - */ - asm("rlwimi %0,%0,8,16,23\n" - "\trlwimi %0,%0,16,0,15\n" - : "=r" (c) : "0" (c)); - return c; -} -void attribute_hidden *__memset(void *to, int c, size_t n) -{ - unsigned long rem, chunks; - unsigned char *tmp_to; - - chunks = n / 8; - tmp_to = to - 4; - c = expand_byte_word(c); - if (!chunks) - goto lessthan8; - rem = (unsigned long )tmp_to % 4; - if (rem) - goto align; - copy_chunks: - do { - *(unsigned long *)(tmp_to+4) = c; - tmp_to += 4; - *(unsigned long *)(tmp_to+4) = c; - tmp_to += 4; - } while (--chunks); - lessthan8: - n = n % 8; - if (n >= 4) { - *(unsigned long *)(tmp_to+4) = c; - tmp_to += 4; - n = n-4; - } - if (!n ) return to; - tmp_to += 3; - do { - *++tmp_to = c; - } while (--n); - - return to; - align: - rem = 4 - rem; - n = n-rem; - do { - *(tmp_to+4) = c; - ++tmp_to; - } while (--rem); - chunks = n / 8; - if (chunks) - goto copy_chunks; - goto lessthan8; -} -strong_alias(__memset, memset); -#endif - -#ifdef L_bzero -weak_alias(__bzero,bzero); -void __bzero(void *s, size_t n) -{ - (void)memset(s, 0, n); -} -#endif |