From 89c9ff504b29d03cf266aa3598ccb8d2cf1ddd13 Mon Sep 17 00:00:00 2001 From: Diego Biurrun Date: Wed, 28 Jan 2009 00:16:05 +0000 Subject: spelling/grammar/consistency review part I Originally committed as revision 16840 to svn://svn.ffmpeg.org/ffmpeg/trunk --- libavutil/mem.c | 16 ++++++++-------- 1 file changed, 8 insertions(+), 8 deletions(-) (limited to 'libavutil/mem.c') diff --git a/libavutil/mem.c b/libavutil/mem.c index 251b1fadab..9268ab4dfc 100644 --- a/libavutil/mem.c +++ b/libavutil/mem.c @@ -21,7 +21,7 @@ /** * @file mem.c - * default memory allocator for libavutil. + * default memory allocator for libavutil */ #include "config.h" @@ -35,14 +35,14 @@ #include "mem.h" -/* here we can use OS dependent allocation functions */ +/* here we can use OS-dependent allocation functions */ #undef free #undef malloc #undef realloc -/* you can redefine av_malloc and av_free in your project to use your +/* You can redefine av_malloc and av_free in your project to use your memory allocator. You do not need to suppress this file because the - linker will do it automatically */ + linker will do it automatically. */ void *av_malloc(unsigned int size) { @@ -70,7 +70,7 @@ void *av_malloc(unsigned int size) Indeed, we should align it: on 4 for 386 on 16 for 486 - on 32 for 586, PPro - k6-III + on 32 for 586, PPro - K6-III on 64 for K7 (maybe for P3 too). Because L1 and L2 caches are aligned on those values. But I don't want to code such logic here! @@ -78,10 +78,10 @@ void *av_malloc(unsigned int size) /* Why 16? Because some CPUs need alignment, for example SSE2 on P4, & most RISC CPUs it will just trigger an exception and the unaligned load will be done in the - exception handler or it will just segfault (SSE2 on P4) + exception handler or it will just segfault (SSE2 on P4). Why not larger? Because I did not see a difference in benchmarks ... */ - /* benchmarks with p3 + /* benchmarks with P3 memalign(64)+1 3071,3051,3032 memalign(64)+2 3051,3032,3041 memalign(64)+4 2911,2896,2915 @@ -90,7 +90,7 @@ void *av_malloc(unsigned int size) memalign(64)+32 2546,2545,2571 memalign(64)+64 2570,2533,2558 - btw, malloc seems to do 8 byte alignment by default here + BTW, malloc seems to do 8-byte alignment by default here. */ #else ptr = malloc(size); -- cgit v1.2.3