2020#include "random.h"
2121#include "util.h"
2222
23+ #if CONFIG_BLOCK_OPS_CHECK_SIZE && !defined(HAS_ARM_MTE )
24+ #include "musl.h"
25+ #endif
26+
2327#ifdef USE_PKEY
2428#include <sys/mman.h>
2529#endif
@@ -528,7 +532,7 @@ static void set_canary(UNUSED const struct slab_metadata *metadata, UNUSED void
528532 }
529533#endif
530534
531- memcpy ((char * )p + size - canary_size , & metadata -> canary_value , canary_size );
535+ h_memcpy_internal ((char * )p + size - canary_size , & metadata -> canary_value , canary_size );
532536#endif
533537}
534538
@@ -541,7 +545,7 @@ static void check_canary(UNUSED const struct slab_metadata *metadata, UNUSED con
541545#endif
542546
543547 u64 canary_value ;
544- memcpy (& canary_value , (const char * )p + size - canary_size , canary_size );
548+ h_memcpy_internal (& canary_value , (const char * )p + size - canary_size , canary_size );
545549
546550#ifdef HAS_ARM_MTE
547551 if (unlikely (canary_value == 0 )) {
@@ -831,7 +835,7 @@ static inline void deallocate_small(void *p, const size_t *expected_size) {
831835#endif
832836
833837 if (ZERO_ON_FREE && !skip_zero ) {
834- memset (p , 0 , size - canary_size );
838+ h_memset_internal (p , 0 , size - canary_size );
835839 }
836840 }
837841
@@ -1502,7 +1506,7 @@ EXPORT void *h_calloc(size_t nmemb, size_t size) {
15021506 total_size = adjust_size_for_canary (total_size );
15031507 void * p = alloc (total_size );
15041508 if (!ZERO_ON_FREE && likely (p != NULL ) && total_size && total_size <= max_slab_size_class ) {
1505- memset (p , 0 , total_size - canary_size );
1509+ h_memset_internal (p , 0 , total_size - canary_size );
15061510 }
15071511#ifdef HAS_ARM_MTE
15081512 // use an assert instead of adding a conditional to memset() above (freed memory is always
@@ -1624,7 +1628,7 @@ EXPORT void *h_realloc(void *old, size_t size) {
16241628 mutex_unlock (& ra -> lock );
16251629
16261630 if (memory_remap_fixed (old , old_size , new , size )) {
1627- memcpy (new , old , copy_size );
1631+ h_memcpy_internal (new , old , copy_size );
16281632 deallocate_pages (old , old_size , old_guard_size );
16291633 } else {
16301634 memory_unmap ((char * )old - old_guard_size , old_guard_size );
@@ -1646,7 +1650,7 @@ EXPORT void *h_realloc(void *old, size_t size) {
16461650 if (copy_size > 0 && copy_size <= max_slab_size_class ) {
16471651 copy_size -= canary_size ;
16481652 }
1649- memcpy (new , old_orig , copy_size );
1653+ h_memcpy_internal (new , old_orig , copy_size );
16501654 if (old_size <= max_slab_size_class ) {
16511655 deallocate_small (old , NULL );
16521656 } else {
@@ -1874,6 +1878,88 @@ EXPORT size_t h_malloc_object_size_fast(const void *p) {
18741878 return SIZE_MAX ;
18751879}
18761880
1881+ #if CONFIG_BLOCK_OPS_CHECK_SIZE && !defined(HAS_ARM_MTE )
1882+ EXPORT void * memcpy (void * restrict dst , const void * restrict src , size_t len ) {
1883+ if (unlikely (dst == src || len == 0 )) {
1884+ return dst ;
1885+ }
1886+ if (unlikely (dst < (src + len ) && (dst + len ) > src )) {
1887+ fatal_error ("memcpy overlap" );
1888+ }
1889+ if (unlikely (len > malloc_object_size (src ))) {
1890+ fatal_error ("memcpy read overflow" );
1891+ }
1892+ if (unlikely (len > malloc_object_size (dst ))) {
1893+ fatal_error ("memcpy buffer overflow" );
1894+ }
1895+ return musl_memcpy (dst , src , len );
1896+ }
1897+
1898+ EXPORT void * memmove (void * dst , const void * src , size_t len ) {
1899+ if (unlikely (dst == src || len == 0 )) {
1900+ return dst ;
1901+ }
1902+ if (unlikely (len > malloc_object_size (src ))) {
1903+ fatal_error ("memmove read overflow" );
1904+ }
1905+ if (unlikely (len > malloc_object_size (dst ))) {
1906+ fatal_error ("memmove buffer overflow" );
1907+ }
1908+ return musl_memmove (dst , src , len );
1909+ }
1910+
1911+ EXPORT void * memset (void * dst , int value , size_t len ) {
1912+ if (unlikely (len == 0 )) {
1913+ return dst ;
1914+ }
1915+ if (unlikely (len > malloc_object_size (dst ))) {
1916+ fatal_error ("memset buffer overflow" );
1917+ }
1918+ return musl_memset (dst , value , len );
1919+ }
1920+
1921+ EXPORT wchar_t * wmemcpy (wchar_t * restrict dst , const wchar_t * restrict src , size_t len ) {
1922+ if (unlikely (dst == src || len == 0 )) {
1923+ return dst ;
1924+ }
1925+ if (unlikely (dst < (src + len ) && (dst + len ) > src )) {
1926+ fatal_error ("wmemcpy overlap" );
1927+ }
1928+ size_t lenAdj = len * sizeof (wchar_t );
1929+ if (unlikely (lenAdj > malloc_object_size (src ))) {
1930+ fatal_error ("wmemcpy read overflow" );
1931+ }
1932+ if (unlikely (lenAdj > malloc_object_size (dst ))) {
1933+ fatal_error ("wmemcpy buffer overflow" );
1934+ }
1935+ return (wchar_t * )musl_memcpy ((char * )dst , (const char * )src , lenAdj );
1936+ }
1937+
1938+ EXPORT wchar_t * wmemmove (wchar_t * dst , const wchar_t * src , size_t len ) {
1939+ if (unlikely (dst == src || len == 0 )) {
1940+ return dst ;
1941+ }
1942+ size_t lenAdj = len * sizeof (wchar_t );
1943+ if (unlikely (lenAdj > malloc_object_size (src ))) {
1944+ fatal_error ("wmemmove read overflow" );
1945+ }
1946+ if (unlikely (lenAdj > malloc_object_size (dst ))) {
1947+ fatal_error ("wmemmove buffer overflow" );
1948+ }
1949+ return (wchar_t * )musl_memmove ((char * )dst , (const char * )src , lenAdj );
1950+ }
1951+
1952+ EXPORT wchar_t * wmemset (wchar_t * dst , wchar_t value , size_t len ) {
1953+ if (unlikely (len == 0 )) {
1954+ return dst ;
1955+ }
1956+ if (unlikely ((len * sizeof (wchar_t )) > malloc_object_size (dst ))) {
1957+ fatal_error ("wmemset buffer overflow" );
1958+ }
1959+ return musl_wmemset (dst , value , len );
1960+ }
1961+ #endif /* CONFIG_BLOCK_OPS_CHECK_SIZE && !defined(HAS_ARM_MTE) */
1962+
18771963EXPORT int h_mallopt (UNUSED int param , UNUSED int value ) {
18781964#ifdef __ANDROID__
18791965 if (param == M_PURGE ) {
0 commit comments