Forráskód Böngészése

Kernel+AK: Remove AK/StdLibExtras.cpp, moving kernel stuff to Kernel/.

We had some kernel-specific gizmos in AK that should really just be in the
Kernel subdirectory instead. The only thing remaining after moving those
was mmx_memcpy() which I moved to the ARCH(i386)-specific section of
LibC/string.cpp.
Andreas Kling 6 éve
szülő
commit
57c29491a3
6 módosított fájl, 146 hozzáadás és 161 törlés
  1. 0 157
      AK/StdLibExtras.cpp
  2. 2 2
      AK/StdLibExtras.h
  3. 0 1
      Kernel/Makefile
  4. 96 0
      Kernel/StdLib.cpp
  5. 0 1
      Libraries/LibC/Makefile
  6. 48 0
      Libraries/LibC/string.cpp

+ 0 - 157
AK/StdLibExtras.cpp

@@ -1,157 +0,0 @@
-#include <AK/Assertions.h>
-#include <AK/StdLibExtras.h>
-#include <AK/Types.h>
-#include <AK/kstdio.h>
-
-extern "C" {
-
-#if ARCH(I386)
-#ifndef KERNEL
-void* mmx_memcpy(void* dest, const void* src, size_t len)
-{
-    ASSERT(len >= 1024);
-
-    auto* dest_ptr = (u8*)dest;
-    auto* src_ptr = (const u8*)src;
-
-    if ((u32)dest_ptr & 7) {
-        u32 prologue = 8 - ((u32)dest_ptr & 7);
-        len -= prologue;
-        asm volatile(
-            "rep movsb\n"
-            : "=S"(src_ptr), "=D"(dest_ptr), "=c"(prologue)
-            : "0"(src_ptr), "1"(dest_ptr), "2"(prologue)
-            : "memory");
-    }
-    for (u32 i = len / 64; i; --i) {
-        asm volatile(
-            "movq (%0), %%mm0\n"
-            "movq 8(%0), %%mm1\n"
-            "movq 16(%0), %%mm2\n"
-            "movq 24(%0), %%mm3\n"
-            "movq 32(%0), %%mm4\n"
-            "movq 40(%0), %%mm5\n"
-            "movq 48(%0), %%mm6\n"
-            "movq 56(%0), %%mm7\n"
-            "movq %%mm0, (%1)\n"
-            "movq %%mm1, 8(%1)\n"
-            "movq %%mm2, 16(%1)\n"
-            "movq %%mm3, 24(%1)\n"
-            "movq %%mm4, 32(%1)\n"
-            "movq %%mm5, 40(%1)\n"
-            "movq %%mm6, 48(%1)\n"
-            "movq %%mm7, 56(%1)\n" ::"r"(src_ptr),
-            "r"(dest_ptr)
-            : "memory");
-        src_ptr += 64;
-        dest_ptr += 64;
-    }
-    asm volatile("emms" ::
-                     : "memory");
-    // Whatever remains we'll have to memcpy.
-    len %= 64;
-    if (len)
-        memcpy(dest_ptr, src_ptr, len);
-    return dest;
-}
-#endif
-#endif
-
-#ifdef KERNEL
-
-static inline uint32_t divq(uint64_t n, uint32_t d)
-{
-    uint32_t n1 = n >> 32;
-    uint32_t n0 = n;
-    uint32_t q;
-    uint32_t r;
-    asm volatile("divl %4"
-                 : "=d"(r), "=a"(q)
-                 : "0"(n1), "1"(n0), "rm"(d));
-    return q;
-}
-
-static uint64_t unsigned_divide64(uint64_t n, uint64_t d)
-{
-    if ((d >> 32) == 0) {
-        uint64_t b = 1ULL << 32;
-        uint32_t n1 = n >> 32;
-        uint32_t n0 = n;
-        uint32_t d0 = d;
-        return divq(b * (n1 % d0) + n0, d0) + b * (n1 / d0);
-    }
-    if (n < d)
-        return 0;
-    uint32_t d1 = d >> 32u;
-    int s = __builtin_clz(d1);
-    uint64_t q = divq(n >> 1, (d << s) >> 32) >> (31 - s);
-    return n - (q - 1) * d < d ? q - 1 : q;
-}
-
-static uint32_t unsigned_modulo64(uint64_t n, uint64_t d)
-{
-    return n - d * unsigned_divide64(n, d);
-}
-
-static int64_t signed_divide64(int64_t n, int64_t d)
-{
-    uint64_t n_abs = n >= 0 ? (uint64_t)n : -(uint64_t)n;
-    uint64_t d_abs = d >= 0 ? (uint64_t)d : -(uint64_t)d;
-    uint64_t q_abs = unsigned_divide64(n_abs, d_abs);
-    return (n < 0) == (d < 0) ? (int64_t)q_abs : -(int64_t)q_abs;
-}
-
-static int32_t signed_modulo64(int64_t n, int64_t d)
-{
-    return n - d * signed_divide64(n, d);
-}
-
-int64_t __divdi3(int64_t n, int64_t d)
-{
-    return signed_divide64(n, d);
-}
-
-int64_t __moddi3(int64_t n, int64_t d)
-{
-    return signed_modulo64(n, d);
-}
-
-uint64_t __udivdi3(uint64_t n, uint64_t d)
-{
-    return unsigned_divide64(n, d);
-}
-
-uint64_t __umoddi3(uint64_t n, uint64_t d)
-{
-    return unsigned_modulo64(n, d);
-}
-
-uint64_t __udivmoddi4(uint64_t n, uint64_t d, uint64_t* r)
-{
-    uint64_t q = 0;
-    uint64_t qbit = 1;
-
-    if (!d)
-        return 1 / ((unsigned)d);
-
-    while ((int64_t)d >= 0) {
-        d <<= 1;
-        qbit <<= 1;
-    }
-
-    while (qbit) {
-        if (d <= n) {
-            n -= d;
-            q += qbit;
-        }
-        d >>= 1;
-        qbit >>= 1;
-    }
-
-    if (r)
-        *r = n;
-
-    return q;
-}
-#endif
-}

+ 2 - 2
AK/StdLibExtras.h

@@ -11,13 +11,13 @@
 
 #include <AK/Types.h>
 
-#ifndef KERNEL
+#if defined(__serenity__) && !defined(KERNEL)
 extern "C" void* mmx_memcpy(void* to, const void* from, size_t);
 #endif
 
 [[gnu::always_inline]] inline void fast_u32_copy(u32* dest, const u32* src, size_t count)
 {
-#ifndef KERNEL
+#if defined(__serenity__) && !defined(KERNEL)
     if (count >= 256) {
         mmx_memcpy(dest, src, count * sizeof(count));
         return;

+ 0 - 1
Kernel/Makefile

@@ -89,7 +89,6 @@ AK_OBJS = \
     ../AK/StringBuilder.o \
     ../AK/StringView.o \
     ../AK/FileSystemPath.o \
-    ../AK/StdLibExtras.o \
     ../AK/JsonObject.o \
     ../AK/JsonValue.o \
     ../AK/JsonArray.o \

+ 96 - 0
Kernel/StdLib.cpp

@@ -134,4 +134,100 @@ int memcmp(const void* v1, const void* v2, size_t n)
 {
     ASSERT_NOT_REACHED();
 }
+
+static inline uint32_t divq(uint64_t n, uint32_t d)
+{
+    uint32_t n1 = n >> 32;
+    uint32_t n0 = n;
+    uint32_t q;
+    uint32_t r;
+    asm volatile("divl %4"
+                 : "=d"(r), "=a"(q)
+                 : "0"(n1), "1"(n0), "rm"(d));
+    return q;
+}
+
+static uint64_t unsigned_divide64(uint64_t n, uint64_t d)
+{
+    if ((d >> 32) == 0) {
+        uint64_t b = 1ULL << 32;
+        uint32_t n1 = n >> 32;
+        uint32_t n0 = n;
+        uint32_t d0 = d;
+        return divq(b * (n1 % d0) + n0, d0) + b * (n1 / d0);
+    }
+    if (n < d)
+        return 0;
+    uint32_t d1 = d >> 32u;
+    int s = __builtin_clz(d1);
+    uint64_t q = divq(n >> 1, (d << s) >> 32) >> (31 - s);
+    return n - (q - 1) * d < d ? q - 1 : q;
+}
+
+static uint32_t unsigned_modulo64(uint64_t n, uint64_t d)
+{
+    return n - d * unsigned_divide64(n, d);
+}
+
+static int64_t signed_divide64(int64_t n, int64_t d)
+{
+    uint64_t n_abs = n >= 0 ? (uint64_t)n : -(uint64_t)n;
+    uint64_t d_abs = d >= 0 ? (uint64_t)d : -(uint64_t)d;
+    uint64_t q_abs = unsigned_divide64(n_abs, d_abs);
+    return (n < 0) == (d < 0) ? (int64_t)q_abs : -(int64_t)q_abs;
+}
+
+static int32_t signed_modulo64(int64_t n, int64_t d)
+{
+    return n - d * signed_divide64(n, d);
+}
+
+int64_t __divdi3(int64_t n, int64_t d)
+{
+    return signed_divide64(n, d);
+}
+
+int64_t __moddi3(int64_t n, int64_t d)
+{
+    return signed_modulo64(n, d);
+}
+
+uint64_t __udivdi3(uint64_t n, uint64_t d)
+{
+    return unsigned_divide64(n, d);
+}
+
+uint64_t __umoddi3(uint64_t n, uint64_t d)
+{
+    return unsigned_modulo64(n, d);
+}
+
+uint64_t __udivmoddi4(uint64_t n, uint64_t d, uint64_t* r)
+{
+    uint64_t q = 0;
+    uint64_t qbit = 1;
+
+    if (!d)
+        return 1 / ((unsigned)d);
+
+    while ((int64_t)d >= 0) {
+        d <<= 1;
+        qbit <<= 1;
+    }
+
+    while (qbit) {
+        if (d <= n) {
+            n -= d;
+            q += qbit;
+        }
+        d >>= 1;
+        qbit >>= 1;
+    }
+
+    if (r)
+        *r = n;
+
+    return q;
+}
+
 }

+ 0 - 1
Libraries/LibC/Makefile

@@ -6,7 +6,6 @@ AK_OBJS = \
     ../../AK/StringView.o \
     ../../AK/StringBuilder.o \
     ../../AK/FileSystemPath.o \
-    ../../AK/StdLibExtras.o \
     ../../AK/JsonValue.o \
     ../../AK/JsonArray.o \
     ../../AK/JsonObject.o \

+ 48 - 0
Libraries/LibC/string.cpp

@@ -134,6 +134,54 @@ int memcmp(const void* v1, const void* v2, size_t n)
 }
 
 #if ARCH(I386)
+void* mmx_memcpy(void* dest, const void* src, size_t len)
+{
+    ASSERT(len >= 1024);
+
+    auto* dest_ptr = (u8*)dest;
+    auto* src_ptr = (const u8*)src;
+
+    if ((u32)dest_ptr & 7) {
+        u32 prologue = 8 - ((u32)dest_ptr & 7);
+        len -= prologue;
+        asm volatile(
+            "rep movsb\n"
+            : "=S"(src_ptr), "=D"(dest_ptr), "=c"(prologue)
+            : "0"(src_ptr), "1"(dest_ptr), "2"(prologue)
+            : "memory");
+    }
+    for (u32 i = len / 64; i; --i) {
+        asm volatile(
+            "movq (%0), %%mm0\n"
+            "movq 8(%0), %%mm1\n"
+            "movq 16(%0), %%mm2\n"
+            "movq 24(%0), %%mm3\n"
+            "movq 32(%0), %%mm4\n"
+            "movq 40(%0), %%mm5\n"
+            "movq 48(%0), %%mm6\n"
+            "movq 56(%0), %%mm7\n"
+            "movq %%mm0, (%1)\n"
+            "movq %%mm1, 8(%1)\n"
+            "movq %%mm2, 16(%1)\n"
+            "movq %%mm3, 24(%1)\n"
+            "movq %%mm4, 32(%1)\n"
+            "movq %%mm5, 40(%1)\n"
+            "movq %%mm6, 48(%1)\n"
+            "movq %%mm7, 56(%1)\n" ::"r"(src_ptr),
+            "r"(dest_ptr)
+            : "memory");
+        src_ptr += 64;
+        dest_ptr += 64;
+    }
+    asm volatile("emms" ::
+                     : "memory");
+    // Whatever remains we'll have to memcpy.
+    len %= 64;
+    if (len)
+        memcpy(dest_ptr, src_ptr, len);
+    return dest;
+}
+
 void* memcpy(void* dest_ptr, const void* src_ptr, size_t n)
 {
     if (n >= 1024)