StdLibExtras.cpp 3.4 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157
  1. #include <AK/Assertions.h>
  2. #include <AK/StdLibExtras.h>
  3. #include <AK/Types.h>
  4. #include <AK/kstdio.h>
  5. extern "C" {
  6. #if ARCH(I386)
  7. #ifndef KERNEL
  8. void* mmx_memcpy(void* dest, const void* src, size_t len)
  9. {
  10. ASSERT(len >= 1024);
  11. auto* dest_ptr = (u8*)dest;
  12. auto* src_ptr = (const u8*)src;
  13. if ((u32)dest_ptr & 7) {
  14. u32 prologue = 8 - ((u32)dest_ptr & 7);
  15. len -= prologue;
  16. asm volatile(
  17. "rep movsb\n"
  18. : "=S"(src_ptr), "=D"(dest_ptr), "=c"(prologue)
  19. : "0"(src_ptr), "1"(dest_ptr), "2"(prologue)
  20. : "memory");
  21. }
  22. for (u32 i = len / 64; i; --i) {
  23. asm volatile(
  24. "movq (%0), %%mm0\n"
  25. "movq 8(%0), %%mm1\n"
  26. "movq 16(%0), %%mm2\n"
  27. "movq 24(%0), %%mm3\n"
  28. "movq 32(%0), %%mm4\n"
  29. "movq 40(%0), %%mm5\n"
  30. "movq 48(%0), %%mm6\n"
  31. "movq 56(%0), %%mm7\n"
  32. "movq %%mm0, (%1)\n"
  33. "movq %%mm1, 8(%1)\n"
  34. "movq %%mm2, 16(%1)\n"
  35. "movq %%mm3, 24(%1)\n"
  36. "movq %%mm4, 32(%1)\n"
  37. "movq %%mm5, 40(%1)\n"
  38. "movq %%mm6, 48(%1)\n"
  39. "movq %%mm7, 56(%1)\n" ::"r"(src_ptr),
  40. "r"(dest_ptr)
  41. : "memory");
  42. src_ptr += 64;
  43. dest_ptr += 64;
  44. }
  45. asm volatile("emms" ::
  46. : "memory");
  47. // Whatever remains we'll have to memcpy.
  48. len %= 64;
  49. if (len)
  50. memcpy(dest_ptr, src_ptr, len);
  51. return dest;
  52. }
  53. #endif
  54. #endif
  55. #ifdef KERNEL
  56. static inline uint32_t divq(uint64_t n, uint32_t d)
  57. {
  58. uint32_t n1 = n >> 32;
  59. uint32_t n0 = n;
  60. uint32_t q;
  61. uint32_t r;
  62. asm volatile("divl %4"
  63. : "=d"(r), "=a"(q)
  64. : "0"(n1), "1"(n0), "rm"(d));
  65. return q;
  66. }
  67. static uint64_t unsigned_divide64(uint64_t n, uint64_t d)
  68. {
  69. if ((d >> 32) == 0) {
  70. uint64_t b = 1ULL << 32;
  71. uint32_t n1 = n >> 32;
  72. uint32_t n0 = n;
  73. uint32_t d0 = d;
  74. return divq(b * (n1 % d0) + n0, d0) + b * (n1 / d0);
  75. }
  76. if (n < d)
  77. return 0;
  78. uint32_t d1 = d >> 32u;
  79. int s = __builtin_clz(d1);
  80. uint64_t q = divq(n >> 1, (d << s) >> 32) >> (31 - s);
  81. return n - (q - 1) * d < d ? q - 1 : q;
  82. }
  83. static uint32_t unsigned_modulo64(uint64_t n, uint64_t d)
  84. {
  85. return n - d * unsigned_divide64(n, d);
  86. }
  87. static int64_t signed_divide64(int64_t n, int64_t d)
  88. {
  89. uint64_t n_abs = n >= 0 ? (uint64_t)n : -(uint64_t)n;
  90. uint64_t d_abs = d >= 0 ? (uint64_t)d : -(uint64_t)d;
  91. uint64_t q_abs = unsigned_divide64(n_abs, d_abs);
  92. return (n < 0) == (d < 0) ? (int64_t)q_abs : -(int64_t)q_abs;
  93. }
  94. static int32_t signed_modulo64(int64_t n, int64_t d)
  95. {
  96. return n - d * signed_divide64(n, d);
  97. }
  98. int64_t __divdi3(int64_t n, int64_t d)
  99. {
  100. return signed_divide64(n, d);
  101. }
  102. int64_t __moddi3(int64_t n, int64_t d)
  103. {
  104. return signed_modulo64(n, d);
  105. }
  106. uint64_t __udivdi3(uint64_t n, uint64_t d)
  107. {
  108. return unsigned_divide64(n, d);
  109. }
  110. uint64_t __umoddi3(uint64_t n, uint64_t d)
  111. {
  112. return unsigned_modulo64(n, d);
  113. }
  114. uint64_t __udivmoddi4(uint64_t n, uint64_t d, uint64_t* r)
  115. {
  116. uint64_t q = 0;
  117. uint64_t qbit = 1;
  118. if (!d)
  119. return 1 / ((unsigned)d);
  120. while ((int64_t)d >= 0) {
  121. d <<= 1;
  122. qbit <<= 1;
  123. }
  124. while (qbit) {
  125. if (d <= n) {
  126. n -= d;
  127. q += qbit;
  128. }
  129. d >>= 1;
  130. qbit >>= 1;
  131. }
  132. if (r)
  133. *r = n;
  134. return q;
  135. }
  136. #endif
  137. }