ladybird/Kernel/Task.cpp

776 lines
19 KiB
C++
Raw Normal View History

#include "types.h"
#include "Task.h"
#include "kmalloc.h"
#include "VGA.h"
#include "StdLib.h"
#include "i386.h"
#include "system.h"
2018-10-18 08:28:09 +00:00
#include <VirtualFileSystem/FileHandle.h>
#include <VirtualFileSystem/VirtualFileSystem.h>
#include <ELFLoader/ExecSpace.h>
#include "MemoryManager.h"
//#define DEBUG_IO
//#define TASK_DEBUG
static const DWORD defaultStackSize = 16384;
Task* current;
Task* s_kernelTask;
static pid_t next_pid;
2018-10-17 08:55:43 +00:00
static InlineLinkedList<Task>* s_tasks;
static InlineLinkedList<Task>* s_deadTasks;
static bool contextSwitch(Task*);
static void redoKernelTaskTSS()
{
if (!s_kernelTask->selector())
s_kernelTask->setSelector(allocateGDTEntry());
auto& tssDescriptor = getGDTEntry(s_kernelTask->selector());
tssDescriptor.setBase(&s_kernelTask->tss());
tssDescriptor.setLimit(0xffff);
tssDescriptor.dpl = 0;
tssDescriptor.segment_present = 1;
tssDescriptor.granularity = 1;
tssDescriptor.zero = 0;
tssDescriptor.operation_size = 1;
tssDescriptor.descriptor_type = 0;
tssDescriptor.type = 9;
flushGDT();
}
void Task::prepForIRETToNewTask()
{
redoKernelTaskTSS();
s_kernelTask->tss().backlink = current->selector();
loadTaskRegister(s_kernelTask->selector());
}
void Task::initialize()
{
current = nullptr;
next_pid = 0;
2018-10-17 08:55:43 +00:00
s_tasks = new InlineLinkedList<Task>;
s_deadTasks = new InlineLinkedList<Task>;
s_kernelTask = new Task(0, "colonel", IPC::Handle::Any, Task::Ring0);
redoKernelTaskTSS();
loadTaskRegister(s_kernelTask->selector());
}
#ifdef TASK_SANITY_CHECKS
void Task::checkSanity(const char* msg)
{
char ch = current->name()[0];
kprintf("<%p> %s{%u}%b [%d] :%b: sanity check <%s>\n",
current->name().characters(),
current->name().characters(),
current->name().length(),
current->name()[current->name().length() - 1],
current->pid(), ch, msg ? msg : "");
ASSERT((ch >= 'a' && ch <= 'z') || (ch >= 'A' && ch <= 'Z'));
}
#endif
void Task::allocateLDT()
{
ASSERT(!m_tss.ldt);
static const WORD numLDTEntries = 4;
WORD newLDTSelector = allocateGDTEntry();
m_ldtEntries = new Descriptor[numLDTEntries];
2018-10-18 12:53:00 +00:00
#if 0
kprintf("new ldt selector = %x\n", newLDTSelector);
kprintf("new ldt table at = %p\n", m_ldtEntries);
kprintf("new ldt table size = %u\n", (numLDTEntries * 8) - 1);
#endif
Descriptor& ldt = getGDTEntry(newLDTSelector);
ldt.setBase(m_ldtEntries);
ldt.setLimit(numLDTEntries * 8 - 1);
ldt.dpl = 0;
ldt.segment_present = 1;
ldt.granularity = 0;
ldt.zero = 0;
ldt.operation_size = 1;
ldt.descriptor_type = 0;
ldt.type = Descriptor::LDT;
m_tss.ldt = newLDTSelector;
}
Vector<Task*> Task::allTasks()
{
Vector<Task*> tasks;
tasks.ensureCapacity(s_tasks->sizeSlow());
for (auto* task = s_tasks->head(); task; task = task->next())
tasks.append(task);
return tasks;
}
2018-10-18 12:53:00 +00:00
Task::Region* Task::allocateRegion(size_t size, String&& name)
{
2018-10-18 12:53:00 +00:00
// FIXME: This needs sanity checks. What if this overlaps existing regions?
auto zone = MemoryManager::the().createZone(size);
2018-10-18 12:53:00 +00:00
ASSERT(zone);
m_regions.append(make<Region>(m_nextRegion, size, move(zone), move(name)));
m_nextRegion = m_nextRegion.offset(size).offset(16384);
return m_regions.last().ptr();
}
int Task::sys$spawn(const char* path)
{
auto* child = Task::create(path, m_uid, m_gid);
if (child)
return child->pid();
return -1;
}
Task* Task::create(const String& path, uid_t uid, gid_t gid)
{
auto parts = path.split('/');
if (parts.isEmpty())
return nullptr;
auto handle = VirtualFileSystem::the().open(path);
if (!handle)
return nullptr;
auto elfData = handle->readEntireFile();
if (!elfData)
return nullptr;
cli();
Task* t = new Task(parts.takeLast(), uid, gid);
ExecSpace space;
space.hookableAlloc = [&] (const String& name, size_t size) {
if (!size)
return (void*)nullptr;
size = ((size / 4096) + 1) * 4096;
Region* region = t->allocateRegion(size, String(name));
ASSERT(region);
MemoryManager::the().mapRegion(*t, *region);
return (void*)region->linearAddress.asPtr();
};
bool success = space.loadELF(move(elfData));
if (!success) {
delete t;
return nullptr;
}
t->m_tss.eip = (dword)space.symbolPtr("_start");
if (!t->m_tss.eip) {
delete t;
return nullptr;
}
MemoryManager::the().unmapRegionsForTask(*t);
MemoryManager::the().mapRegionsForTask(*current);
s_tasks->prepend(t);
system.nprocess++;
#ifdef TASK_DEBUG
kprintf("Task %u (%s) spawned @ %p\n", t->pid(), t->name().characters(), t->m_tss.eip);
#endif
sti();
return t;
}
Task::Task(String&& name, uid_t uid, gid_t gid)
: m_name(move(name))
, m_pid(next_pid++)
, m_uid(uid)
, m_gid(gid)
, m_state(Runnable)
, m_ring(Ring3)
{
m_nextRegion = LinearAddress(0x600000);
memset(&m_tss, 0, sizeof(m_tss));
memset(&m_ldtEntries, 0, sizeof(m_ldtEntries));
allocateLDT();
// Only IF is set when a task boots.
m_tss.eflags = 0x0202;
WORD codeSegment = 0x1b;
WORD dataSegment = 0x23;
WORD stackSegment = dataSegment;
m_tss.ds = dataSegment;
m_tss.es = dataSegment;
m_tss.fs = dataSegment;
m_tss.gs = dataSegment;
m_tss.ss = stackSegment;
m_tss.cs = codeSegment;
m_tss.cr3 = MemoryManager::the().pageDirectoryBase().get();
auto* region = allocateRegion(defaultStackSize, "stack");
ASSERT(region);
m_stackTop = region->linearAddress.offset(defaultStackSize).get() & 0xfffffff8;
m_tss.esp = m_stackTop;
// Set up a separate stack for Ring0.
m_kernelStack = kmalloc(defaultStackSize);
DWORD ring0StackTop = ((DWORD)m_kernelStack + defaultStackSize) & 0xffffff8;
m_tss.ss0 = 0x10;
m_tss.esp0 = ring0StackTop;
// HACK: Ring2 SS in the TSS is the current PID.
m_tss.ss2 = m_pid;
m_farPtr.offset = 0x98765432;
ASSERT(m_pid);
}
Task::Task(void (*e)(), const char* n, IPC::Handle h, RingLevel ring)
: m_name(n)
, m_entry(e)
, m_pid(next_pid++)
, m_handle(h)
, m_state(Runnable)
, m_ring(ring)
{
2018-10-18 12:53:00 +00:00
m_nextRegion = LinearAddress(0x600000);
2018-10-18 12:53:00 +00:00
Region* codeRegion = nullptr;
if (!isRing0()) {
codeRegion = allocateRegion(4096, "code");
ASSERT(codeRegion);
bool success = copyToZone(*codeRegion->zone, (void*)e, PAGE_SIZE);
ASSERT(success);
}
memset(&m_tss, 0, sizeof(m_tss));
memset(&m_ldtEntries, 0, sizeof(m_ldtEntries));
if (ring == Ring3) {
allocateLDT();
}
// Only IF is set when a task boots.
m_tss.eflags = 0x0202;
WORD dataSegment;
WORD stackSegment;
WORD codeSegment;
if (ring == Ring0) {
codeSegment = 0x08;
dataSegment = 0x10;
stackSegment = dataSegment;
} else {
codeSegment = 0x1b;
dataSegment = 0x23;
stackSegment = dataSegment;
}
m_tss.ds = dataSegment;
m_tss.es = dataSegment;
m_tss.fs = dataSegment;
m_tss.gs = dataSegment;
m_tss.ss = stackSegment;
m_tss.cs = codeSegment;
ASSERT((codeSegment & 3) == (stackSegment & 3));
m_tss.cr3 = MemoryManager::the().pageDirectoryBase().get();
if (isRing0()) {
m_tss.eip = (DWORD)m_entry;
} else {
2018-10-18 12:53:00 +00:00
m_tss.eip = codeRegion->linearAddress.get();
}
2018-10-18 12:53:00 +00:00
if (isRing0()) {
// FIXME: This memory is leaked.
// But uh, there's also no kernel task termination, so I guess it's not technically leaked...
dword stackBottom = (dword)kmalloc(defaultStackSize);
m_stackTop = (stackBottom + defaultStackSize) & 0xffffff8;
2018-10-18 12:53:00 +00:00
m_tss.esp = m_stackTop;
} else {
auto* region = allocateRegion(defaultStackSize, "stack");
ASSERT(region);
m_stackTop = region->linearAddress.offset(defaultStackSize).get() & 0xfffffff8;
m_tss.esp = m_stackTop;
}
if (ring == Ring3) {
// Set up a separate stack for Ring0.
// FIXME: Don't leak this stack either.
2018-10-18 12:53:00 +00:00
m_kernelStack = kmalloc(defaultStackSize);
DWORD ring0StackTop = ((DWORD)m_kernelStack + defaultStackSize) & 0xffffff8;
m_tss.ss0 = 0x10;
m_tss.esp0 = ring0StackTop;
}
// HACK: Ring2 SS in the TSS is the current PID.
m_tss.ss2 = m_pid;
m_farPtr.offset = 0x12345678;
// Don't add task 0 (kernel dummy task) to task list.
// FIXME: This doesn't belong in the constructor.
if (m_pid == 0)
return;
// Add it to head of task list (meaning it's next to run too, ATM.)
s_tasks->prepend(this);
system.nprocess++;
#ifdef TASK_DEBUG
kprintf("Task %u (%s) spawned @ %p\n", m_pid, m_name.characters(), m_tss.eip);
#endif
}
Task::~Task()
{
system.nprocess--;
delete [] m_ldtEntries;
m_ldtEntries = nullptr;
2018-10-18 12:53:00 +00:00
if (m_kernelStack) {
kfree(m_kernelStack);
m_kernelStack = nullptr;
}
}
void Task::dumpRegions()
{
kprintf("Task %s(%u) regions:\n", name().characters(), pid());
kprintf("BEGIN END SIZE NAME\n");
for (auto& region : m_regions) {
kprintf("%x -- %x %x %s\n",
region->linearAddress.get(),
region->linearAddress.offset(region->size - 1).get(),
region->size,
region->name.characters());
}
}
void Task::sys$exit(int status)
{
cli();
#ifdef TASK_DEBUG
kprintf("sys$exit: %s(%u) exit with status %d\n", name().characters(), pid(), status);
#endif
setState(Exiting);
MemoryManager::the().unmapRegionsForTask(*this);
s_tasks->remove(this);
if (!scheduleNewTask()) {
kprintf("Task::taskDidCrash: Failed to schedule a new task :(\n");
HANG;
}
s_deadTasks->append(this);
switchNow();
}
void Task::taskDidCrash(Task* crashedTask)
{
// NOTE: This is called from an excepton handler, so interrupts are disabled.
crashedTask->setState(Crashing);
crashedTask->dumpRegions();
2018-10-18 12:53:00 +00:00
s_tasks->remove(crashedTask);
MemoryManager::the().unmapRegionsForTask(*crashedTask);
if (!scheduleNewTask()) {
kprintf("Task::taskDidCrash: Failed to schedule a new task :(\n");
HANG;
}
s_deadTasks->append(crashedTask);
switchNow();
}
void Task::doHouseKeeping()
{
Task* next = nullptr;
for (auto* deadTask = s_deadTasks->head(); deadTask; deadTask = next) {
next = deadTask->next();
delete deadTask;
}
s_deadTasks->clear();
}
void yield()
{
if (!current) {
kprintf( "PANIC: yield() with !current" );
HANG;
}
//kprintf("%s<%u> yield()\n", current->name().characters(), current->pid());
cli();
if (!scheduleNewTask()) {
sti();
return;
}
//kprintf("yield() jumping to new task: %x (%s)\n", current->farPtr().selector, current->name().characters());
switchNow();
}
void switchNow()
{
Descriptor& descriptor = getGDTEntry(current->selector());
descriptor.type = 9;
flushGDT();
asm("sti\n"
"ljmp *(%%eax)\n"
::"a"(&current->farPtr())
);
}
bool scheduleNewTask()
{
if (!current) {
// XXX: The first ever context_switch() goes to the idle task.
// This to setup a reliable place we can return to.
return contextSwitch(Task::kernelTask());
}
#if 0
kprintf("Scheduler choices:\n");
for (auto* task = s_tasks->head(); task; task = task->next()) {
kprintf("%p %u %s\n", task, task->pid(), task->name().characters());
}
#endif
// Check and unblock tasks whose wait conditions have been met.
for (auto* task = s_tasks->head(); task; task = task->next()) {
if (task->state() == Task::BlockedReceive && (task->ipc.msg.isValid() || task->ipc.notifies)) {
task->unblock();
continue;
}
if (task->state() == Task::BlockedSend) {
Task* peer = Task::fromIPCHandle(task->ipc.dst);
if (peer && peer->state() == Task::BlockedReceive && peer->acceptsMessageFrom(*task)) {
task->unblock();
continue;
}
}
if (task->state() == Task::BlockedSleep) {
if (task->wakeupTime() <= system.uptime) {
task->unblock();
continue;
}
}
if (task->state() == Task::BlockedWait) {
if (!Task::fromPID(task->waitee())) {
task->unblock();
continue;
}
}
}
auto* prevHead = s_tasks->head();
for (;;) {
// Move head to tail.
s_tasks->append(s_tasks->removeHead());
auto* task = s_tasks->head();
if (task->state() == Task::Runnable || task->state() == Task::Running) {
//kprintf("switch to %s (%p vs %p)\n", task->name().characters(), task, current);
return contextSwitch(task);
}
if (task == prevHead) {
// Back at task_head, nothing wants to run.
kprintf("Switch to kernel task\n");
return contextSwitch(Task::kernelTask());
}
}
}
static bool contextSwitch(Task* t)
{
//kprintf("c_s to %s (same:%u)\n", t->name().characters(), current == t);
t->setTicksLeft(5);
if (current == t)
return false;
// Some sanity checking to force a crash earlier.
auto csRPL = t->tss().cs & 3;
auto ssRPL = t->tss().ss & 3;
if (csRPL != ssRPL) {
kprintf("Fuckup! Switching from %s(%u) to %s(%u) has RPL mismatch\n",
current->name().characters(), current->pid(),
t->name().characters(), t->pid()
);
kprintf("code: %w:%x\n", t->tss().cs, t->tss().eip);
kprintf(" stk: %w:%x\n", t->tss().ss, t->tss().esp);
ASSERT(csRPL == ssRPL);
}
2018-10-21 19:57:43 +00:00
if (current) {
// If the last task hasn't blocked (still marked as running),
// mark it as runnable for the next round.
if (current->state() == Task::Running)
current->setState(Task::Runnable);
2018-10-21 19:57:43 +00:00
bool success = MemoryManager::the().unmapRegionsForTask(*current);
ASSERT(success);
}
bool success = MemoryManager::the().mapRegionsForTask(*t);
ASSERT(success);
current = t;
t->setState(Task::Running);
if (!t->selector())
t->setSelector(allocateGDTEntry());
auto& tssDescriptor = getGDTEntry(t->selector());
tssDescriptor.limit_hi = 0;
tssDescriptor.limit_lo = 0xFFFF;
tssDescriptor.base_lo = (DWORD)(&t->tss()) & 0xFFFF;
tssDescriptor.base_hi = ((DWORD)(&t->tss()) >> 16) & 0xFF;
tssDescriptor.base_hi2 = ((DWORD)(&t->tss()) >> 24) & 0xFF;
tssDescriptor.dpl = 0;
tssDescriptor.segment_present = 1;
tssDescriptor.granularity = 1;
tssDescriptor.zero = 0;
tssDescriptor.operation_size = 1;
tssDescriptor.descriptor_type = 0;
tssDescriptor.type = 11; // Busy TSS
flushGDT();
t->didSchedule();
return true;
}
Task* Task::fromPID(pid_t pid)
{
for (auto* task = s_tasks->head(); task; task = task->next()) {
if (task->pid() == pid)
return task;
}
return nullptr;
}
Task* Task::fromIPCHandle(IPC::Handle handle)
{
for (auto* task = s_tasks->head(); task; task = task->next()) {
if (task->handle() == handle)
return task;
}
return nullptr;
}
FileHandle* Task::fileHandleIfExists(int fd)
{
if (fd < 0)
return nullptr;
if ((unsigned)fd < m_fileHandles.size())
2018-10-18 08:28:09 +00:00
return m_fileHandles[fd].ptr();
return nullptr;
}
int Task::sys$seek(int fd, int offset)
{
2018-10-18 08:28:09 +00:00
auto* handle = fileHandleIfExists(fd);
if (!handle)
return -1;
2018-10-18 08:28:09 +00:00
return handle->seek(offset, SEEK_SET);
}
2018-10-18 08:28:09 +00:00
ssize_t Task::sys$read(int fd, void* outbuf, size_t nread)
{
Task::checkSanity("Task::sys$read");
#ifdef DEBUG_IO
kprintf("Task::sys$read: called(%d, %p, %u)\n", fd, outbuf, nread);
#endif
2018-10-18 08:28:09 +00:00
auto* handle = fileHandleIfExists(fd);
#ifdef DEBUG_IO
kprintf("Task::sys$read: handle=%p\n", handle);
#endif
if (!handle) {
kprintf("Task::sys$read: handle not found :(\n");
return -1;
}
#ifdef DEBUG_IO
kprintf("call read on handle=%p\n", handle);
#endif
2018-10-18 08:28:09 +00:00
nread = handle->read((byte*)outbuf, nread);
#ifdef DEBUG_IO
kprintf("Task::sys$read: nread=%u\n", nread);
#endif
return nread;
}
int Task::sys$close(int fd)
{
2018-10-18 08:28:09 +00:00
auto* handle = fileHandleIfExists(fd);
if (!handle)
return -1;
2018-10-18 08:28:09 +00:00
// FIXME: Implement.
return 0;
}
int Task::sys$open(const char* path, size_t pathLength)
{
Task::checkSanity("sys$open");
#ifdef DEBUG_IO
kprintf("Task::sys$open(): PID=%u, path=%s {%u}\n", m_pid, path, pathLength);
#endif
auto* handle = current->openFile(String(path, pathLength));
if (handle)
return handle->fd();
return -1;
}
FileHandle* Task::openFile(String&& path)
{
2018-10-18 08:28:09 +00:00
auto handle = VirtualFileSystem::the().open(move(path));
if (!handle) {
#ifdef DEBUG_IO
2018-10-18 08:28:09 +00:00
kprintf("vfs::open() failed\n");
#endif
return nullptr;
}
2018-10-18 08:28:09 +00:00
handle->setFD(m_fileHandles.size());
#ifdef DEBUG_IO
2018-10-18 08:28:09 +00:00
kprintf("vfs::open() worked! handle=%p, fd=%d\n", handle.ptr(), handle->fd());
#endif
2018-10-18 08:28:09 +00:00
m_fileHandles.append(move(handle)); // FIXME: allow non-move Vector::append
return m_fileHandles.last().ptr();
}
int Task::sys$kill(pid_t pid, int sig)
{
(void) sig;
if (pid == 0) {
// FIXME: Send to same-group processes.
ASSERT(pid != 0);
}
if (pid == -1) {
// FIXME: Send to all processes.
ASSERT(pid != -1);
}
ASSERT_NOT_REACHED();
Task* peer = Task::fromPID(pid);
if (!peer) {
// errno = ESRCH;
return -1;
}
#if 0
send(peer->handle(), IPC::Message(SYS_KILL, DataBuffer::copy((BYTE*)&sig, sizeof(sig))));
IPC::Message response = receive(peer->handle());
return *(int*)response.data();
#endif
return -1;
}
uid_t Task::sys$getuid()
{
return m_uid;
}
gid_t Task::sys$getgid()
{
return m_gid;
}
pid_t Task::sys$getpid()
{
return m_pid;
}
pid_t Task::sys$waitpid(pid_t waitee)
{
if (!Task::fromPID(waitee))
return -1;
m_waitee = waitee;
block(BlockedWait);
yield();
return m_waitee;
}
bool Task::acceptsMessageFrom(Task& peer)
{
return !ipc.msg.isValid() && (ipc.src == IPC::Handle::Any || ipc.src == peer.handle());
}
void Task::unblock()
{
ASSERT(m_state != Task::Runnable && m_state != Task::Running);
system.nblocked--;
m_state = Task::Runnable;
}
void Task::block(Task::State state)
{
ASSERT(current->state() == Task::Running);
system.nblocked++;
current->setState(state);
}
void block(Task::State state)
{
current->block(state);
yield();
}
void sleep(DWORD ticks)
{
ASSERT(current->state() == Task::Running);
current->setWakeupTime(system.uptime + ticks);
current->block(Task::BlockedSleep);
yield();
}
void Task::sys$sleep(DWORD ticks)
{
ASSERT(this == current);
sleep(ticks);
}
Task* Task::kernelTask()
{
ASSERT(s_kernelTask);
return s_kernelTask;
}
void Task::setError(int error)
{
m_error = error;
}
2018-10-18 12:53:00 +00:00
Task::Region::Region(LinearAddress a, size_t s, RetainPtr<Zone>&& z, String&& n)
: linearAddress(a)
, size(s)
, zone(move(z))
, name(move(n))
{
}
Task::Region::~Region()
{
}