From 761fe30bf38c8b0832e5eef977e1646f9da5b69e Mon Sep 17 00:00:00 2001
From: Jeremy Soller <jackpot51@gmail.com>
Date: Mon, 29 Jan 2018 21:29:24 -0700
Subject: [PATCH] Add linked list allocator with automatic resizing Fix memory
 leaks in exec Remove warnings

---
 Cargo.lock                        |   1 +
 Cargo.toml                        |   6 +-
 src/allocator/linked_list.rs      |  70 ++++++++++++++
 src/allocator/mod.rs              |  36 ++++++++
 src/{memory => allocator}/slab.rs |  12 ++-
 src/arch/x86_64/macros.rs         |  57 ++++++++----
 src/arch/x86_64/pti.rs            |   4 +
 src/arch/x86_64/start.rs          |  23 +----
 src/consts.rs                     |   2 +-
 src/lib.rs                        |   6 +-
 src/memory/mod.rs                 |   1 -
 src/syscall/process.rs            | 148 +++++++++++++++++-------------
 12 files changed, 252 insertions(+), 114 deletions(-)
 create mode 100644 src/allocator/linked_list.rs
 create mode 100644 src/allocator/mod.rs
 rename src/{memory => allocator}/slab.rs (89%)

diff --git a/Cargo.lock b/Cargo.lock
index d50efe97..eb05733b 100644
--- a/Cargo.lock
+++ b/Cargo.lock
@@ -125,6 +125,7 @@ dependencies = [
  "bitflags 1.0.1 (registry+https://github.com/rust-lang/crates.io-index)",
  "clippy 0.0.182 (registry+https://github.com/rust-lang/crates.io-index)",
  "goblin 0.0.10 (registry+https://github.com/rust-lang/crates.io-index)",
+ "linked_list_allocator 0.5.0 (registry+https://github.com/rust-lang/crates.io-index)",
  "raw-cpuid 3.0.0 (registry+https://github.com/rust-lang/crates.io-index)",
  "redox_syscall 0.1.37",
  "slab_allocator 0.3.0",
diff --git a/Cargo.toml b/Cargo.toml
index 6e33b585..56e09c55 100644
--- a/Cargo.toml
+++ b/Cargo.toml
@@ -11,10 +11,11 @@ crate-type = ["staticlib"]
 [dependencies]
 bitflags = "1"
 clippy = { version = "*", optional = true }
-slab_allocator = { path = "slab_allocator" }
-spin = "0.4"
+linked_list_allocator = "0.5"
 raw-cpuid = "3.0"
 redox_syscall = { path = "syscall" }
+slab_allocator = { path = "slab_allocator" }
+spin = "0.4"
 
 [dependencies.goblin]
 version = "0.0.10"
@@ -31,3 +32,4 @@ doc = []
 live = []
 multi_core = []
 pti = []
+slab = []
diff --git a/src/allocator/linked_list.rs b/src/allocator/linked_list.rs
new file mode 100644
index 00000000..989101c4
--- /dev/null
+++ b/src/allocator/linked_list.rs
@@ -0,0 +1,70 @@
+use alloc::heap::{Alloc, AllocErr, Layout};
+use linked_list_allocator::Heap;
+use spin::Mutex;
+
+use paging::ActivePageTable;
+
+static HEAP: Mutex<Option<Heap>> = Mutex::new(None);
+
+pub struct Allocator;
+
+impl Allocator {
+    pub unsafe fn init(offset: usize, size: usize) {
+        *HEAP.lock() = Some(Heap::new(offset, size));
+    }
+}
+
+unsafe impl<'a> Alloc for &'a Allocator {
+    unsafe fn alloc(&mut self, mut layout: Layout) -> Result<*mut u8, AllocErr> {
+        loop {
+            let res = if let Some(ref mut heap) = *HEAP.lock() {
+                heap.allocate_first_fit(layout)
+            } else {
+                panic!("__rust_allocate: heap not initialized");
+            };
+
+            match res {
+                Err(AllocErr::Exhausted { request }) => {
+                    layout = request;
+
+                    let size = if let Some(ref heap) = *HEAP.lock() {
+                        heap.size()
+                    } else {
+                        panic!("__rust_allocate: heap not initialized");
+                    };
+
+                    println!("Expand heap at {} MB by {} MB", size/1024/1024, ::KERNEL_HEAP_SIZE/1024/1024);
+
+                    super::map_heap(&mut ActivePageTable::new(), ::KERNEL_HEAP_OFFSET + size, ::KERNEL_HEAP_SIZE);
+
+                    if let Some(ref mut heap) = *HEAP.lock() {
+                        heap.extend(::KERNEL_HEAP_SIZE);
+                    } else {
+                        panic!("__rust_allocate: heap not initialized");
+                    }
+                },
+                other => return other,
+            }
+        }
+    }
+
+    unsafe fn dealloc(&mut self, ptr: *mut u8, layout: Layout) {
+        if let Some(ref mut heap) = *HEAP.lock() {
+            heap.deallocate(ptr, layout)
+        } else {
+            panic!("__rust_deallocate: heap not initialized");
+        }
+    }
+
+    fn oom(&mut self, error: AllocErr) -> ! {
+        panic!("Out of memory: {:?}", error);
+    }
+
+    fn usable_size(&self, layout: &Layout) -> (usize, usize) {
+        if let Some(ref mut heap) = *HEAP.lock() {
+            heap.usable_size(layout)
+        } else {
+            panic!("__rust_usable_size: heap not initialized");
+        }
+    }
+}
diff --git a/src/allocator/mod.rs b/src/allocator/mod.rs
new file mode 100644
index 00000000..5b5b0cce
--- /dev/null
+++ b/src/allocator/mod.rs
@@ -0,0 +1,36 @@
+use paging::{ActivePageTable, Page, VirtualAddress};
+use paging::entry::EntryFlags;
+use paging::mapper::MapperFlushAll;
+
+#[cfg(not(feature="slab"))]
+pub use self::linked_list::Allocator;
+
+#[cfg(feature="slab")]
+pub use self::slab::Allocator;
+
+mod linked_list;
+mod slab;
+
+unsafe fn map_heap(active_table: &mut ActivePageTable, offset: usize, size: usize) {
+    let mut flush_all = MapperFlushAll::new();
+
+    let heap_start_page = Page::containing_address(VirtualAddress::new(offset));
+    let heap_end_page = Page::containing_address(VirtualAddress::new(offset + size-1));
+    for page in Page::range_inclusive(heap_start_page, heap_end_page) {
+        let result = active_table.map(page, EntryFlags::PRESENT | EntryFlags::GLOBAL | EntryFlags::WRITABLE | EntryFlags::NO_EXECUTE);
+        flush_all.consume(result);
+    }
+
+    flush_all.flush(active_table);
+}
+
+pub unsafe fn init(active_table: &mut ActivePageTable) {
+    let offset = ::KERNEL_HEAP_OFFSET;
+    let size = ::KERNEL_HEAP_SIZE;
+
+    // Map heap pages
+    map_heap(active_table, offset, size);
+
+    // Initialize global heap
+    Allocator::init(offset, size);
+}
diff --git a/src/memory/slab.rs b/src/allocator/slab.rs
similarity index 89%
rename from src/memory/slab.rs
rename to src/allocator/slab.rs
index 286e146d..ddc50501 100644
--- a/src/memory/slab.rs
+++ b/src/allocator/slab.rs
@@ -4,12 +4,14 @@ use slab_allocator::Heap;
 
 static HEAP: Mutex<Option<Heap>> = Mutex::new(None);
 
-pub unsafe fn init_heap(offset: usize, size: usize) {
-    *HEAP.lock() = Some(Heap::new(offset, size));
-}
-
 pub struct Allocator;
 
+impl Allocator {
+    pub unsafe fn init(offset: usize, size: usize) {
+        *HEAP.lock() = Some(Heap::new(offset, size));
+    }
+}
+
 unsafe impl<'a> Alloc for &'a Allocator {
     unsafe fn alloc(&mut self, layout: Layout) -> Result<*mut u8, AllocErr> {
         if let Some(ref mut heap) = *HEAP.lock() {
@@ -38,4 +40,4 @@ unsafe impl<'a> Alloc for &'a Allocator {
             panic!("__rust_usable_size: heap not initialized");
         }
     }
-}
\ No newline at end of file
+}
diff --git a/src/arch/x86_64/macros.rs b/src/arch/x86_64/macros.rs
index a7b5a3b0..29a3024b 100644
--- a/src/arch/x86_64/macros.rs
+++ b/src/arch/x86_64/macros.rs
@@ -31,15 +31,25 @@ pub struct ScratchRegisters {
 
 impl ScratchRegisters {
     pub fn dump(&self) {
-        println!("RAX:   {:>016X}", self.rax);
-        println!("RCX:   {:>016X}", self.rcx);
-        println!("RDX:   {:>016X}", self.rdx);
-        println!("RDI:   {:>016X}", self.rdi);
-        println!("RSI:   {:>016X}", self.rsi);
-        println!("R8:    {:>016X}", self.r8);
-        println!("R9:    {:>016X}", self.r9);
-        println!("R10:   {:>016X}", self.r10);
-        println!("R11:   {:>016X}", self.r11);
+        let rax = self.rax;
+        let rcx = self.rcx;
+        let rdx = self.rdx;
+        let rdi = self.rdi;
+        let rsi = self.rsi;
+        let r8 = self.r8;
+        let r9 = self.r9;
+        let r10 = self.r10;
+        let r11 = self.r11;
+
+        println!("RAX:   {:>016X}", rax);
+        println!("RCX:   {:>016X}", rcx);
+        println!("RDX:   {:>016X}", rdx);
+        println!("RDI:   {:>016X}", rdi);
+        println!("RSI:   {:>016X}", rsi);
+        println!("R8:    {:>016X}", r8);
+        println!("R9:    {:>016X}", r9);
+        println!("R10:   {:>016X}", r10);
+        println!("R11:   {:>016X}", r11);
     }
 }
 
@@ -86,12 +96,19 @@ pub struct PreservedRegisters {
 
 impl PreservedRegisters {
     pub fn dump(&self) {
-        println!("RBX:   {:>016X}", self.rbx);
-        println!("RBP:   {:>016X}", self.rbp);
-        println!("R12:   {:>016X}", self.r12);
-        println!("R13:   {:>016X}", self.r13);
-        println!("R14:   {:>016X}", self.r14);
-        println!("R15:   {:>016X}", self.r15);
+        let rbx = self.rbx;
+        let rbp = self.rbp;
+        let r12 = self.r12;
+        let r13 = self.r13;
+        let r14 = self.r14;
+        let r15 = self.r15;
+
+        println!("RBX:   {:>016X}", rbx);
+        println!("RBP:   {:>016X}", rbp);
+        println!("R12:   {:>016X}", r12);
+        println!("R13:   {:>016X}", r13);
+        println!("R14:   {:>016X}", r14);
+        println!("R15:   {:>016X}", r15);
     }
 }
 
@@ -145,9 +162,13 @@ pub struct IretRegisters {
 
 impl IretRegisters {
     pub fn dump(&self) {
-        println!("RFLAG: {:>016X}", self.rflags);
-        println!("CS:    {:>016X}", self.cs);
-        println!("RIP:   {:>016X}", self.rip);
+        let rflags = self.rflags;
+        let cs = self.cs;
+        let rip = self.rip;
+
+        println!("RFLAG: {:>016X}", rflags);
+        println!("CS:    {:>016X}", cs);
+        println!("RIP:   {:>016X}", rip);
     }
 }
 
diff --git a/src/arch/x86_64/pti.rs b/src/arch/x86_64/pti.rs
index 9124c92f..33637155 100644
--- a/src/arch/x86_64/pti.rs
+++ b/src/arch/x86_64/pti.rs
@@ -1,7 +1,11 @@
+#[cfg(feature = "pti")]
 use core::ptr;
 
+#[cfg(feature = "pti")]
 use memory::Frame;
+#[cfg(feature = "pti")]
 use paging::ActivePageTable;
+#[cfg(feature = "pti")]
 use paging::entry::EntryFlags;
 
 #[cfg(feature = "pti")]
diff --git a/src/arch/x86_64/start.rs b/src/arch/x86_64/start.rs
index 26ad295b..96d24586 100644
--- a/src/arch/x86_64/start.rs
+++ b/src/arch/x86_64/start.rs
@@ -6,6 +6,7 @@
 use core::slice;
 use core::sync::atomic::{AtomicBool, ATOMIC_BOOL_INIT, AtomicUsize, ATOMIC_USIZE_INIT, Ordering};
 
+use allocator;
 use acpi;
 use arch::x86_64::pti;
 use device;
@@ -13,10 +14,7 @@ use gdt;
 use idt;
 use interrupt;
 use memory;
-use memory::slab as allocator;
-use paging::{self, Page, VirtualAddress};
-use paging::entry::EntryFlags;
-use paging::mapper::MapperFlushAll;
+use paging;
 
 /// Test of zero values in BSS.
 static BSS_TEST_ZERO: usize = 0;
@@ -99,22 +97,7 @@ pub unsafe extern fn kstart(args_ptr: *const KernelArgs) -> ! {
         BSP_READY.store(false, Ordering::SeqCst);
 
         // Setup kernel heap
-        {
-            let mut flush_all = MapperFlushAll::new();
-
-            // Map heap pages
-            let heap_start_page = Page::containing_address(VirtualAddress::new(::KERNEL_HEAP_OFFSET));
-            let heap_end_page = Page::containing_address(VirtualAddress::new(::KERNEL_HEAP_OFFSET + ::KERNEL_HEAP_SIZE-1));
-            for page in Page::range_inclusive(heap_start_page, heap_end_page) {
-                let result = active_table.map(page, EntryFlags::PRESENT | EntryFlags::GLOBAL | EntryFlags::WRITABLE | EntryFlags::NO_EXECUTE);
-                flush_all.consume(result);
-            }
-
-            flush_all.flush(&mut active_table);
-
-            // Init the allocator
-            allocator::init_heap(::KERNEL_HEAP_OFFSET, ::KERNEL_HEAP_SIZE);
-        }
+        allocator::init(&mut active_table);
 
         // Initialize devices
         device::init(&mut active_table);
diff --git a/src/consts.rs b/src/consts.rs
index 309c4935..3bf54f6f 100644
--- a/src/consts.rs
+++ b/src/consts.rs
@@ -19,7 +19,7 @@
     pub const KERNEL_HEAP_OFFSET: usize = KERNEL_OFFSET - PML4_SIZE;
     pub const KERNEL_HEAP_PML4: usize = (KERNEL_HEAP_OFFSET & PML4_MASK)/PML4_SIZE;
     /// Size of kernel heap
-    pub const KERNEL_HEAP_SIZE: usize = 256 * 1024 * 1024; // 256 MB
+    pub const KERNEL_HEAP_SIZE: usize = 1 * 1024 * 1024; // 1 MB
 
     /// Offset to kernel percpu variables
     //TODO: Use 64-bit fs offset to enable this pub const KERNEL_PERCPU_OFFSET: usize = KERNEL_HEAP_OFFSET - PML4_SIZE;
diff --git a/src/lib.rs b/src/lib.rs
index c73a218b..7e5a6b96 100644
--- a/src/lib.rs
+++ b/src/lib.rs
@@ -41,6 +41,7 @@ extern crate alloc;
 #[macro_use]
 extern crate bitflags;
 extern crate goblin;
+extern crate linked_list_allocator;
 extern crate spin;
 extern crate slab_allocator;
 
@@ -64,6 +65,9 @@ pub use arch::*;
 /// Constants like memory locations
 pub mod consts;
 
+/// Heap allocators
+pub mod allocator;
+
 /// ACPI table parsing
 mod acpi;
 
@@ -104,7 +108,7 @@ pub mod time;
 pub mod tests;
 
 #[global_allocator]
-static ALLOCATOR: memory::slab::Allocator = memory::slab::Allocator;
+static ALLOCATOR: allocator::Allocator = allocator::Allocator;
 
 /// A unique number that identifies the current CPU - used for scheduling
 #[thread_local]
diff --git a/src/memory/mod.rs b/src/memory/mod.rs
index 39cdf5db..6ab6f0fb 100644
--- a/src/memory/mod.rs
+++ b/src/memory/mod.rs
@@ -10,7 +10,6 @@ use spin::Mutex;
 
 pub mod bump;
 pub mod recycle;
-pub mod slab;
 
 /// The current memory map. It's size is maxed out to 512 entries, due to it being
 /// from 0x500 to 0x5000 (800 is the absolute total)
diff --git a/src/syscall/process.rs b/src/syscall/process.rs
index 049582ad..97776bfd 100644
--- a/src/syscall/process.rs
+++ b/src/syscall/process.rs
@@ -542,8 +542,14 @@ impl Drop for ExecFile {
     }
 }
 
-fn exec_noreturn(elf: elf::Elf, canonical: Box<[u8]>, setuid: Option<u32>, setgid: Option<u32>, args: Box<[Box<[u8]>]>) -> ! {
-    let entry = elf.entry();
+fn exec_noreturn(
+    canonical: Box<[u8]>,
+    setuid: Option<u32>,
+    setgid: Option<u32>,
+    data: Box<[u8]>,
+    args: Box<[Box<[u8]>]>
+) -> ! {
+    let entry;
     let mut sp = ::USER_STACK_OFFSET + ::USER_STACK_SIZE - 256;
 
     {
@@ -567,69 +573,76 @@ fn exec_noreturn(elf: elf::Elf, canonical: Box<[u8]>, setuid: Option<u32>, setgi
 
             // Map and copy new segments
             let mut tls_option = None;
-            for segment in elf.segments() {
-                if segment.p_type == program_header::PT_LOAD {
-                    let voff = segment.p_vaddr % 4096;
-                    let vaddr = segment.p_vaddr - voff;
-
-                    let mut memory = context::memory::Memory::new(
-                        VirtualAddress::new(vaddr as usize),
-                        segment.p_memsz as usize + voff as usize,
-                        EntryFlags::NO_EXECUTE | EntryFlags::WRITABLE,
-                        true
-                    );
+            {
+                let elf = elf::Elf::from(&data).unwrap();
+                entry = elf.entry();
+                for segment in elf.segments() {
+                    if segment.p_type == program_header::PT_LOAD {
+                        let voff = segment.p_vaddr % 4096;
+                        let vaddr = segment.p_vaddr - voff;
+
+                        let mut memory = context::memory::Memory::new(
+                            VirtualAddress::new(vaddr as usize),
+                            segment.p_memsz as usize + voff as usize,
+                            EntryFlags::NO_EXECUTE | EntryFlags::WRITABLE,
+                            true
+                        );
 
-                    unsafe {
-                        // Copy file data
-                        intrinsics::copy((elf.data.as_ptr() as usize + segment.p_offset as usize) as *const u8,
-                                        segment.p_vaddr as *mut u8,
-                                        segment.p_filesz as usize);
-                    }
+                        unsafe {
+                            // Copy file data
+                            intrinsics::copy((elf.data.as_ptr() as usize + segment.p_offset as usize) as *const u8,
+                                            segment.p_vaddr as *mut u8,
+                                            segment.p_filesz as usize);
+                        }
 
-                    let mut flags = EntryFlags::NO_EXECUTE | EntryFlags::USER_ACCESSIBLE;
+                        let mut flags = EntryFlags::NO_EXECUTE | EntryFlags::USER_ACCESSIBLE;
 
-                    if segment.p_flags & program_header::PF_R == program_header::PF_R {
-                        flags.insert(EntryFlags::PRESENT);
-                    }
+                        if segment.p_flags & program_header::PF_R == program_header::PF_R {
+                            flags.insert(EntryFlags::PRESENT);
+                        }
 
-                    // W ^ X. If it is executable, do not allow it to be writable, even if requested
-                    if segment.p_flags & program_header::PF_X == program_header::PF_X {
-                        flags.remove(EntryFlags::NO_EXECUTE);
-                    } else if segment.p_flags & program_header::PF_W == program_header::PF_W {
-                        flags.insert(EntryFlags::WRITABLE);
-                    }
+                        // W ^ X. If it is executable, do not allow it to be writable, even if requested
+                        if segment.p_flags & program_header::PF_X == program_header::PF_X {
+                            flags.remove(EntryFlags::NO_EXECUTE);
+                        } else if segment.p_flags & program_header::PF_W == program_header::PF_W {
+                            flags.insert(EntryFlags::WRITABLE);
+                        }
 
-                    memory.remap(flags);
+                        memory.remap(flags);
 
-                    context.image.push(memory.to_shared());
-                } else if segment.p_type == program_header::PT_TLS {
-                    let memory = context::memory::Memory::new(
-                        VirtualAddress::new(::USER_TCB_OFFSET),
-                        4096,
-                        EntryFlags::NO_EXECUTE | EntryFlags::WRITABLE | EntryFlags::USER_ACCESSIBLE,
-                        true
-                    );
-                    let aligned_size = if segment.p_align > 0 {
-                        ((segment.p_memsz + (segment.p_align - 1))/segment.p_align) * segment.p_align
-                    } else {
-                        segment.p_memsz
-                    };
-                    let rounded_size = ((aligned_size + 4095)/4096) * 4096;
-                    let rounded_offset = rounded_size - aligned_size;
-                    let tcb_offset = ::USER_TLS_OFFSET + rounded_size as usize;
-                    unsafe { *(::USER_TCB_OFFSET as *mut usize) = tcb_offset; }
-
-                    context.image.push(memory.to_shared());
-
-                    tls_option = Some((
-                        VirtualAddress::new(segment.p_vaddr as usize),
-                        segment.p_filesz as usize,
-                        rounded_size as usize,
-                        rounded_offset as usize,
-                    ));
+                        context.image.push(memory.to_shared());
+                    } else if segment.p_type == program_header::PT_TLS {
+                        let memory = context::memory::Memory::new(
+                            VirtualAddress::new(::USER_TCB_OFFSET),
+                            4096,
+                            EntryFlags::NO_EXECUTE | EntryFlags::WRITABLE | EntryFlags::USER_ACCESSIBLE,
+                            true
+                        );
+                        let aligned_size = if segment.p_align > 0 {
+                            ((segment.p_memsz + (segment.p_align - 1))/segment.p_align) * segment.p_align
+                        } else {
+                            segment.p_memsz
+                        };
+                        let rounded_size = ((aligned_size + 4095)/4096) * 4096;
+                        let rounded_offset = rounded_size - aligned_size;
+                        let tcb_offset = ::USER_TLS_OFFSET + rounded_size as usize;
+                        unsafe { *(::USER_TCB_OFFSET as *mut usize) = tcb_offset; }
+
+                        context.image.push(memory.to_shared());
+
+                        tls_option = Some((
+                            VirtualAddress::new(segment.p_vaddr as usize),
+                            segment.p_filesz as usize,
+                            rounded_size as usize,
+                            rounded_offset as usize,
+                        ));
+                    }
                 }
             }
 
+            // Data no longer required, can deallocate
+            drop(data);
+
             // Map heap
             context.heap = Some(context::memory::Memory::new(
                 VirtualAddress::new(::USER_HEAP_OFFSET),
@@ -713,6 +726,9 @@ fn exec_noreturn(elf: elf::Elf, canonical: Box<[u8]>, setuid: Option<u32>, setgi
                 context.image.push(memory.to_shared());
             }
 
+            // Args no longer required, can deallocate
+            drop(args);
+
             context.actions = Arc::new(Mutex::new(vec![(
                 SigAction {
                     sa_handler: unsafe { mem::transmute(SIG_DFL) },
@@ -861,10 +877,6 @@ pub fn exec(path: &[u8], arg_ptrs: &[[usize; 2]]) -> Result<usize> {
 
     match elf::Elf::from(&data) {
         Ok(elf) => {
-            // Drop so that usage is not allowed after unmapping context
-            drop(path);
-            drop(arg_ptrs);
-
             // We check the validity of all loadable sections here
             for segment in elf.segments() {
                 if segment.p_type == program_header::PT_LOAD {
@@ -880,17 +892,21 @@ pub fn exec(path: &[u8], arg_ptrs: &[[usize; 2]]) -> Result<usize> {
                     }
                 }
             }
-
-            // This is the point of no return, quite literaly. Any checks for validity need
-            // to be done before, and appropriate errors returned. Otherwise, we have nothing
-            // to return to.
-            exec_noreturn(elf, canonical.into_boxed_slice(), setuid, setgid, args.into_boxed_slice());
         },
         Err(err) => {
             println!("exec: failed to execute {}: {}", unsafe { str::from_utf8_unchecked(path) }, err);
-            Err(Error::new(ENOEXEC))
+            return Err(Error::new(ENOEXEC));
         }
     }
+
+    // Drop so that usage is not allowed after unmapping context
+    drop(path);
+    drop(arg_ptrs);
+
+    // This is the point of no return, quite literaly. Any checks for validity need
+    // to be done before, and appropriate errors returned. Otherwise, we have nothing
+    // to return to.
+    exec_noreturn(canonical.into_boxed_slice(), setuid, setgid, data.into_boxed_slice(), args.into_boxed_slice());
 }
 
 pub fn exit(status: usize) -> ! {
-- 
GitLab