diff options
author | Natanael Copa <ncopa@alpinelinux.org> | 2011-11-22 10:24:42 +0000 |
---|---|---|
committer | Natanael Copa <ncopa@alpinelinux.org> | 2011-11-22 10:24:42 +0000 |
commit | 972686408af9f0fd661213df86d37b77834b08e2 (patch) | |
tree | b5c48470a5609fd897d5c3e13f25283fe27e70de /main/linux-vserver | |
parent | cfab7149b49724e392a726baee37c9d4a3de7447 (diff) | |
download | aports-972686408af9f0fd661213df86d37b77834b08e2.tar.bz2 aports-972686408af9f0fd661213df86d37b77834b08e2.tar.xz |
main/*-vserver: upgrade to 3.0.10-vs2.3.2.1-unofficial
Diffstat (limited to 'main/linux-vserver')
-rw-r--r-- | main/linux-vserver/APKBUILD | 10 | ||||
-rw-r--r-- | main/linux-vserver/patch-3.0.10-vs2.3.2.1-unofficial.diff | 26517 |
2 files changed, 26522 insertions, 5 deletions
diff --git a/main/linux-vserver/APKBUILD b/main/linux-vserver/APKBUILD index b9ad3a3964..c2b0e0bbf4 100644 --- a/main/linux-vserver/APKBUILD +++ b/main/linux-vserver/APKBUILD @@ -2,8 +2,8 @@ _flavor=vserver pkgname=linux-${_flavor} -pkgver=3.0.9 -pkgrel=1 +pkgver=3.0.10 +pkgrel=0 _vsver=vs2.3.2.1 if [ "${pkgver##*.*.*}" = "$pkgver" ]; then @@ -21,7 +21,7 @@ _config=${config:-kernelconfig.${CARCH}} install= source="http://www.kernel.org/pub/linux/kernel/v3.0/linux-$_kernver.tar.bz2 http://www.kernel.org/pub/linux/kernel/v3.0/patch-$pkgver.bz2 - http://vserver.13thfloor.at/Experimental/patch-$pkgver-$_vsver.diff + patch-3.0.10-vs2.3.2.1-unofficial.diff kernelconfig.x86 kernelconfig.x86_64 " @@ -132,7 +132,7 @@ dev() { } md5sums="398e95866794def22b12dfbc15ce89c0 linux-3.0.tar.bz2 -0154d21e63d3f14fc1084cdb130fab2d patch-3.0.9.bz2 -da9ad65521398d86afb61223bf4d19a1 patch-3.0.9-vs2.3.2.1.diff +8e0391dd935d838ee3023ab9daa6b447 patch-3.0.10.bz2 +6d2fd60b055272429dde5e4043280fcb patch-3.0.10-vs2.3.2.1-unofficial.diff fdf2737ad7b77bcfe4290ef2b0112138 kernelconfig.x86 f82d261dc117ced42e2a3dcd4b6850a5 kernelconfig.x86_64" diff --git a/main/linux-vserver/patch-3.0.10-vs2.3.2.1-unofficial.diff b/main/linux-vserver/patch-3.0.10-vs2.3.2.1-unofficial.diff new file mode 100644 index 0000000000..2c09d43edf --- /dev/null +++ b/main/linux-vserver/patch-3.0.10-vs2.3.2.1-unofficial.diff @@ -0,0 +1,26517 @@ +diff -NurpP --minimal linux-3.0.9/Documentation/vserver/debug.txt linux-3.0.9-vs2.3.2.1/Documentation/vserver/debug.txt +--- linux-3.0.9/Documentation/vserver/debug.txt 1970-01-01 01:00:00.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/Documentation/vserver/debug.txt 2011-06-10 22:11:24.000000000 +0200 +@@ -0,0 +1,154 @@ ++ ++debug_cvirt: ++ ++ 2 4 "vx_map_tgid: %p/%llx: %d -> %d" ++ "vx_rmap_tgid: %p/%llx: %d -> %d" ++ ++debug_dlim: ++ ++ 0 1 "ALLOC (%p,#%d)%c inode (%d)" ++ "FREE (%p,#%d)%c inode" ++ 1 2 "ALLOC (%p,#%d)%c %lld bytes (%d)" ++ "FREE (%p,#%d)%c %lld bytes" ++ 2 4 "ADJUST: %lld,%lld on %ld,%ld [mult=%d]" ++ 3 8 "ext3_has_free_blocks(%p): %lu<%lu+1, %c, %u!=%u r=%d" ++ "ext3_has_free_blocks(%p): free=%lu, root=%lu" ++ "rcu_free_dl_info(%p)" ++ 4 10 "alloc_dl_info(%p,%d) = %p" ++ "dealloc_dl_info(%p)" ++ "get_dl_info(%p[#%d.%d])" ++ "put_dl_info(%p[#%d.%d])" ++ 5 20 "alloc_dl_info(%p,%d)*" ++ 6 40 "__hash_dl_info: %p[#%d]" ++ "__unhash_dl_info: %p[#%d]" ++ 7 80 "locate_dl_info(%p,#%d) = %p" ++ ++debug_misc: ++ ++ 0 1 "destroy_dqhash: %p [#0x%08x] c=%d" ++ "new_dqhash: %p [#0x%08x]" ++ "vroot[%d]_clr_dev: dev=%p[%lu,%d:%d]" ++ "vroot[%d]_get_real_bdev: dev=%p[%lu,%d:%d]" ++ "vroot[%d]_set_dev: dev=%p[%lu,%d:%d]" ++ "vroot_get_real_bdev not set" ++ 1 2 "cow_break_link(»%s«)" ++ "temp copy »%s«" ++ 2 4 "dentry_open(new): %p" ++ "dentry_open(old): %p" ++ "lookup_create(new): %p" ++ "old path »%s«" ++ "path_lookup(old): %d" ++ "vfs_create(new): %d" ++ "vfs_rename: %d" ++ "vfs_sendfile: %d" ++ 3 8 "fput(new_file=%p[#%d])" ++ "fput(old_file=%p[#%d])" ++ 4 10 "vx_info_kill(%p[#%d],%d,%d) = %d" ++ "vx_info_kill(%p[#%d],%d,%d)*" ++ 5 20 "vs_reboot(%p[#%d],%d)" ++ 6 40 "dropping task %p[#%u,%u] for %p[#%u,%u]" ++ ++debug_net: ++ ++ 2 4 "nx_addr_conflict(%p,%p) %d.%d,%d.%d" ++ 3 8 "inet_bind(%p) %d.%d.%d.%d, %d.%d.%d.%d, %d.%d.%d.%d" ++ "inet_bind(%p)* %p,%p;%lx %d.%d.%d.%d" ++ 4 10 "ip_route_connect(%p) %p,%p;%lx" ++ 5 20 "__addr_in_socket(%p,%d.%d.%d.%d) %p:%d.%d.%d.%d %p;%lx" ++ 6 40 "sk,egf: %p [#%d] (from %d)" ++ "sk,egn: %p [#%d] (from %d)" ++ "sk,req: %p [#%d] (from %d)" ++ "sk: %p [#%d] (from %d)" ++ "tw: %p [#%d] (from %d)" ++ 7 80 "__sock_recvmsg: %p[%p,%p,%p;%d]:%d/%d" ++ "__sock_sendmsg: %p[%p,%p,%p;%d]:%d/%d" ++ ++debug_nid: ++ ++ 0 1 "__lookup_nx_info(#%u): %p[#%u]" ++ "alloc_nx_info(%d) = %p" ++ "create_nx_info(%d) (dynamic rejected)" ++ "create_nx_info(%d) = %p (already there)" ++ "create_nx_info(%d) = %p (new)" ++ "dealloc_nx_info(%p)" ++ 1 2 "alloc_nx_info(%d)*" ++ "create_nx_info(%d)*" ++ 2 4 "get_nx_info(%p[#%d.%d])" ++ "put_nx_info(%p[#%d.%d])" ++ 3 8 "claim_nx_info(%p[#%d.%d.%d]) %p" ++ "clr_nx_info(%p[#%d.%d])" ++ "init_nx_info(%p[#%d.%d])" ++ "release_nx_info(%p[#%d.%d.%d]) %p" ++ "set_nx_info(%p[#%d.%d])" ++ 4 10 "__hash_nx_info: %p[#%d]" ++ "__nx_dynamic_id: [#%d]" ++ "__unhash_nx_info: %p[#%d.%d.%d]" ++ 5 20 "moved task %p into nxi:%p[#%d]" ++ "nx_migrate_task(%p,%p[#%d.%d.%d])" ++ "task_get_nx_info(%p)" ++ 6 40 "nx_clear_persistent(%p[#%d])" ++ ++debug_quota: ++ ++ 0 1 "quota_sync_dqh(%p,%d) discard inode %p" ++ 1 2 "quota_sync_dqh(%p,%d)" ++ "sync_dquots(%p,%d)" ++ "sync_dquots_dqh(%p,%d)" ++ 3 8 "do_quotactl(%p,%d,cmd=%d,id=%d,%p)" ++ ++debug_switch: ++ ++ 0 1 "vc: VCMD_%02d_%d[%d], %d,%p [%d,%d,%x,%x]" ++ 1 2 "vc: VCMD_%02d_%d[%d] = %08lx(%ld) [%d,%d]" ++ 4 10 "%s: (%s %s) returned %s with %d" ++ ++debug_tag: ++ ++ 7 80 "dx_parse_tag(»%s«): %d:#%d" ++ "dx_propagate_tag(%p[#%lu.%d]): %d,%d" ++ ++debug_xid: ++ ++ 0 1 "__lookup_vx_info(#%u): %p[#%u]" ++ "alloc_vx_info(%d) = %p" ++ "alloc_vx_info(%d)*" ++ "create_vx_info(%d) (dynamic rejected)" ++ "create_vx_info(%d) = %p (already there)" ++ "create_vx_info(%d) = %p (new)" ++ "dealloc_vx_info(%p)" ++ "loc_vx_info(%d) = %p (found)" ++ "loc_vx_info(%d) = %p (new)" ++ "loc_vx_info(%d) = %p (not available)" ++ 1 2 "create_vx_info(%d)*" ++ "loc_vx_info(%d)*" ++ 2 4 "get_vx_info(%p[#%d.%d])" ++ "put_vx_info(%p[#%d.%d])" ++ 3 8 "claim_vx_info(%p[#%d.%d.%d]) %p" ++ "clr_vx_info(%p[#%d.%d])" ++ "init_vx_info(%p[#%d.%d])" ++ "release_vx_info(%p[#%d.%d.%d]) %p" ++ "set_vx_info(%p[#%d.%d])" ++ 4 10 "__hash_vx_info: %p[#%d]" ++ "__unhash_vx_info: %p[#%d.%d.%d]" ++ "__vx_dynamic_id: [#%d]" ++ 5 20 "enter_vx_info(%p[#%d],%p) %p[#%d,%p]" ++ "leave_vx_info(%p[#%d,%p]) %p[#%d,%p]" ++ "moved task %p into vxi:%p[#%d]" ++ "task_get_vx_info(%p)" ++ "vx_migrate_task(%p,%p[#%d.%d])" ++ 6 40 "vx_clear_persistent(%p[#%d])" ++ "vx_exit_init(%p[#%d],%p[#%d,%d,%d])" ++ "vx_set_init(%p[#%d],%p[#%d,%d,%d])" ++ "vx_set_persistent(%p[#%d])" ++ "vx_set_reaper(%p[#%d],%p[#%d,%d])" ++ 7 80 "vx_child_reaper(%p[#%u,%u]) = %p[#%u,%u]" ++ ++ ++debug_limit: ++ ++ n 2^n "vx_acc_cres[%5d,%s,%2d]: %5d%s" ++ "vx_cres_avail[%5d,%s,%2d]: %5ld > %5d + %5d" ++ ++ m 2^m "vx_acc_page[%5d,%s,%2d]: %5d%s" ++ "vx_acc_pages[%5d,%s,%2d]: %5d += %5d" ++ "vx_pages_avail[%5d,%s,%2d]: %5ld > %5d + %5d" +diff -NurpP --minimal linux-3.0.9/Makefile linux-3.0.9-vs2.3.2.1/Makefile +--- linux-3.0.9/Makefile 2011-11-15 16:40:41.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/Makefile 2011-11-17 23:18:51.000000000 +0100 +@@ -1,7 +1,7 @@ + VERSION = 3 + PATCHLEVEL = 0 + SUBLEVEL = 10 +-EXTRAVERSION = ++EXTRAVERSION = -vs2.3.2.1 + NAME = Sneaky Weasel + + # *DOCUMENTATION* +diff -NurpP --minimal linux-3.0.9/arch/alpha/Kconfig linux-3.0.9-vs2.3.2.1/arch/alpha/Kconfig +--- linux-3.0.9/arch/alpha/Kconfig 2011-07-22 11:17:32.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/arch/alpha/Kconfig 2011-06-10 22:11:24.000000000 +0200 +@@ -668,6 +668,8 @@ config DUMMY_CONSOLE + depends on VGA_HOSE + default y + ++source "kernel/vserver/Kconfig" ++ + source "security/Kconfig" + + source "crypto/Kconfig" +diff -NurpP --minimal linux-3.0.9/arch/alpha/kernel/entry.S linux-3.0.9-vs2.3.2.1/arch/alpha/kernel/entry.S +--- linux-3.0.9/arch/alpha/kernel/entry.S 2010-10-21 13:06:45.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/arch/alpha/kernel/entry.S 2011-06-10 22:11:24.000000000 +0200 +@@ -860,24 +860,15 @@ sys_getxgid: + .globl sys_getxpid + .ent sys_getxpid + sys_getxpid: ++ lda $sp, -16($sp) ++ stq $26, 0($sp) + .prologue 0 +- ldq $2, TI_TASK($8) + +- /* See linux/kernel/timer.c sys_getppid for discussion +- about this loop. */ +- ldq $3, TASK_GROUP_LEADER($2) +- ldq $4, TASK_REAL_PARENT($3) +- ldl $0, TASK_TGID($2) +-1: ldl $1, TASK_TGID($4) +-#ifdef CONFIG_SMP +- mov $4, $5 +- mb +- ldq $3, TASK_GROUP_LEADER($2) +- ldq $4, TASK_REAL_PARENT($3) +- cmpeq $4, $5, $5 +- beq $5, 1b +-#endif +- stq $1, 80($sp) ++ lda $16, 96($sp) ++ jsr $26, do_getxpid ++ ldq $26, 0($sp) ++ ++ lda $sp, 16($sp) + ret + .end sys_getxpid + +diff -NurpP --minimal linux-3.0.9/arch/alpha/kernel/ptrace.c linux-3.0.9-vs2.3.2.1/arch/alpha/kernel/ptrace.c +--- linux-3.0.9/arch/alpha/kernel/ptrace.c 2011-01-05 21:48:40.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/arch/alpha/kernel/ptrace.c 2011-06-10 22:11:24.000000000 +0200 +@@ -13,6 +13,7 @@ + #include <linux/user.h> + #include <linux/security.h> + #include <linux/signal.h> ++#include <linux/vs_base.h> + + #include <asm/uaccess.h> + #include <asm/pgtable.h> +diff -NurpP --minimal linux-3.0.9/arch/alpha/kernel/systbls.S linux-3.0.9-vs2.3.2.1/arch/alpha/kernel/systbls.S +--- linux-3.0.9/arch/alpha/kernel/systbls.S 2011-07-22 11:17:32.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/arch/alpha/kernel/systbls.S 2011-06-10 22:11:24.000000000 +0200 +@@ -446,7 +446,7 @@ sys_call_table: + .quad sys_stat64 /* 425 */ + .quad sys_lstat64 + .quad sys_fstat64 +- .quad sys_ni_syscall /* sys_vserver */ ++ .quad sys_vserver /* sys_vserver */ + .quad sys_ni_syscall /* sys_mbind */ + .quad sys_ni_syscall /* sys_get_mempolicy */ + .quad sys_ni_syscall /* sys_set_mempolicy */ +diff -NurpP --minimal linux-3.0.9/arch/alpha/kernel/traps.c linux-3.0.9-vs2.3.2.1/arch/alpha/kernel/traps.c +--- linux-3.0.9/arch/alpha/kernel/traps.c 2010-10-21 13:06:46.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/arch/alpha/kernel/traps.c 2011-06-10 22:11:24.000000000 +0200 +@@ -183,7 +183,8 @@ die_if_kernel(char * str, struct pt_regs + #ifdef CONFIG_SMP + printk("CPU %d ", hard_smp_processor_id()); + #endif +- printk("%s(%d): %s %ld\n", current->comm, task_pid_nr(current), str, err); ++ printk("%s(%d[#%u]): %s %ld\n", current->comm, ++ task_pid_nr(current), current->xid, str, err); + dik_show_regs(regs, r9_15); + add_taint(TAINT_DIE); + dik_show_trace((unsigned long *)(regs+1)); +diff -NurpP --minimal linux-3.0.9/arch/arm/Kconfig linux-3.0.9-vs2.3.2.1/arch/arm/Kconfig +--- linux-3.0.9/arch/arm/Kconfig 2011-11-15 16:40:41.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/arch/arm/Kconfig 2011-10-18 13:51:13.000000000 +0200 +@@ -2049,6 +2049,8 @@ source "fs/Kconfig" + + source "arch/arm/Kconfig.debug" + ++source "kernel/vserver/Kconfig" ++ + source "security/Kconfig" + + source "crypto/Kconfig" +diff -NurpP --minimal linux-3.0.9/arch/arm/kernel/calls.S linux-3.0.9-vs2.3.2.1/arch/arm/kernel/calls.S +--- linux-3.0.9/arch/arm/kernel/calls.S 2011-07-22 11:17:32.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/arch/arm/kernel/calls.S 2011-06-10 22:11:24.000000000 +0200 +@@ -322,7 +322,7 @@ + /* 310 */ CALL(sys_request_key) + CALL(sys_keyctl) + CALL(ABI(sys_semtimedop, sys_oabi_semtimedop)) +-/* vserver */ CALL(sys_ni_syscall) ++ CALL(sys_vserver) + CALL(sys_ioprio_set) + /* 315 */ CALL(sys_ioprio_get) + CALL(sys_inotify_init) +diff -NurpP --minimal linux-3.0.9/arch/arm/kernel/process.c linux-3.0.9-vs2.3.2.1/arch/arm/kernel/process.c +--- linux-3.0.9/arch/arm/kernel/process.c 2011-05-22 16:16:47.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/arch/arm/kernel/process.c 2011-06-10 22:11:24.000000000 +0200 +@@ -315,7 +315,8 @@ void __show_regs(struct pt_regs *regs) + void show_regs(struct pt_regs * regs) + { + printk("\n"); +- printk("Pid: %d, comm: %20s\n", task_pid_nr(current), current->comm); ++ printk("Pid: %d[#%u], comm: %20s\n", ++ task_pid_nr(current), current->xid, current->comm); + __show_regs(regs); + __backtrace(); + } +diff -NurpP --minimal linux-3.0.9/arch/arm/kernel/traps.c linux-3.0.9-vs2.3.2.1/arch/arm/kernel/traps.c +--- linux-3.0.9/arch/arm/kernel/traps.c 2011-07-22 11:17:32.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/arch/arm/kernel/traps.c 2011-06-22 12:39:12.000000000 +0200 +@@ -242,8 +242,8 @@ static int __die(const char *str, int er + + print_modules(); + __show_regs(regs); +- printk(KERN_EMERG "Process %.*s (pid: %d, stack limit = 0x%p)\n", +- TASK_COMM_LEN, tsk->comm, task_pid_nr(tsk), thread + 1); ++ printk(KERN_EMERG "Process %.*s (pid: %d:#%u, stack limit = 0x%p)\n", ++ TASK_COMM_LEN, tsk->comm, task_pid_nr(tsk), tsk->xid, thread + 1); + + if (!user_mode(regs) || in_interrupt()) { + dump_mem(KERN_EMERG, "Stack: ", regs->ARM_sp, +diff -NurpP --minimal linux-3.0.9/arch/cris/Kconfig linux-3.0.9-vs2.3.2.1/arch/cris/Kconfig +--- linux-3.0.9/arch/cris/Kconfig 2011-07-22 11:17:35.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/arch/cris/Kconfig 2011-06-10 22:11:24.000000000 +0200 +@@ -678,6 +678,8 @@ source "drivers/staging/Kconfig" + + source "arch/cris/Kconfig.debug" + ++source "kernel/vserver/Kconfig" ++ + source "security/Kconfig" + + source "crypto/Kconfig" +diff -NurpP --minimal linux-3.0.9/arch/frv/kernel/kernel_thread.S linux-3.0.9-vs2.3.2.1/arch/frv/kernel/kernel_thread.S +--- linux-3.0.9/arch/frv/kernel/kernel_thread.S 2008-12-25 00:26:37.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/arch/frv/kernel/kernel_thread.S 2011-06-10 22:11:24.000000000 +0200 +@@ -37,7 +37,7 @@ kernel_thread: + + # start by forking the current process, but with shared VM + setlos.p #__NR_clone,gr7 ; syscall number +- ori gr10,#CLONE_VM,gr8 ; first syscall arg [clone_flags] ++ ori gr10,#CLONE_KT,gr8 ; first syscall arg [clone_flags] + sethi.p #0xe4e4,gr9 ; second syscall arg [newsp] + setlo #0xe4e4,gr9 + setlos.p #0,gr10 ; third syscall arg [parent_tidptr] +diff -NurpP --minimal linux-3.0.9/arch/h8300/Kconfig linux-3.0.9-vs2.3.2.1/arch/h8300/Kconfig +--- linux-3.0.9/arch/h8300/Kconfig 2011-07-22 11:17:35.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/arch/h8300/Kconfig 2011-06-10 22:11:24.000000000 +0200 +@@ -213,6 +213,8 @@ source "fs/Kconfig" + + source "arch/h8300/Kconfig.debug" + ++source "kernel/vserver/Kconfig" ++ + source "security/Kconfig" + + source "crypto/Kconfig" +diff -NurpP --minimal linux-3.0.9/arch/ia64/Kconfig linux-3.0.9-vs2.3.2.1/arch/ia64/Kconfig +--- linux-3.0.9/arch/ia64/Kconfig 2011-07-22 11:17:35.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/arch/ia64/Kconfig 2011-06-10 22:11:24.000000000 +0200 +@@ -671,6 +671,8 @@ source "fs/Kconfig" + + source "arch/ia64/Kconfig.debug" + ++source "kernel/vserver/Kconfig" ++ + source "security/Kconfig" + + source "crypto/Kconfig" +diff -NurpP --minimal linux-3.0.9/arch/ia64/include/asm/tlb.h linux-3.0.9-vs2.3.2.1/arch/ia64/include/asm/tlb.h +--- linux-3.0.9/arch/ia64/include/asm/tlb.h 2011-07-22 11:17:35.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/arch/ia64/include/asm/tlb.h 2011-06-10 22:11:24.000000000 +0200 +@@ -40,6 +40,7 @@ + #include <linux/mm.h> + #include <linux/pagemap.h> + #include <linux/swap.h> ++#include <linux/vs_memory.h> + + #include <asm/pgalloc.h> + #include <asm/processor.h> +diff -NurpP --minimal linux-3.0.9/arch/ia64/kernel/entry.S linux-3.0.9-vs2.3.2.1/arch/ia64/kernel/entry.S +--- linux-3.0.9/arch/ia64/kernel/entry.S 2011-07-22 11:17:35.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/arch/ia64/kernel/entry.S 2011-06-10 22:11:24.000000000 +0200 +@@ -1714,7 +1714,7 @@ sys_call_table: + data8 sys_mq_notify + data8 sys_mq_getsetattr + data8 sys_kexec_load +- data8 sys_ni_syscall // reserved for vserver ++ data8 sys_vserver + data8 sys_waitid // 1270 + data8 sys_add_key + data8 sys_request_key +diff -NurpP --minimal linux-3.0.9/arch/ia64/kernel/perfmon.c linux-3.0.9-vs2.3.2.1/arch/ia64/kernel/perfmon.c +--- linux-3.0.9/arch/ia64/kernel/perfmon.c 2011-03-15 18:06:39.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/arch/ia64/kernel/perfmon.c 2011-06-10 22:11:24.000000000 +0200 +@@ -42,6 +42,7 @@ + #include <linux/completion.h> + #include <linux/tracehook.h> + #include <linux/slab.h> ++#include <linux/vs_memory.h> + + #include <asm/errno.h> + #include <asm/intrinsics.h> +diff -NurpP --minimal linux-3.0.9/arch/ia64/kernel/process.c linux-3.0.9-vs2.3.2.1/arch/ia64/kernel/process.c +--- linux-3.0.9/arch/ia64/kernel/process.c 2011-03-15 18:06:39.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/arch/ia64/kernel/process.c 2011-06-10 22:11:24.000000000 +0200 +@@ -109,8 +109,8 @@ show_regs (struct pt_regs *regs) + unsigned long ip = regs->cr_iip + ia64_psr(regs)->ri; + + print_modules(); +- printk("\nPid: %d, CPU %d, comm: %20s\n", task_pid_nr(current), +- smp_processor_id(), current->comm); ++ printk("\nPid: %d[#%u], CPU %d, comm: %20s\n", task_pid_nr(current), ++ current->xid, smp_processor_id(), current->comm); + printk("psr : %016lx ifs : %016lx ip : [<%016lx>] %s (%s)\n", + regs->cr_ipsr, regs->cr_ifs, ip, print_tainted(), + init_utsname()->release); +diff -NurpP --minimal linux-3.0.9/arch/ia64/kernel/ptrace.c linux-3.0.9-vs2.3.2.1/arch/ia64/kernel/ptrace.c +--- linux-3.0.9/arch/ia64/kernel/ptrace.c 2011-01-05 21:48:59.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/arch/ia64/kernel/ptrace.c 2011-06-10 22:11:24.000000000 +0200 +@@ -21,6 +21,7 @@ + #include <linux/regset.h> + #include <linux/elf.h> + #include <linux/tracehook.h> ++#include <linux/vs_base.h> + + #include <asm/pgtable.h> + #include <asm/processor.h> +diff -NurpP --minimal linux-3.0.9/arch/ia64/kernel/traps.c linux-3.0.9-vs2.3.2.1/arch/ia64/kernel/traps.c +--- linux-3.0.9/arch/ia64/kernel/traps.c 2010-07-07 18:31:01.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/arch/ia64/kernel/traps.c 2011-06-10 22:11:24.000000000 +0200 +@@ -59,8 +59,9 @@ die (const char *str, struct pt_regs *re + put_cpu(); + + if (++die.lock_owner_depth < 3) { +- printk("%s[%d]: %s %ld [%d]\n", +- current->comm, task_pid_nr(current), str, err, ++die_counter); ++ printk("%s[%d[#%u]]: %s %ld [%d]\n", ++ current->comm, task_pid_nr(current), current->xid, ++ str, err, ++die_counter); + if (notify_die(DIE_OOPS, str, regs, err, 255, SIGSEGV) + != NOTIFY_STOP) + show_regs(regs); +@@ -323,8 +324,9 @@ handle_fpu_swa (int fp_fault, struct pt_ + if ((last.count & 15) < 5 && (ia64_fetchadd(1, &last.count, acq) & 15) < 5) { + last.time = current_jiffies + 5 * HZ; + printk(KERN_WARNING +- "%s(%d): floating-point assist fault at ip %016lx, isr %016lx\n", +- current->comm, task_pid_nr(current), regs->cr_iip + ia64_psr(regs)->ri, isr); ++ "%s(%d[#%u]): floating-point assist fault at ip %016lx, isr %016lx\n", ++ current->comm, task_pid_nr(current), current->xid, ++ regs->cr_iip + ia64_psr(regs)->ri, isr); + } + } + } +diff -NurpP --minimal linux-3.0.9/arch/ia64/mm/fault.c linux-3.0.9-vs2.3.2.1/arch/ia64/mm/fault.c +--- linux-3.0.9/arch/ia64/mm/fault.c 2011-07-22 11:17:35.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/arch/ia64/mm/fault.c 2011-06-10 22:28:23.000000000 +0200 +@@ -11,6 +11,7 @@ + #include <linux/kprobes.h> + #include <linux/kdebug.h> + #include <linux/prefetch.h> ++#include <linux/vs_memory.h> + + #include <asm/pgtable.h> + #include <asm/processor.h> +diff -NurpP --minimal linux-3.0.9/arch/m32r/kernel/traps.c linux-3.0.9-vs2.3.2.1/arch/m32r/kernel/traps.c +--- linux-3.0.9/arch/m32r/kernel/traps.c 2009-12-03 20:01:57.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/arch/m32r/kernel/traps.c 2011-06-10 22:11:24.000000000 +0200 +@@ -196,8 +196,9 @@ static void show_registers(struct pt_reg + } else { + printk("SPI: %08lx\n", sp); + } +- printk("Process %s (pid: %d, process nr: %d, stackpage=%08lx)", +- current->comm, task_pid_nr(current), 0xffff & i, 4096+(unsigned long)current); ++ printk("Process %s (pid: %d[#%u], process nr: %d, stackpage=%08lx)", ++ current->comm, task_pid_nr(current), current->xid, ++ 0xffff & i, 4096+(unsigned long)current); + + /* + * When in-kernel, we also print out the stack and code at the +diff -NurpP --minimal linux-3.0.9/arch/m68k/Kconfig linux-3.0.9-vs2.3.2.1/arch/m68k/Kconfig +--- linux-3.0.9/arch/m68k/Kconfig 2011-07-22 11:17:35.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/arch/m68k/Kconfig 2011-06-10 22:11:24.000000000 +0200 +@@ -241,6 +241,8 @@ source "fs/Kconfig" + + source "arch/m68k/Kconfig.debug" + ++source "kernel/vserver/Kconfig" ++ + source "security/Kconfig" + + source "crypto/Kconfig" +diff -NurpP --minimal linux-3.0.9/arch/mips/Kconfig linux-3.0.9-vs2.3.2.1/arch/mips/Kconfig +--- linux-3.0.9/arch/mips/Kconfig 2011-07-22 11:17:35.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/arch/mips/Kconfig 2011-06-10 22:11:24.000000000 +0200 +@@ -2485,6 +2485,8 @@ source "fs/Kconfig" + + source "arch/mips/Kconfig.debug" + ++source "kernel/vserver/Kconfig" ++ + source "security/Kconfig" + + source "crypto/Kconfig" +diff -NurpP --minimal linux-3.0.9/arch/mips/kernel/ptrace.c linux-3.0.9-vs2.3.2.1/arch/mips/kernel/ptrace.c +--- linux-3.0.9/arch/mips/kernel/ptrace.c 2011-07-22 11:17:36.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/arch/mips/kernel/ptrace.c 2011-06-10 22:11:24.000000000 +0200 +@@ -25,6 +25,7 @@ + #include <linux/security.h> + #include <linux/audit.h> + #include <linux/seccomp.h> ++#include <linux/vs_base.h> + + #include <asm/byteorder.h> + #include <asm/cpu.h> +@@ -263,6 +264,9 @@ long arch_ptrace(struct task_struct *chi + void __user *datavp = (void __user *) data; + unsigned long __user *datalp = (void __user *) data; + ++ if (!vx_check(vx_task_xid(child), VS_WATCH_P | VS_IDENT)) ++ goto out; ++ + switch (request) { + /* when I and D space are separate, these will need to be fixed. */ + case PTRACE_PEEKTEXT: /* read word at location addr. */ +diff -NurpP --minimal linux-3.0.9/arch/mips/kernel/scall32-o32.S linux-3.0.9-vs2.3.2.1/arch/mips/kernel/scall32-o32.S +--- linux-3.0.9/arch/mips/kernel/scall32-o32.S 2011-07-22 11:17:36.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/arch/mips/kernel/scall32-o32.S 2011-06-10 22:11:24.000000000 +0200 +@@ -523,7 +523,7 @@ einval: li v0, -ENOSYS + sys sys_mq_timedreceive 5 + sys sys_mq_notify 2 /* 4275 */ + sys sys_mq_getsetattr 3 +- sys sys_ni_syscall 0 /* sys_vserver */ ++ sys sys_vserver 3 + sys sys_waitid 5 + sys sys_ni_syscall 0 /* available, was setaltroot */ + sys sys_add_key 5 /* 4280 */ +diff -NurpP --minimal linux-3.0.9/arch/mips/kernel/scall64-64.S linux-3.0.9-vs2.3.2.1/arch/mips/kernel/scall64-64.S +--- linux-3.0.9/arch/mips/kernel/scall64-64.S 2011-07-22 11:17:36.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/arch/mips/kernel/scall64-64.S 2011-06-10 22:11:24.000000000 +0200 +@@ -362,7 +362,7 @@ sys_call_table: + PTR sys_mq_timedreceive + PTR sys_mq_notify + PTR sys_mq_getsetattr /* 5235 */ +- PTR sys_ni_syscall /* sys_vserver */ ++ PTR sys_vserver + PTR sys_waitid + PTR sys_ni_syscall /* available, was setaltroot */ + PTR sys_add_key +diff -NurpP --minimal linux-3.0.9/arch/mips/kernel/scall64-n32.S linux-3.0.9-vs2.3.2.1/arch/mips/kernel/scall64-n32.S +--- linux-3.0.9/arch/mips/kernel/scall64-n32.S 2011-07-22 11:17:36.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/arch/mips/kernel/scall64-n32.S 2011-06-10 22:11:24.000000000 +0200 +@@ -361,7 +361,7 @@ EXPORT(sysn32_call_table) + PTR compat_sys_mq_timedreceive + PTR compat_sys_mq_notify + PTR compat_sys_mq_getsetattr +- PTR sys_ni_syscall /* 6240, sys_vserver */ ++ PTR sys32_vserver /* 6240 */ + PTR compat_sys_waitid + PTR sys_ni_syscall /* available, was setaltroot */ + PTR sys_add_key +diff -NurpP --minimal linux-3.0.9/arch/mips/kernel/scall64-o32.S linux-3.0.9-vs2.3.2.1/arch/mips/kernel/scall64-o32.S +--- linux-3.0.9/arch/mips/kernel/scall64-o32.S 2011-07-22 11:17:36.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/arch/mips/kernel/scall64-o32.S 2011-06-10 22:11:24.000000000 +0200 +@@ -480,7 +480,7 @@ sys_call_table: + PTR compat_sys_mq_timedreceive + PTR compat_sys_mq_notify /* 4275 */ + PTR compat_sys_mq_getsetattr +- PTR sys_ni_syscall /* sys_vserver */ ++ PTR sys32_vserver + PTR sys_32_waitid + PTR sys_ni_syscall /* available, was setaltroot */ + PTR sys_add_key /* 4280 */ +diff -NurpP --minimal linux-3.0.9/arch/mips/kernel/traps.c linux-3.0.9-vs2.3.2.1/arch/mips/kernel/traps.c +--- linux-3.0.9/arch/mips/kernel/traps.c 2011-05-22 16:17:00.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/arch/mips/kernel/traps.c 2011-06-10 22:11:24.000000000 +0200 +@@ -343,9 +343,10 @@ void show_registers(struct pt_regs *regs + + __show_regs(regs); + print_modules(); +- printk("Process %s (pid: %d, threadinfo=%p, task=%p, tls=%0*lx)\n", +- current->comm, current->pid, current_thread_info(), current, +- field, current_thread_info()->tp_value); ++ printk("Process %s (pid: %d:#%u, threadinfo=%p, task=%p, tls=%0*lx)\n", ++ current->comm, task_pid_nr(current), current->xid, ++ current_thread_info(), current, ++ field, current_thread_info()->tp_value); + if (cpu_has_userlocal) { + unsigned long tls; + +diff -NurpP --minimal linux-3.0.9/arch/parisc/Kconfig linux-3.0.9-vs2.3.2.1/arch/parisc/Kconfig +--- linux-3.0.9/arch/parisc/Kconfig 2011-07-22 11:17:36.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/arch/parisc/Kconfig 2011-06-10 22:11:24.000000000 +0200 +@@ -279,6 +279,8 @@ source "fs/Kconfig" + + source "arch/parisc/Kconfig.debug" + ++source "kernel/vserver/Kconfig" ++ + source "security/Kconfig" + + source "crypto/Kconfig" +diff -NurpP --minimal linux-3.0.9/arch/parisc/kernel/syscall_table.S linux-3.0.9-vs2.3.2.1/arch/parisc/kernel/syscall_table.S +--- linux-3.0.9/arch/parisc/kernel/syscall_table.S 2011-11-15 16:40:42.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/arch/parisc/kernel/syscall_table.S 2011-08-29 03:45:07.000000000 +0200 +@@ -361,7 +361,7 @@ + ENTRY_COMP(mbind) /* 260 */ + ENTRY_COMP(get_mempolicy) + ENTRY_COMP(set_mempolicy) +- ENTRY_SAME(ni_syscall) /* 263: reserved for vserver */ ++ ENTRY_DIFF(vserver) + ENTRY_SAME(add_key) + ENTRY_SAME(request_key) /* 265 */ + ENTRY_SAME(keyctl) +diff -NurpP --minimal linux-3.0.9/arch/parisc/kernel/traps.c linux-3.0.9-vs2.3.2.1/arch/parisc/kernel/traps.c +--- linux-3.0.9/arch/parisc/kernel/traps.c 2009-09-10 15:25:40.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/arch/parisc/kernel/traps.c 2011-06-10 22:11:24.000000000 +0200 +@@ -236,8 +236,9 @@ void die_if_kernel(char *str, struct pt_ + if (err == 0) + return; /* STFU */ + +- printk(KERN_CRIT "%s (pid %d): %s (code %ld) at " RFMT "\n", +- current->comm, task_pid_nr(current), str, err, regs->iaoq[0]); ++ printk(KERN_CRIT "%s (pid %d:#%u): %s (code %ld) at " RFMT "\n", ++ current->comm, task_pid_nr(current), current->xid, ++ str, err, regs->iaoq[0]); + #ifdef PRINT_USER_FAULTS + /* XXX for debugging only */ + show_regs(regs); +@@ -270,8 +271,8 @@ void die_if_kernel(char *str, struct pt_ + pdc_console_restart(); + + if (err) +- printk(KERN_CRIT "%s (pid %d): %s (code %ld)\n", +- current->comm, task_pid_nr(current), str, err); ++ printk(KERN_CRIT "%s (pid %d:#%u): %s (code %ld)\n", ++ current->comm, task_pid_nr(current), current->xid, str, err); + + /* Wot's wrong wif bein' racy? */ + if (current->thread.flags & PARISC_KERNEL_DEATH) { +diff -NurpP --minimal linux-3.0.9/arch/parisc/mm/fault.c linux-3.0.9-vs2.3.2.1/arch/parisc/mm/fault.c +--- linux-3.0.9/arch/parisc/mm/fault.c 2010-08-02 16:52:06.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/arch/parisc/mm/fault.c 2011-06-10 22:11:24.000000000 +0200 +@@ -237,8 +237,9 @@ bad_area: + + #ifdef PRINT_USER_FAULTS + printk(KERN_DEBUG "\n"); +- printk(KERN_DEBUG "do_page_fault() pid=%d command='%s' type=%lu address=0x%08lx\n", +- task_pid_nr(tsk), tsk->comm, code, address); ++ printk(KERN_DEBUG "do_page_fault() pid=%d:#%u " ++ "command='%s' type=%lu address=0x%08lx\n", ++ task_pid_nr(tsk), tsk->xid, tsk->comm, code, address); + if (vma) { + printk(KERN_DEBUG "vm_start = 0x%08lx, vm_end = 0x%08lx\n", + vma->vm_start, vma->vm_end); +diff -NurpP --minimal linux-3.0.9/arch/powerpc/Kconfig linux-3.0.9-vs2.3.2.1/arch/powerpc/Kconfig +--- linux-3.0.9/arch/powerpc/Kconfig 2011-07-22 11:17:36.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/arch/powerpc/Kconfig 2011-06-10 22:11:24.000000000 +0200 +@@ -978,6 +978,8 @@ source "lib/Kconfig" + + source "arch/powerpc/Kconfig.debug" + ++source "kernel/vserver/Kconfig" ++ + source "security/Kconfig" + + config KEYS_COMPAT +diff -NurpP --minimal linux-3.0.9/arch/powerpc/include/asm/unistd.h linux-3.0.9-vs2.3.2.1/arch/powerpc/include/asm/unistd.h +--- linux-3.0.9/arch/powerpc/include/asm/unistd.h 2011-07-22 11:17:40.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/arch/powerpc/include/asm/unistd.h 2011-06-10 22:11:24.000000000 +0200 +@@ -275,7 +275,7 @@ + #endif + #define __NR_rtas 255 + #define __NR_sys_debug_setcontext 256 +-/* Number 257 is reserved for vserver */ ++#define __NR_vserver 257 + #define __NR_migrate_pages 258 + #define __NR_mbind 259 + #define __NR_get_mempolicy 260 +diff -NurpP --minimal linux-3.0.9/arch/powerpc/kernel/process.c linux-3.0.9-vs2.3.2.1/arch/powerpc/kernel/process.c +--- linux-3.0.9/arch/powerpc/kernel/process.c 2011-07-22 11:17:40.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/arch/powerpc/kernel/process.c 2011-06-10 22:11:24.000000000 +0200 +@@ -656,8 +656,9 @@ void show_regs(struct pt_regs * regs) + #else + printk("DAR: "REG", DSISR: %08lx\n", regs->dar, regs->dsisr); + #endif +- printk("TASK = %p[%d] '%s' THREAD: %p", +- current, task_pid_nr(current), current->comm, task_thread_info(current)); ++ printk("TASK = %p[%d,#%u] '%s' THREAD: %p", ++ current, task_pid_nr(current), current->xid, ++ current->comm, task_thread_info(current)); + + #ifdef CONFIG_SMP + printk(" CPU: %d", raw_smp_processor_id()); +diff -NurpP --minimal linux-3.0.9/arch/powerpc/kernel/traps.c linux-3.0.9-vs2.3.2.1/arch/powerpc/kernel/traps.c +--- linux-3.0.9/arch/powerpc/kernel/traps.c 2011-07-22 11:17:40.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/arch/powerpc/kernel/traps.c 2011-07-19 00:44:39.000000000 +0200 +@@ -1075,8 +1075,9 @@ void nonrecoverable_exception(struct pt_ + + void trace_syscall(struct pt_regs *regs) + { +- printk("Task: %p(%d), PC: %08lX/%08lX, Syscall: %3ld, Result: %s%ld %s\n", +- current, task_pid_nr(current), regs->nip, regs->link, regs->gpr[0], ++ printk("Task: %p(%d[#%u]), PC: %08lX/%08lX, Syscall: %3ld, Result: %s%ld %s\n", ++ current, task_pid_nr(current), current->xid, ++ regs->nip, regs->link, regs->gpr[0], + regs->ccr&0x10000000?"Error=":"", regs->gpr[3], print_tainted()); + } + +diff -NurpP --minimal linux-3.0.9/arch/powerpc/kernel/vdso.c linux-3.0.9-vs2.3.2.1/arch/powerpc/kernel/vdso.c +--- linux-3.0.9/arch/powerpc/kernel/vdso.c 2011-05-22 16:17:02.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/arch/powerpc/kernel/vdso.c 2011-06-10 22:11:24.000000000 +0200 +@@ -23,6 +23,7 @@ + #include <linux/security.h> + #include <linux/bootmem.h> + #include <linux/memblock.h> ++#include <linux/vs_memory.h> + + #include <asm/pgtable.h> + #include <asm/system.h> +diff -NurpP --minimal linux-3.0.9/arch/s390/Kconfig linux-3.0.9-vs2.3.2.1/arch/s390/Kconfig +--- linux-3.0.9/arch/s390/Kconfig 2011-07-22 11:17:41.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/arch/s390/Kconfig 2011-07-01 11:35:34.000000000 +0200 +@@ -628,6 +628,8 @@ source "fs/Kconfig" + + source "arch/s390/Kconfig.debug" + ++source "kernel/vserver/Kconfig" ++ + source "security/Kconfig" + + source "crypto/Kconfig" +diff -NurpP --minimal linux-3.0.9/arch/s390/include/asm/tlb.h linux-3.0.9-vs2.3.2.1/arch/s390/include/asm/tlb.h +--- linux-3.0.9/arch/s390/include/asm/tlb.h 2011-07-22 11:17:41.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/arch/s390/include/asm/tlb.h 2011-06-15 02:40:14.000000000 +0200 +@@ -24,6 +24,8 @@ + #include <linux/mm.h> + #include <linux/pagemap.h> + #include <linux/swap.h> ++#include <linux/vs_memory.h> ++ + #include <asm/processor.h> + #include <asm/pgalloc.h> + #include <asm/tlbflush.h> +diff -NurpP --minimal linux-3.0.9/arch/s390/include/asm/unistd.h linux-3.0.9-vs2.3.2.1/arch/s390/include/asm/unistd.h +--- linux-3.0.9/arch/s390/include/asm/unistd.h 2011-07-22 11:17:41.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/arch/s390/include/asm/unistd.h 2011-06-10 22:11:24.000000000 +0200 +@@ -202,7 +202,7 @@ + #define __NR_clock_gettime (__NR_timer_create+6) + #define __NR_clock_getres (__NR_timer_create+7) + #define __NR_clock_nanosleep (__NR_timer_create+8) +-/* Number 263 is reserved for vserver */ ++#define __NR_vserver 263 + #define __NR_statfs64 265 + #define __NR_fstatfs64 266 + #define __NR_remap_file_pages 267 +diff -NurpP --minimal linux-3.0.9/arch/s390/kernel/ptrace.c linux-3.0.9-vs2.3.2.1/arch/s390/kernel/ptrace.c +--- linux-3.0.9/arch/s390/kernel/ptrace.c 2011-11-15 16:40:42.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/arch/s390/kernel/ptrace.c 2011-11-15 17:37:04.000000000 +0100 +@@ -20,6 +20,7 @@ + #include <linux/regset.h> + #include <linux/tracehook.h> + #include <linux/seccomp.h> ++#include <linux/vs_base.h> + #include <trace/syscall.h> + #include <asm/compat.h> + #include <asm/segment.h> +diff -NurpP --minimal linux-3.0.9/arch/s390/kernel/syscalls.S linux-3.0.9-vs2.3.2.1/arch/s390/kernel/syscalls.S +--- linux-3.0.9/arch/s390/kernel/syscalls.S 2011-07-22 11:17:41.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/arch/s390/kernel/syscalls.S 2011-06-10 22:11:24.000000000 +0200 +@@ -271,7 +271,7 @@ SYSCALL(sys_clock_settime,sys_clock_sett + SYSCALL(sys_clock_gettime,sys_clock_gettime,sys32_clock_gettime_wrapper) /* 260 */ + SYSCALL(sys_clock_getres,sys_clock_getres,sys32_clock_getres_wrapper) + SYSCALL(sys_clock_nanosleep,sys_clock_nanosleep,sys32_clock_nanosleep_wrapper) +-NI_SYSCALL /* reserved for vserver */ ++SYSCALL(sys_vserver,sys_vserver,sys32_vserver) + SYSCALL(sys_s390_fadvise64_64,sys_ni_syscall,sys32_fadvise64_64_wrapper) + SYSCALL(sys_statfs64,sys_statfs64,compat_sys_statfs64_wrapper) + SYSCALL(sys_fstatfs64,sys_fstatfs64,compat_sys_fstatfs64_wrapper) +diff -NurpP --minimal linux-3.0.9/arch/sh/Kconfig linux-3.0.9-vs2.3.2.1/arch/sh/Kconfig +--- linux-3.0.9/arch/sh/Kconfig 2011-07-22 11:17:41.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/arch/sh/Kconfig 2011-07-19 00:44:39.000000000 +0200 +@@ -893,6 +893,8 @@ source "fs/Kconfig" + + source "arch/sh/Kconfig.debug" + ++source "kernel/vserver/Kconfig" ++ + source "security/Kconfig" + + source "crypto/Kconfig" +diff -NurpP --minimal linux-3.0.9/arch/sh/kernel/irq.c linux-3.0.9-vs2.3.2.1/arch/sh/kernel/irq.c +--- linux-3.0.9/arch/sh/kernel/irq.c 2011-07-22 11:17:41.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/arch/sh/kernel/irq.c 2011-07-19 00:45:06.000000000 +0200 +@@ -14,6 +14,7 @@ + #include <linux/ftrace.h> + #include <linux/delay.h> + #include <linux/ratelimit.h> ++// #include <linux/vs_context.h> + #include <asm/processor.h> + #include <asm/machvec.h> + #include <asm/uaccess.h> +diff -NurpP --minimal linux-3.0.9/arch/sh/kernel/vsyscall/vsyscall.c linux-3.0.9-vs2.3.2.1/arch/sh/kernel/vsyscall/vsyscall.c +--- linux-3.0.9/arch/sh/kernel/vsyscall/vsyscall.c 2011-05-22 16:17:07.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/arch/sh/kernel/vsyscall/vsyscall.c 2011-06-10 22:11:24.000000000 +0200 +@@ -18,6 +18,7 @@ + #include <linux/elf.h> + #include <linux/sched.h> + #include <linux/err.h> ++#include <linux/vs_memory.h> + + /* + * Should the kernel map a VDSO page into processes and pass its +diff -NurpP --minimal linux-3.0.9/arch/sparc/Kconfig linux-3.0.9-vs2.3.2.1/arch/sparc/Kconfig +--- linux-3.0.9/arch/sparc/Kconfig 2011-11-15 16:40:42.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/arch/sparc/Kconfig 2011-10-18 13:51:13.000000000 +0200 +@@ -602,6 +602,8 @@ source "fs/Kconfig" + + source "arch/sparc/Kconfig.debug" + ++source "kernel/vserver/Kconfig" ++ + source "security/Kconfig" + + source "crypto/Kconfig" +diff -NurpP --minimal linux-3.0.9/arch/sparc/include/asm/tlb_64.h linux-3.0.9-vs2.3.2.1/arch/sparc/include/asm/tlb_64.h +--- linux-3.0.9/arch/sparc/include/asm/tlb_64.h 2011-07-22 11:17:42.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/arch/sparc/include/asm/tlb_64.h 2011-06-10 22:11:24.000000000 +0200 +@@ -3,6 +3,7 @@ + + #include <linux/swap.h> + #include <linux/pagemap.h> ++#include <linux/vs_memory.h> + #include <asm/pgalloc.h> + #include <asm/tlbflush.h> + #include <asm/mmu_context.h> +diff -NurpP --minimal linux-3.0.9/arch/sparc/include/asm/unistd.h linux-3.0.9-vs2.3.2.1/arch/sparc/include/asm/unistd.h +--- linux-3.0.9/arch/sparc/include/asm/unistd.h 2011-07-22 11:17:42.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/arch/sparc/include/asm/unistd.h 2011-06-10 22:11:24.000000000 +0200 +@@ -335,7 +335,7 @@ + #define __NR_timer_getoverrun 264 + #define __NR_timer_delete 265 + #define __NR_timer_create 266 +-/* #define __NR_vserver 267 Reserved for VSERVER */ ++#define __NR_vserver 267 + #define __NR_io_setup 268 + #define __NR_io_destroy 269 + #define __NR_io_submit 270 +diff -NurpP --minimal linux-3.0.9/arch/sparc/kernel/systbls_32.S linux-3.0.9-vs2.3.2.1/arch/sparc/kernel/systbls_32.S +--- linux-3.0.9/arch/sparc/kernel/systbls_32.S 2011-07-22 11:17:42.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/arch/sparc/kernel/systbls_32.S 2011-06-10 22:11:24.000000000 +0200 +@@ -70,7 +70,7 @@ sys_call_table: + /*250*/ .long sys_mremap, sys_sysctl, sys_getsid, sys_fdatasync, sys_nfsservctl + /*255*/ .long sys_sync_file_range, sys_clock_settime, sys_clock_gettime, sys_clock_getres, sys_clock_nanosleep + /*260*/ .long sys_sched_getaffinity, sys_sched_setaffinity, sys_timer_settime, sys_timer_gettime, sys_timer_getoverrun +-/*265*/ .long sys_timer_delete, sys_timer_create, sys_nis_syscall, sys_io_setup, sys_io_destroy ++/*265*/ .long sys_timer_delete, sys_timer_create, sys_vserver, sys_io_setup, sys_io_destroy + /*270*/ .long sys_io_submit, sys_io_cancel, sys_io_getevents, sys_mq_open, sys_mq_unlink + /*275*/ .long sys_mq_timedsend, sys_mq_timedreceive, sys_mq_notify, sys_mq_getsetattr, sys_waitid + /*280*/ .long sys_tee, sys_add_key, sys_request_key, sys_keyctl, sys_openat +diff -NurpP --minimal linux-3.0.9/arch/sparc/kernel/systbls_64.S linux-3.0.9-vs2.3.2.1/arch/sparc/kernel/systbls_64.S +--- linux-3.0.9/arch/sparc/kernel/systbls_64.S 2011-07-22 11:17:42.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/arch/sparc/kernel/systbls_64.S 2011-06-10 22:11:24.000000000 +0200 +@@ -71,7 +71,7 @@ sys_call_table32: + /*250*/ .word sys_mremap, compat_sys_sysctl, sys32_getsid, sys_fdatasync, sys32_nfsservctl + .word sys32_sync_file_range, compat_sys_clock_settime, compat_sys_clock_gettime, compat_sys_clock_getres, sys32_clock_nanosleep + /*260*/ .word compat_sys_sched_getaffinity, compat_sys_sched_setaffinity, sys32_timer_settime, compat_sys_timer_gettime, sys_timer_getoverrun +- .word sys_timer_delete, compat_sys_timer_create, sys_ni_syscall, compat_sys_io_setup, sys_io_destroy ++ .word sys_timer_delete, compat_sys_timer_create, sys32_vserver, compat_sys_io_setup, sys_io_destroy + /*270*/ .word sys32_io_submit, sys_io_cancel, compat_sys_io_getevents, sys32_mq_open, sys_mq_unlink + .word compat_sys_mq_timedsend, compat_sys_mq_timedreceive, compat_sys_mq_notify, compat_sys_mq_getsetattr, compat_sys_waitid + /*280*/ .word sys32_tee, sys_add_key, sys_request_key, sys_keyctl, compat_sys_openat +@@ -148,7 +148,7 @@ sys_call_table: + /*250*/ .word sys_64_mremap, sys_sysctl, sys_getsid, sys_fdatasync, sys_nfsservctl + .word sys_sync_file_range, sys_clock_settime, sys_clock_gettime, sys_clock_getres, sys_clock_nanosleep + /*260*/ .word sys_sched_getaffinity, sys_sched_setaffinity, sys_timer_settime, sys_timer_gettime, sys_timer_getoverrun +- .word sys_timer_delete, sys_timer_create, sys_ni_syscall, sys_io_setup, sys_io_destroy ++ .word sys_timer_delete, sys_timer_create, sys_vserver, sys_io_setup, sys_io_destroy + /*270*/ .word sys_io_submit, sys_io_cancel, sys_io_getevents, sys_mq_open, sys_mq_unlink + .word sys_mq_timedsend, sys_mq_timedreceive, sys_mq_notify, sys_mq_getsetattr, sys_waitid + /*280*/ .word sys_tee, sys_add_key, sys_request_key, sys_keyctl, sys_openat +diff -NurpP --minimal linux-3.0.9/arch/um/Kconfig.rest linux-3.0.9-vs2.3.2.1/arch/um/Kconfig.rest +--- linux-3.0.9/arch/um/Kconfig.rest 2009-06-11 17:12:19.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/arch/um/Kconfig.rest 2011-06-10 22:11:24.000000000 +0200 +@@ -18,6 +18,8 @@ source "drivers/connector/Kconfig" + + source "fs/Kconfig" + ++source "kernel/vserver/Kconfig" ++ + source "security/Kconfig" + + source "crypto/Kconfig" +diff -NurpP --minimal linux-3.0.9/arch/um/include/asm/tlb.h linux-3.0.9-vs2.3.2.1/arch/um/include/asm/tlb.h +--- linux-3.0.9/arch/um/include/asm/tlb.h 2011-07-22 11:17:42.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/arch/um/include/asm/tlb.h 2011-06-10 22:11:24.000000000 +0200 +@@ -3,6 +3,7 @@ + + #include <linux/pagemap.h> + #include <linux/swap.h> ++#include <linux/vs_memory.h> + #include <asm/percpu.h> + #include <asm/pgalloc.h> + #include <asm/tlbflush.h> +diff -NurpP --minimal linux-3.0.9/arch/um/include/shared/kern_constants.h linux-3.0.9-vs2.3.2.1/arch/um/include/shared/kern_constants.h +--- linux-3.0.9/arch/um/include/shared/kern_constants.h 1970-01-01 01:00:00.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/arch/um/include/shared/kern_constants.h 2011-06-10 22:11:24.000000000 +0200 +@@ -0,0 +1 @@ ++#include "../../../../include/generated/asm-offsets.h" +diff -NurpP --minimal linux-3.0.9/arch/um/include/shared/user_constants.h linux-3.0.9-vs2.3.2.1/arch/um/include/shared/user_constants.h +--- linux-3.0.9/arch/um/include/shared/user_constants.h 1970-01-01 01:00:00.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/arch/um/include/shared/user_constants.h 2011-06-10 22:11:24.000000000 +0200 +@@ -0,0 +1,40 @@ ++/* ++ * DO NOT MODIFY. ++ * ++ * This file was generated by arch/um/Makefile ++ * ++ */ ++ ++#define HOST_SC_CR2 176 /* offsetof(struct sigcontext, cr2) # */ ++#define HOST_SC_ERR 152 /* offsetof(struct sigcontext, err) # */ ++#define HOST_SC_TRAPNO 160 /* offsetof(struct sigcontext, trapno) # */ ++#define HOST_FP_SIZE 64 /* sizeof(struct _fpstate) / sizeof(unsigned long) # */ ++#define HOST_RBX 5 /* RBX # */ ++#define HOST_RCX 11 /* RCX # */ ++#define HOST_RDI 14 /* RDI # */ ++#define HOST_RSI 13 /* RSI # */ ++#define HOST_RDX 12 /* RDX # */ ++#define HOST_RBP 4 /* RBP # */ ++#define HOST_RAX 10 /* RAX # */ ++#define HOST_R8 9 /* R8 # */ ++#define HOST_R9 8 /* R9 # */ ++#define HOST_R10 7 /* R10 # */ ++#define HOST_R11 6 /* R11 # */ ++#define HOST_R12 3 /* R12 # */ ++#define HOST_R13 2 /* R13 # */ ++#define HOST_R14 1 /* R14 # */ ++#define HOST_R15 0 /* R15 # */ ++#define HOST_ORIG_RAX 15 /* ORIG_RAX # */ ++#define HOST_CS 17 /* CS # */ ++#define HOST_SS 20 /* SS # */ ++#define HOST_EFLAGS 18 /* EFLAGS # */ ++#define HOST_IP 16 /* RIP # */ ++#define HOST_SP 19 /* RSP # */ ++#define UM_FRAME_SIZE 216 /* sizeof(struct user_regs_struct) # */ ++#define UM_POLLIN 1 /* POLLIN # */ ++#define UM_POLLPRI 2 /* POLLPRI # */ ++#define UM_POLLOUT 4 /* POLLOUT # */ ++#define UM_PROT_READ 1 /* PROT_READ # */ ++#define UM_PROT_WRITE 2 /* PROT_WRITE # */ ++#define UM_PROT_EXEC 4 /* PROT_EXEC # */ ++ +diff -NurpP --minimal linux-3.0.9/arch/x86/Kconfig linux-3.0.9-vs2.3.2.1/arch/x86/Kconfig +--- linux-3.0.9/arch/x86/Kconfig 2011-07-22 11:17:42.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/arch/x86/Kconfig 2011-07-22 11:20:39.000000000 +0200 +@@ -2159,6 +2159,8 @@ source "fs/Kconfig" + + source "arch/x86/Kconfig.debug" + ++source "kernel/vserver/Kconfig" ++ + source "security/Kconfig" + + source "crypto/Kconfig" +diff -NurpP --minimal linux-3.0.9/arch/x86/ia32/ia32entry.S linux-3.0.9-vs2.3.2.1/arch/x86/ia32/ia32entry.S +--- linux-3.0.9/arch/x86/ia32/ia32entry.S 2011-07-22 11:17:42.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/arch/x86/ia32/ia32entry.S 2011-06-10 22:11:24.000000000 +0200 +@@ -776,7 +776,7 @@ ia32_sys_call_table: + .quad sys_tgkill /* 270 */ + .quad compat_sys_utimes + .quad sys32_fadvise64_64 +- .quad quiet_ni_syscall /* sys_vserver */ ++ .quad sys32_vserver + .quad sys_mbind + .quad compat_sys_get_mempolicy /* 275 */ + .quad sys_set_mempolicy +diff -NurpP --minimal linux-3.0.9/arch/x86/include/asm/unistd_64.h linux-3.0.9-vs2.3.2.1/arch/x86/include/asm/unistd_64.h +--- linux-3.0.9/arch/x86/include/asm/unistd_64.h 2011-07-22 11:17:43.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/arch/x86/include/asm/unistd_64.h 2011-06-10 22:11:24.000000000 +0200 +@@ -535,7 +535,7 @@ __SYSCALL(__NR_tgkill, sys_tgkill) + #define __NR_utimes 235 + __SYSCALL(__NR_utimes, sys_utimes) + #define __NR_vserver 236 +-__SYSCALL(__NR_vserver, sys_ni_syscall) ++__SYSCALL(__NR_vserver, sys_vserver) + #define __NR_mbind 237 + __SYSCALL(__NR_mbind, sys_mbind) + #define __NR_set_mempolicy 238 +diff -NurpP --minimal linux-3.0.9/arch/x86/kernel/syscall_table_32.S linux-3.0.9-vs2.3.2.1/arch/x86/kernel/syscall_table_32.S +--- linux-3.0.9/arch/x86/kernel/syscall_table_32.S 2011-07-22 11:17:43.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/arch/x86/kernel/syscall_table_32.S 2011-06-10 22:11:24.000000000 +0200 +@@ -272,7 +272,7 @@ ENTRY(sys_call_table) + .long sys_tgkill /* 270 */ + .long sys_utimes + .long sys_fadvise64_64 +- .long sys_ni_syscall /* sys_vserver */ ++ .long sys_vserver + .long sys_mbind + .long sys_get_mempolicy + .long sys_set_mempolicy +diff -NurpP --minimal linux-3.0.9/block/genhd.c linux-3.0.9-vs2.3.2.1/block/genhd.c +--- linux-3.0.9/block/genhd.c 2011-11-15 16:40:42.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/block/genhd.c 2011-11-15 17:37:04.000000000 +0100 +@@ -1162,17 +1162,17 @@ static int diskstats_show(struct seq_fil + cpu = part_stat_lock(); + part_round_stats(cpu, hd); + part_stat_unlock(); +- seq_printf(seqf, "%4d %7d %s %lu %lu %llu " +- "%u %lu %lu %llu %u %u %u %u\n", ++ seq_printf(seqf, "%4d %7d %s %lu %lu %lu " ++ "%u %lu %lu %lu %u %u %u %u\n", + MAJOR(part_devt(hd)), MINOR(part_devt(hd)), + disk_name(gp, hd->partno, buf), + part_stat_read(hd, ios[READ]), + part_stat_read(hd, merges[READ]), +- (unsigned long long)part_stat_read(hd, sectors[READ]), ++ part_stat_read(hd, sectors[READ]), + jiffies_to_msecs(part_stat_read(hd, ticks[READ])), + part_stat_read(hd, ios[WRITE]), + part_stat_read(hd, merges[WRITE]), +- (unsigned long long)part_stat_read(hd, sectors[WRITE]), ++ part_stat_read(hd, sectors[WRITE]), + jiffies_to_msecs(part_stat_read(hd, ticks[WRITE])), + part_in_flight(hd), + jiffies_to_msecs(part_stat_read(hd, io_ticks)), +diff -NurpP --minimal linux-3.0.9/drivers/block/Kconfig linux-3.0.9-vs2.3.2.1/drivers/block/Kconfig +--- linux-3.0.9/drivers/block/Kconfig 2011-07-22 11:17:44.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/drivers/block/Kconfig 2011-06-10 22:11:24.000000000 +0200 +@@ -273,6 +273,13 @@ config BLK_DEV_CRYPTOLOOP + + source "drivers/block/drbd/Kconfig" + ++config BLK_DEV_VROOT ++ tristate "Virtual Root device support" ++ depends on QUOTACTL ++ ---help--- ++ Saying Y here will allow you to use quota/fs ioctls on a shared ++ partition within a virtual server without compromising security. ++ + config BLK_DEV_NBD + tristate "Network block device support" + depends on NET +diff -NurpP --minimal linux-3.0.9/drivers/block/Makefile linux-3.0.9-vs2.3.2.1/drivers/block/Makefile +--- linux-3.0.9/drivers/block/Makefile 2011-07-22 11:17:44.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/drivers/block/Makefile 2011-06-10 22:11:24.000000000 +0200 +@@ -34,6 +34,7 @@ obj-$(CONFIG_VIODASD) += viodasd.o + obj-$(CONFIG_BLK_DEV_SX8) += sx8.o + obj-$(CONFIG_BLK_DEV_UB) += ub.o + obj-$(CONFIG_BLK_DEV_HD) += hd.o ++obj-$(CONFIG_BLK_DEV_VROOT) += vroot.o + + obj-$(CONFIG_XEN_BLKDEV_FRONTEND) += xen-blkfront.o + obj-$(CONFIG_XEN_BLKDEV_BACKEND) += xen-blkback/ +diff -NurpP --minimal linux-3.0.9/drivers/block/loop.c linux-3.0.9-vs2.3.2.1/drivers/block/loop.c +--- linux-3.0.9/drivers/block/loop.c 2011-11-15 16:40:42.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/drivers/block/loop.c 2011-08-31 19:37:44.000000000 +0200 +@@ -75,6 +75,7 @@ + #include <linux/kthread.h> + #include <linux/splice.h> + #include <linux/sysfs.h> ++#include <linux/vs_context.h> + + #include <asm/uaccess.h> + +@@ -891,6 +892,7 @@ static int loop_set_fd(struct loop_devic + lo->lo_blocksize = lo_blocksize; + lo->lo_device = bdev; + lo->lo_flags = lo_flags; ++ lo->lo_xid = vx_current_xid(); + lo->lo_backing_file = file; + lo->transfer = transfer_none; + lo->ioctl = NULL; +@@ -1021,6 +1023,7 @@ static int loop_clr_fd(struct loop_devic + lo->lo_encrypt_key_size = 0; + lo->lo_flags = 0; + lo->lo_thread = NULL; ++ lo->lo_xid = 0; + memset(lo->lo_encrypt_key, 0, LO_KEY_SIZE); + memset(lo->lo_crypt_name, 0, LO_NAME_SIZE); + memset(lo->lo_file_name, 0, LO_NAME_SIZE); +@@ -1059,7 +1062,7 @@ loop_set_status(struct loop_device *lo, + + if (lo->lo_encrypt_key_size && + lo->lo_key_owner != uid && +- !capable(CAP_SYS_ADMIN)) ++ !vx_capable(CAP_SYS_ADMIN, VXC_ADMIN_CLOOP)) + return -EPERM; + if (lo->lo_state != Lo_bound) + return -ENXIO; +@@ -1143,7 +1146,8 @@ loop_get_status(struct loop_device *lo, + memcpy(info->lo_crypt_name, lo->lo_crypt_name, LO_NAME_SIZE); + info->lo_encrypt_type = + lo->lo_encryption ? lo->lo_encryption->number : 0; +- if (lo->lo_encrypt_key_size && capable(CAP_SYS_ADMIN)) { ++ if (lo->lo_encrypt_key_size && ++ vx_capable(CAP_SYS_ADMIN, VXC_ADMIN_CLOOP)) { + info->lo_encrypt_key_size = lo->lo_encrypt_key_size; + memcpy(info->lo_encrypt_key, lo->lo_encrypt_key, + lo->lo_encrypt_key_size); +@@ -1489,6 +1493,9 @@ static int lo_open(struct block_device * + { + struct loop_device *lo = bdev->bd_disk->private_data; + ++ if (!vx_check(lo->lo_xid, VS_IDENT|VS_HOSTID|VS_ADMIN_P)) ++ return -EACCES; ++ + mutex_lock(&lo->lo_ctl_mutex); + lo->lo_refcnt++; + mutex_unlock(&lo->lo_ctl_mutex); +diff -NurpP --minimal linux-3.0.9/drivers/block/vroot.c linux-3.0.9-vs2.3.2.1/drivers/block/vroot.c +--- linux-3.0.9/drivers/block/vroot.c 1970-01-01 01:00:00.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/drivers/block/vroot.c 2011-06-10 22:11:24.000000000 +0200 +@@ -0,0 +1,292 @@ ++/* ++ * linux/drivers/block/vroot.c ++ * ++ * written by Herbert Pötzl, 9/11/2002 ++ * ported to 2.6.10 by Herbert Pötzl, 30/12/2004 ++ * ++ * based on the loop.c code by Theodore Ts'o. ++ * ++ * Copyright (C) 2002-2007 by Herbert Pötzl. ++ * Redistribution of this file is permitted under the ++ * GNU General Public License. ++ * ++ */ ++ ++#include <linux/module.h> ++#include <linux/moduleparam.h> ++#include <linux/file.h> ++#include <linux/major.h> ++#include <linux/blkdev.h> ++#include <linux/slab.h> ++ ++#include <linux/vroot.h> ++#include <linux/vs_context.h> ++ ++ ++static int max_vroot = 8; ++ ++static struct vroot_device *vroot_dev; ++static struct gendisk **disks; ++ ++ ++static int vroot_set_dev( ++ struct vroot_device *vr, ++ struct block_device *bdev, ++ unsigned int arg) ++{ ++ struct block_device *real_bdev; ++ struct file *file; ++ struct inode *inode; ++ int error; ++ ++ error = -EBUSY; ++ if (vr->vr_state != Vr_unbound) ++ goto out; ++ ++ error = -EBADF; ++ file = fget(arg); ++ if (!file) ++ goto out; ++ ++ error = -EINVAL; ++ inode = file->f_dentry->d_inode; ++ ++ ++ if (S_ISBLK(inode->i_mode)) { ++ real_bdev = inode->i_bdev; ++ vr->vr_device = real_bdev; ++ __iget(real_bdev->bd_inode); ++ } else ++ goto out_fput; ++ ++ vxdprintk(VXD_CBIT(misc, 0), ++ "vroot[%d]_set_dev: dev=" VXF_DEV, ++ vr->vr_number, VXD_DEV(real_bdev)); ++ ++ vr->vr_state = Vr_bound; ++ error = 0; ++ ++ out_fput: ++ fput(file); ++ out: ++ return error; ++} ++ ++static int vroot_clr_dev( ++ struct vroot_device *vr, ++ struct block_device *bdev) ++{ ++ struct block_device *real_bdev; ++ ++ if (vr->vr_state != Vr_bound) ++ return -ENXIO; ++ if (vr->vr_refcnt > 1) /* we needed one fd for the ioctl */ ++ return -EBUSY; ++ ++ real_bdev = vr->vr_device; ++ ++ vxdprintk(VXD_CBIT(misc, 0), ++ "vroot[%d]_clr_dev: dev=" VXF_DEV, ++ vr->vr_number, VXD_DEV(real_bdev)); ++ ++ bdput(real_bdev); ++ vr->vr_state = Vr_unbound; ++ vr->vr_device = NULL; ++ return 0; ++} ++ ++ ++static int vr_ioctl(struct block_device *bdev, fmode_t mode, ++ unsigned int cmd, unsigned long arg) ++{ ++ struct vroot_device *vr = bdev->bd_disk->private_data; ++ int err; ++ ++ down(&vr->vr_ctl_mutex); ++ switch (cmd) { ++ case VROOT_SET_DEV: ++ err = vroot_set_dev(vr, bdev, arg); ++ break; ++ case VROOT_CLR_DEV: ++ err = vroot_clr_dev(vr, bdev); ++ break; ++ default: ++ err = -EINVAL; ++ break; ++ } ++ up(&vr->vr_ctl_mutex); ++ return err; ++} ++ ++static int vr_open(struct block_device *bdev, fmode_t mode) ++{ ++ struct vroot_device *vr = bdev->bd_disk->private_data; ++ ++ down(&vr->vr_ctl_mutex); ++ vr->vr_refcnt++; ++ up(&vr->vr_ctl_mutex); ++ return 0; ++} ++ ++static int vr_release(struct gendisk *disk, fmode_t mode) ++{ ++ struct vroot_device *vr = disk->private_data; ++ ++ down(&vr->vr_ctl_mutex); ++ --vr->vr_refcnt; ++ up(&vr->vr_ctl_mutex); ++ return 0; ++} ++ ++static struct block_device_operations vr_fops = { ++ .owner = THIS_MODULE, ++ .open = vr_open, ++ .release = vr_release, ++ .ioctl = vr_ioctl, ++}; ++ ++static int vroot_make_request(struct request_queue *q, struct bio *bio) ++{ ++ printk("vroot_make_request %p, %p\n", q, bio); ++ bio_io_error(bio); ++ return 0; ++} ++ ++struct block_device *__vroot_get_real_bdev(struct block_device *bdev) ++{ ++ struct inode *inode = bdev->bd_inode; ++ struct vroot_device *vr; ++ struct block_device *real_bdev; ++ int minor = iminor(inode); ++ ++ vr = &vroot_dev[minor]; ++ real_bdev = vr->vr_device; ++ ++ vxdprintk(VXD_CBIT(misc, 0), ++ "vroot[%d]_get_real_bdev: dev=" VXF_DEV, ++ vr->vr_number, VXD_DEV(real_bdev)); ++ ++ if (vr->vr_state != Vr_bound) ++ return ERR_PTR(-ENXIO); ++ ++ __iget(real_bdev->bd_inode); ++ return real_bdev; ++} ++ ++ ++ ++/* ++ * And now the modules code and kernel interface. ++ */ ++ ++module_param(max_vroot, int, 0); ++ ++MODULE_PARM_DESC(max_vroot, "Maximum number of vroot devices (1-256)"); ++MODULE_LICENSE("GPL"); ++MODULE_ALIAS_BLOCKDEV_MAJOR(VROOT_MAJOR); ++ ++MODULE_AUTHOR ("Herbert Pötzl"); ++MODULE_DESCRIPTION ("Virtual Root Device Mapper"); ++ ++ ++int __init vroot_init(void) ++{ ++ int err, i; ++ ++ if (max_vroot < 1 || max_vroot > 256) { ++ max_vroot = MAX_VROOT_DEFAULT; ++ printk(KERN_WARNING "vroot: invalid max_vroot " ++ "(must be between 1 and 256), " ++ "using default (%d)\n", max_vroot); ++ } ++ ++ if (register_blkdev(VROOT_MAJOR, "vroot")) ++ return -EIO; ++ ++ err = -ENOMEM; ++ vroot_dev = kmalloc(max_vroot * sizeof(struct vroot_device), GFP_KERNEL); ++ if (!vroot_dev) ++ goto out_mem1; ++ memset(vroot_dev, 0, max_vroot * sizeof(struct vroot_device)); ++ ++ disks = kmalloc(max_vroot * sizeof(struct gendisk *), GFP_KERNEL); ++ if (!disks) ++ goto out_mem2; ++ ++ for (i = 0; i < max_vroot; i++) { ++ disks[i] = alloc_disk(1); ++ if (!disks[i]) ++ goto out_mem3; ++ disks[i]->queue = blk_alloc_queue(GFP_KERNEL); ++ if (!disks[i]->queue) ++ goto out_mem3; ++ blk_queue_make_request(disks[i]->queue, vroot_make_request); ++ } ++ ++ for (i = 0; i < max_vroot; i++) { ++ struct vroot_device *vr = &vroot_dev[i]; ++ struct gendisk *disk = disks[i]; ++ ++ memset(vr, 0, sizeof(*vr)); ++ sema_init(&vr->vr_ctl_mutex, 1); ++ vr->vr_number = i; ++ disk->major = VROOT_MAJOR; ++ disk->first_minor = i; ++ disk->fops = &vr_fops; ++ sprintf(disk->disk_name, "vroot%d", i); ++ disk->private_data = vr; ++ } ++ ++ err = register_vroot_grb(&__vroot_get_real_bdev); ++ if (err) ++ goto out_mem3; ++ ++ for (i = 0; i < max_vroot; i++) ++ add_disk(disks[i]); ++ printk(KERN_INFO "vroot: loaded (max %d devices)\n", max_vroot); ++ return 0; ++ ++out_mem3: ++ while (i--) ++ put_disk(disks[i]); ++ kfree(disks); ++out_mem2: ++ kfree(vroot_dev); ++out_mem1: ++ unregister_blkdev(VROOT_MAJOR, "vroot"); ++ printk(KERN_ERR "vroot: ran out of memory\n"); ++ return err; ++} ++ ++void vroot_exit(void) ++{ ++ int i; ++ ++ if (unregister_vroot_grb(&__vroot_get_real_bdev)) ++ printk(KERN_WARNING "vroot: cannot unregister grb\n"); ++ ++ for (i = 0; i < max_vroot; i++) { ++ del_gendisk(disks[i]); ++ put_disk(disks[i]); ++ } ++ unregister_blkdev(VROOT_MAJOR, "vroot"); ++ ++ kfree(disks); ++ kfree(vroot_dev); ++} ++ ++module_init(vroot_init); ++module_exit(vroot_exit); ++ ++#ifndef MODULE ++ ++static int __init max_vroot_setup(char *str) ++{ ++ max_vroot = simple_strtol(str, NULL, 0); ++ return 1; ++} ++ ++__setup("max_vroot=", max_vroot_setup); ++ ++#endif ++ +diff -NurpP --minimal linux-3.0.9/drivers/infiniband/core/addr.c linux-3.0.9-vs2.3.2.1/drivers/infiniband/core/addr.c +--- linux-3.0.9/drivers/infiniband/core/addr.c 2011-07-22 11:17:45.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/drivers/infiniband/core/addr.c 2011-06-16 14:16:51.000000000 +0200 +@@ -252,7 +252,7 @@ static int addr6_resolve(struct sockaddr + + if (ipv6_addr_any(&fl6.saddr)) { + ret = ipv6_dev_get_saddr(&init_net, ip6_dst_idev(dst)->dev, +- &fl6.daddr, 0, &fl6.saddr); ++ &fl6.daddr, 0, &fl6.saddr, NULL); + if (ret) + goto put; + +diff -NurpP --minimal linux-3.0.9/drivers/infiniband/hw/ipath/ipath_user_pages.c linux-3.0.9-vs2.3.2.1/drivers/infiniband/hw/ipath/ipath_user_pages.c +--- linux-3.0.9/drivers/infiniband/hw/ipath/ipath_user_pages.c 2011-05-22 16:17:16.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/drivers/infiniband/hw/ipath/ipath_user_pages.c 2011-06-10 22:11:24.000000000 +0200 +@@ -35,6 +35,7 @@ + #include <linux/device.h> + #include <linux/slab.h> + #include <linux/sched.h> ++#include <linux/vs_memory.h> + + #include "ipath_kernel.h" + +diff -NurpP --minimal linux-3.0.9/drivers/md/dm-ioctl.c linux-3.0.9-vs2.3.2.1/drivers/md/dm-ioctl.c +--- linux-3.0.9/drivers/md/dm-ioctl.c 2011-05-22 16:17:18.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/drivers/md/dm-ioctl.c 2011-06-10 22:11:24.000000000 +0200 +@@ -16,6 +16,7 @@ + #include <linux/dm-ioctl.h> + #include <linux/hdreg.h> + #include <linux/compat.h> ++#include <linux/vs_context.h> + + #include <asm/uaccess.h> + +@@ -106,7 +107,8 @@ static struct hash_cell *__get_name_cell + unsigned int h = hash_str(str); + + list_for_each_entry (hc, _name_buckets + h, name_list) +- if (!strcmp(hc->name, str)) { ++ if (vx_check(dm_get_xid(hc->md), VS_WATCH_P | VS_IDENT) && ++ !strcmp(hc->name, str)) { + dm_get(hc->md); + return hc; + } +@@ -120,7 +122,8 @@ static struct hash_cell *__get_uuid_cell + unsigned int h = hash_str(str); + + list_for_each_entry (hc, _uuid_buckets + h, uuid_list) +- if (!strcmp(hc->uuid, str)) { ++ if (vx_check(dm_get_xid(hc->md), VS_WATCH_P | VS_IDENT) && ++ !strcmp(hc->uuid, str)) { + dm_get(hc->md); + return hc; + } +@@ -427,6 +430,9 @@ typedef int (*ioctl_fn)(struct dm_ioctl + + static int remove_all(struct dm_ioctl *param, size_t param_size) + { ++ if (!vx_check(0, VS_ADMIN)) ++ return -EPERM; ++ + dm_hash_remove_all(1); + param->data_size = 0; + return 0; +@@ -474,6 +480,8 @@ static int list_devices(struct dm_ioctl + */ + for (i = 0; i < NUM_BUCKETS; i++) { + list_for_each_entry (hc, _name_buckets + i, name_list) { ++ if (!vx_check(dm_get_xid(hc->md), VS_WATCH_P | VS_IDENT)) ++ continue; + needed += sizeof(struct dm_name_list); + needed += strlen(hc->name) + 1; + needed += ALIGN_MASK; +@@ -497,6 +505,8 @@ static int list_devices(struct dm_ioctl + */ + for (i = 0; i < NUM_BUCKETS; i++) { + list_for_each_entry (hc, _name_buckets + i, name_list) { ++ if (!vx_check(dm_get_xid(hc->md), VS_WATCH_P | VS_IDENT)) ++ continue; + if (old_nl) + old_nl->next = (uint32_t) ((void *) nl - + (void *) old_nl); +@@ -731,10 +741,11 @@ static struct hash_cell *__find_device_h + if (!md) + goto out; + +- mdptr = dm_get_mdptr(md); ++ if (vx_check(dm_get_xid(md), VS_WATCH_P | VS_IDENT)) ++ mdptr = dm_get_mdptr(md); ++ + if (!mdptr) + dm_put(md); +- + out: + return mdptr; + } +@@ -1577,8 +1588,8 @@ static int ctl_ioctl(uint command, struc + ioctl_fn fn = NULL; + size_t input_param_size; + +- /* only root can play with this */ +- if (!capable(CAP_SYS_ADMIN)) ++ /* only root and certain contexts can play with this */ ++ if (!vx_capable(CAP_SYS_ADMIN, VXC_ADMIN_MAPPER)) + return -EACCES; + + if (_IOC_TYPE(command) != DM_IOCTL) +diff -NurpP --minimal linux-3.0.9/drivers/md/dm.c linux-3.0.9-vs2.3.2.1/drivers/md/dm.c +--- linux-3.0.9/drivers/md/dm.c 2011-11-15 16:40:44.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/drivers/md/dm.c 2011-08-08 23:04:47.000000000 +0200 +@@ -20,6 +20,7 @@ + #include <linux/idr.h> + #include <linux/hdreg.h> + #include <linux/delay.h> ++#include <linux/vs_base.h> + + #include <trace/events/block.h> + +@@ -121,6 +122,7 @@ struct mapped_device { + rwlock_t map_lock; + atomic_t holders; + atomic_t open_count; ++ xid_t xid; + + unsigned long flags; + +@@ -334,6 +336,7 @@ int dm_deleting_md(struct mapped_device + static int dm_blk_open(struct block_device *bdev, fmode_t mode) + { + struct mapped_device *md; ++ int ret = -ENXIO; + + spin_lock(&_minor_lock); + +@@ -342,18 +345,19 @@ static int dm_blk_open(struct block_devi + goto out; + + if (test_bit(DMF_FREEING, &md->flags) || +- dm_deleting_md(md)) { +- md = NULL; ++ dm_deleting_md(md)) ++ goto out; ++ ++ ret = -EACCES; ++ if (!vx_check(md->xid, VS_IDENT|VS_HOSTID)) + goto out; +- } + + dm_get(md); + atomic_inc(&md->open_count); +- ++ ret = 0; + out: + spin_unlock(&_minor_lock); +- +- return md ? 0 : -ENXIO; ++ return ret; + } + + static int dm_blk_close(struct gendisk *disk, fmode_t mode) +@@ -574,6 +578,14 @@ int dm_set_geometry(struct mapped_device + return 0; + } + ++/* ++ * Get the xid associated with a dm device ++ */ ++xid_t dm_get_xid(struct mapped_device *md) ++{ ++ return md->xid; ++} ++ + /*----------------------------------------------------------------- + * CRUD START: + * A more elegant soln is in the works that uses the queue +@@ -1847,6 +1859,7 @@ static struct mapped_device *alloc_dev(i + INIT_LIST_HEAD(&md->uevent_list); + spin_lock_init(&md->uevent_lock); + ++ md->xid = vx_current_xid(); + md->queue = blk_alloc_queue(GFP_KERNEL); + if (!md->queue) + goto bad_queue; +diff -NurpP --minimal linux-3.0.9/drivers/md/dm.h linux-3.0.9-vs2.3.2.1/drivers/md/dm.h +--- linux-3.0.9/drivers/md/dm.h 2011-05-22 16:17:18.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/drivers/md/dm.h 2011-06-10 22:11:24.000000000 +0200 +@@ -41,6 +41,8 @@ struct dm_dev_internal { + struct dm_table; + struct dm_md_mempools; + ++xid_t dm_get_xid(struct mapped_device *md); ++ + /*----------------------------------------------------------------- + * Internal table functions. + *---------------------------------------------------------------*/ +diff -NurpP --minimal linux-3.0.9/drivers/net/tun.c linux-3.0.9-vs2.3.2.1/drivers/net/tun.c +--- linux-3.0.9/drivers/net/tun.c 2011-11-15 16:40:45.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/drivers/net/tun.c 2011-08-29 03:45:08.000000000 +0200 +@@ -64,6 +64,7 @@ + #include <linux/nsproxy.h> + #include <linux/virtio_net.h> + #include <linux/rcupdate.h> ++#include <linux/vs_network.h> + #include <net/net_namespace.h> + #include <net/netns/generic.h> + #include <net/rtnetlink.h> +@@ -121,6 +122,7 @@ struct tun_struct { + unsigned int flags; + uid_t owner; + gid_t group; ++ nid_t nid; + + struct net_device *dev; + u32 set_features; +@@ -905,6 +907,7 @@ static void tun_setup(struct net_device + + tun->owner = -1; + tun->group = -1; ++ tun->nid = current->nid; + + dev->ethtool_ops = &tun_ethtool_ops; + dev->destructor = tun_free_netdev; +@@ -1055,7 +1058,7 @@ static int tun_set_iff(struct net *net, + + if (((tun->owner != -1 && cred->euid != tun->owner) || + (tun->group != -1 && !in_egroup_p(tun->group))) && +- !capable(CAP_NET_ADMIN)) ++ !cap_raised(current_cap(), CAP_NET_ADMIN)) + return -EPERM; + err = security_tun_dev_attach(tun->socket.sk); + if (err < 0) +@@ -1069,7 +1072,7 @@ static int tun_set_iff(struct net *net, + char *name; + unsigned long flags = 0; + +- if (!capable(CAP_NET_ADMIN)) ++ if (!nx_capable(CAP_NET_ADMIN, NXC_TUN_CREATE)) + return -EPERM; + err = security_tun_dev_create(); + if (err < 0) +@@ -1137,6 +1140,9 @@ static int tun_set_iff(struct net *net, + + sk->sk_destruct = tun_sock_destruct; + ++ if (!nx_check(tun->nid, VS_IDENT | VS_HOSTID | VS_ADMIN_P)) ++ return -EPERM; ++ + err = tun_attach(tun, file); + if (err < 0) + goto failed; +@@ -1318,6 +1324,16 @@ static long __tun_chr_ioctl(struct file + tun_debug(KERN_INFO, tun, "group set to %d\n", tun->group); + break; + ++ case TUNSETNID: ++ if (!capable(CAP_CONTEXT)) ++ return -EPERM; ++ ++ /* Set nid owner of the device */ ++ tun->nid = (nid_t) arg; ++ ++ tun_debug(KERN_INFO, tun, "nid owner set to %u\n", tun->nid); ++ break; ++ + case TUNSETLINK: + /* Only allow setting the type when the interface is down */ + if (tun->dev->flags & IFF_UP) { +diff -NurpP --minimal linux-3.0.9/drivers/tty/sysrq.c linux-3.0.9-vs2.3.2.1/drivers/tty/sysrq.c +--- linux-3.0.9/drivers/tty/sysrq.c 2011-05-22 16:17:44.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/drivers/tty/sysrq.c 2011-06-10 22:11:24.000000000 +0200 +@@ -41,6 +41,7 @@ + #include <linux/oom.h> + #include <linux/slab.h> + #include <linux/input.h> ++#include <linux/vserver/debug.h> + + #include <asm/ptrace.h> + #include <asm/irq_regs.h> +@@ -395,6 +396,21 @@ static struct sysrq_key_op sysrq_unrt_op + .enable_mask = SYSRQ_ENABLE_RTNICE, + }; + ++ ++#ifdef CONFIG_VSERVER_DEBUG ++static void sysrq_handle_vxinfo(int key) ++{ ++ dump_vx_info_inactive((key == 'x') ? 0 : 1); ++} ++ ++static struct sysrq_key_op sysrq_showvxinfo_op = { ++ .handler = sysrq_handle_vxinfo, ++ .help_msg = "conteXt", ++ .action_msg = "Show Context Info", ++ .enable_mask = SYSRQ_ENABLE_DUMP, ++}; ++#endif ++ + /* Key Operations table and lock */ + static DEFINE_SPINLOCK(sysrq_key_table_lock); + +@@ -449,7 +465,11 @@ static struct sysrq_key_op *sysrq_key_ta + NULL, /* v */ + &sysrq_showstate_blocked_op, /* w */ + /* x: May be registered on ppc/powerpc for xmon */ ++#ifdef CONFIG_VSERVER_DEBUG ++ &sysrq_showvxinfo_op, /* x */ ++#else + NULL, /* x */ ++#endif + /* y: May be registered on sparc64 for global register dump */ + NULL, /* y */ + &sysrq_ftrace_dump_op, /* z */ +@@ -464,6 +484,8 @@ static int sysrq_key_table_key2index(int + retval = key - '0'; + else if ((key >= 'a') && (key <= 'z')) + retval = key + 10 - 'a'; ++ else if ((key >= 'A') && (key <= 'Z')) ++ retval = key + 10 - 'A'; + else + retval = -1; + return retval; +diff -NurpP --minimal linux-3.0.9/drivers/tty/tty_io.c linux-3.0.9-vs2.3.2.1/drivers/tty/tty_io.c +--- linux-3.0.9/drivers/tty/tty_io.c 2011-11-15 16:40:46.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/drivers/tty/tty_io.c 2011-11-15 17:37:05.000000000 +0100 +@@ -104,6 +104,7 @@ + + #include <linux/kmod.h> + #include <linux/nsproxy.h> ++#include <linux/vs_pid.h> + + #undef TTY_DEBUG_HANGUP + +@@ -2080,7 +2081,8 @@ static int tiocsti(struct tty_struct *tt + char ch, mbz = 0; + struct tty_ldisc *ld; + +- if ((current->signal->tty != tty) && !capable(CAP_SYS_ADMIN)) ++ if (((current->signal->tty != tty) && ++ !vx_capable(CAP_SYS_ADMIN, VXC_TIOCSTI))) + return -EPERM; + if (get_user(ch, p)) + return -EFAULT; +@@ -2368,6 +2370,7 @@ static int tiocspgrp(struct tty_struct * + return -ENOTTY; + if (get_user(pgrp_nr, p)) + return -EFAULT; ++ pgrp_nr = vx_rmap_pid(pgrp_nr); + if (pgrp_nr < 0) + return -EINVAL; + rcu_read_lock(); +diff -NurpP --minimal linux-3.0.9/fs/attr.c linux-3.0.9-vs2.3.2.1/fs/attr.c +--- linux-3.0.9/fs/attr.c 2011-07-22 11:18:05.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/fs/attr.c 2011-06-10 22:11:24.000000000 +0200 +@@ -13,6 +13,9 @@ + #include <linux/fsnotify.h> + #include <linux/fcntl.h> + #include <linux/security.h> ++#include <linux/proc_fs.h> ++#include <linux/devpts_fs.h> ++#include <linux/vs_tag.h> + + /** + * inode_change_ok - check if attribute changes to an inode are allowed +@@ -73,6 +76,10 @@ int inode_change_ok(const struct inode * + return -EPERM; + } + ++ /* check for inode tag permission */ ++ if (dx_permission(inode, MAY_WRITE)) ++ return -EACCES; ++ + return 0; + } + EXPORT_SYMBOL(inode_change_ok); +@@ -143,6 +150,8 @@ void setattr_copy(struct inode *inode, c + inode->i_uid = attr->ia_uid; + if (ia_valid & ATTR_GID) + inode->i_gid = attr->ia_gid; ++ if ((ia_valid & ATTR_TAG) && IS_TAGGED(inode)) ++ inode->i_tag = attr->ia_tag; + if (ia_valid & ATTR_ATIME) + inode->i_atime = timespec_trunc(attr->ia_atime, + inode->i_sb->s_time_gran); +@@ -170,7 +179,8 @@ int notify_change(struct dentry * dentry + struct timespec now; + unsigned int ia_valid = attr->ia_valid; + +- if (ia_valid & (ATTR_MODE | ATTR_UID | ATTR_GID | ATTR_TIMES_SET)) { ++ if (ia_valid & (ATTR_MODE | ATTR_UID | ATTR_GID | ++ ATTR_TAG | ATTR_TIMES_SET)) { + if (IS_IMMUTABLE(inode) || IS_APPEND(inode)) + return -EPERM; + } +diff -NurpP --minimal linux-3.0.9/fs/binfmt_aout.c linux-3.0.9-vs2.3.2.1/fs/binfmt_aout.c +--- linux-3.0.9/fs/binfmt_aout.c 2010-10-21 13:07:47.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/fs/binfmt_aout.c 2011-06-10 22:11:24.000000000 +0200 +@@ -25,6 +25,7 @@ + #include <linux/init.h> + #include <linux/coredump.h> + #include <linux/slab.h> ++#include <linux/vs_memory.h> + + #include <asm/system.h> + #include <asm/uaccess.h> +diff -NurpP --minimal linux-3.0.9/fs/binfmt_elf.c linux-3.0.9-vs2.3.2.1/fs/binfmt_elf.c +--- linux-3.0.9/fs/binfmt_elf.c 2011-11-15 16:40:46.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/fs/binfmt_elf.c 2011-11-15 17:37:06.000000000 +0100 +@@ -32,6 +32,7 @@ + #include <linux/elf.h> + #include <linux/utsname.h> + #include <linux/coredump.h> ++#include <linux/vs_memory.h> + #include <asm/uaccess.h> + #include <asm/param.h> + #include <asm/page.h> +diff -NurpP --minimal linux-3.0.9/fs/binfmt_flat.c linux-3.0.9-vs2.3.2.1/fs/binfmt_flat.c +--- linux-3.0.9/fs/binfmt_flat.c 2011-07-22 11:18:05.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/fs/binfmt_flat.c 2011-06-10 22:11:24.000000000 +0200 +@@ -35,6 +35,7 @@ + #include <linux/init.h> + #include <linux/flat.h> + #include <linux/syscalls.h> ++#include <linux/vs_memory.h> + + #include <asm/byteorder.h> + #include <asm/system.h> +diff -NurpP --minimal linux-3.0.9/fs/binfmt_som.c linux-3.0.9-vs2.3.2.1/fs/binfmt_som.c +--- linux-3.0.9/fs/binfmt_som.c 2010-02-25 11:52:04.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/fs/binfmt_som.c 2011-06-10 22:11:24.000000000 +0200 +@@ -28,6 +28,7 @@ + #include <linux/shm.h> + #include <linux/personality.h> + #include <linux/init.h> ++#include <linux/vs_memory.h> + + #include <asm/uaccess.h> + #include <asm/pgtable.h> +diff -NurpP --minimal linux-3.0.9/fs/block_dev.c linux-3.0.9-vs2.3.2.1/fs/block_dev.c +--- linux-3.0.9/fs/block_dev.c 2011-11-15 16:40:46.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/fs/block_dev.c 2011-11-15 17:37:06.000000000 +0100 +@@ -25,6 +25,7 @@ + #include <linux/namei.h> + #include <linux/log2.h> + #include <linux/kmemleak.h> ++#include <linux/vs_device.h> + #include <asm/uaccess.h> + #include "internal.h" + +@@ -553,6 +554,7 @@ struct block_device *bdget(dev_t dev) + bdev->bd_invalidated = 0; + inode->i_mode = S_IFBLK; + inode->i_rdev = dev; ++ inode->i_mdev = dev; + inode->i_bdev = bdev; + inode->i_data.a_ops = &def_blk_aops; + mapping_set_gfp_mask(&inode->i_data, GFP_USER); +@@ -599,6 +601,11 @@ EXPORT_SYMBOL(bdput); + static struct block_device *bd_acquire(struct inode *inode) + { + struct block_device *bdev; ++ dev_t mdev; ++ ++ if (!vs_map_blkdev(inode->i_rdev, &mdev, DATTR_OPEN)) ++ return NULL; ++ inode->i_mdev = mdev; + + spin_lock(&bdev_lock); + bdev = inode->i_bdev; +@@ -609,7 +616,7 @@ static struct block_device *bd_acquire(s + } + spin_unlock(&bdev_lock); + +- bdev = bdget(inode->i_rdev); ++ bdev = bdget(mdev); + if (bdev) { + spin_lock(&bdev_lock); + if (!inode->i_bdev) { +diff -NurpP --minimal linux-3.0.9/fs/btrfs/ctree.h linux-3.0.9-vs2.3.2.1/fs/btrfs/ctree.h +--- linux-3.0.9/fs/btrfs/ctree.h 2011-07-22 11:18:05.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/fs/btrfs/ctree.h 2011-07-19 00:44:39.000000000 +0200 +@@ -600,11 +600,14 @@ struct btrfs_inode_item { + /* modification sequence number for NFS */ + __le64 sequence; + ++ __le16 tag; + /* + * a little future expansion, for more than this we can + * just grow the inode item and version it + */ +- __le64 reserved[4]; ++ __le16 reserved16; ++ __le32 reserved32; ++ __le64 reserved[3]; + struct btrfs_timespec atime; + struct btrfs_timespec ctime; + struct btrfs_timespec mtime; +@@ -1359,6 +1362,8 @@ struct btrfs_ioctl_defrag_range_args { + #define BTRFS_MOUNT_AUTO_DEFRAG (1 << 16) + #define BTRFS_MOUNT_INODE_MAP_CACHE (1 << 17) + ++#define BTRFS_MOUNT_TAGGED (1 << 24) ++ + #define btrfs_clear_opt(o, opt) ((o) &= ~BTRFS_MOUNT_##opt) + #define btrfs_set_opt(o, opt) ((o) |= BTRFS_MOUNT_##opt) + #define btrfs_test_opt(root, opt) ((root)->fs_info->mount_opt & \ +@@ -1568,6 +1573,7 @@ BTRFS_SETGET_FUNCS(inode_block_group, st + BTRFS_SETGET_FUNCS(inode_nlink, struct btrfs_inode_item, nlink, 32); + BTRFS_SETGET_FUNCS(inode_uid, struct btrfs_inode_item, uid, 32); + BTRFS_SETGET_FUNCS(inode_gid, struct btrfs_inode_item, gid, 32); ++BTRFS_SETGET_FUNCS(inode_tag, struct btrfs_inode_item, tag, 16); + BTRFS_SETGET_FUNCS(inode_mode, struct btrfs_inode_item, mode, 32); + BTRFS_SETGET_FUNCS(inode_rdev, struct btrfs_inode_item, rdev, 64); + BTRFS_SETGET_FUNCS(inode_flags, struct btrfs_inode_item, flags, 64); +@@ -1621,6 +1627,10 @@ BTRFS_SETGET_FUNCS(extent_flags, struct + + BTRFS_SETGET_FUNCS(extent_refs_v0, struct btrfs_extent_item_v0, refs, 32); + ++#define BTRFS_INODE_IXUNLINK (1 << 24) ++#define BTRFS_INODE_BARRIER (1 << 25) ++#define BTRFS_INODE_COW (1 << 26) ++ + + BTRFS_SETGET_FUNCS(tree_block_level, struct btrfs_tree_block_info, level, 8); + +@@ -2595,6 +2605,7 @@ extern const struct dentry_operations bt + long btrfs_ioctl(struct file *file, unsigned int cmd, unsigned long arg); + void btrfs_update_iflags(struct inode *inode); + void btrfs_inherit_iflags(struct inode *inode, struct inode *dir); ++int btrfs_sync_flags(struct inode *inode, int, int); + int btrfs_defrag_file(struct inode *inode, struct file *file, + struct btrfs_ioctl_defrag_range_args *range, + u64 newer_than, unsigned long max_pages); +diff -NurpP --minimal linux-3.0.9/fs/btrfs/disk-io.c linux-3.0.9-vs2.3.2.1/fs/btrfs/disk-io.c +--- linux-3.0.9/fs/btrfs/disk-io.c 2011-07-22 11:18:05.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/fs/btrfs/disk-io.c 2011-06-22 12:39:15.000000000 +0200 +@@ -1765,6 +1765,9 @@ struct btrfs_root *open_ctree(struct sup + goto fail_alloc; + } + ++ if (btrfs_test_opt(tree_root, TAGGED)) ++ sb->s_flags |= MS_TAGGED; ++ + features = btrfs_super_incompat_flags(disk_super) & + ~BTRFS_FEATURE_INCOMPAT_SUPP; + if (features) { +diff -NurpP --minimal linux-3.0.9/fs/btrfs/inode.c linux-3.0.9-vs2.3.2.1/fs/btrfs/inode.c +--- linux-3.0.9/fs/btrfs/inode.c 2011-11-15 16:40:46.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/fs/btrfs/inode.c 2011-10-18 13:51:13.000000000 +0200 +@@ -38,6 +38,7 @@ + #include <linux/falloc.h> + #include <linux/slab.h> + #include <linux/ratelimit.h> ++#include <linux/vs_tag.h> + #include "compat.h" + #include "ctree.h" + #include "disk-io.h" +@@ -2508,6 +2509,8 @@ static void btrfs_read_locked_inode(stru + struct btrfs_key location; + int maybe_acls; + u32 rdev; ++ uid_t uid; ++ gid_t gid; + int ret; + bool filled = false; + +@@ -2540,8 +2543,13 @@ static void btrfs_read_locked_inode(stru + + inode->i_mode = btrfs_inode_mode(leaf, inode_item); + inode->i_nlink = btrfs_inode_nlink(leaf, inode_item); +- inode->i_uid = btrfs_inode_uid(leaf, inode_item); +- inode->i_gid = btrfs_inode_gid(leaf, inode_item); ++ ++ uid = btrfs_inode_uid(leaf, inode_item); ++ gid = btrfs_inode_gid(leaf, inode_item); ++ inode->i_uid = INOTAG_UID(DX_TAG(inode), uid, gid); ++ inode->i_gid = INOTAG_GID(DX_TAG(inode), uid, gid); ++ inode->i_tag = INOTAG_TAG(DX_TAG(inode), uid, gid, ++ btrfs_inode_tag(leaf, inode_item)); + btrfs_i_size_write(inode, btrfs_inode_size(leaf, inode_item)); + + tspec = btrfs_inode_atime(inode_item); +@@ -2624,6 +2632,9 @@ static void fill_inode_item(struct btrfs + struct btrfs_inode_item *item, + struct inode *inode) + { ++ uid_t uid = TAGINO_UID(DX_TAG(inode), inode->i_uid, inode->i_tag); ++ gid_t gid = TAGINO_GID(DX_TAG(inode), inode->i_gid, inode->i_tag); ++ + if (!leaf->map_token) + map_private_extent_buffer(leaf, (unsigned long)item, + sizeof(struct btrfs_inode_item), +@@ -2631,8 +2642,11 @@ static void fill_inode_item(struct btrfs + &leaf->map_start, &leaf->map_len, + KM_USER1); + +- btrfs_set_inode_uid(leaf, item, inode->i_uid); +- btrfs_set_inode_gid(leaf, item, inode->i_gid); ++ btrfs_set_inode_uid(leaf, item, uid); ++ btrfs_set_inode_gid(leaf, item, gid); ++#ifdef CONFIG_TAGGING_INTERN ++ btrfs_set_inode_tag(leaf, item, inode->i_tag); ++#endif + btrfs_set_inode_size(leaf, item, BTRFS_I(inode)->disk_i_size); + btrfs_set_inode_mode(leaf, item, inode->i_mode); + btrfs_set_inode_nlink(leaf, item, inode->i_nlink); +@@ -7360,6 +7374,7 @@ static const struct inode_operations btr + .listxattr = btrfs_listxattr, + .removexattr = btrfs_removexattr, + .permission = btrfs_permission, ++ .sync_flags = btrfs_sync_flags, + }; + static const struct inode_operations btrfs_dir_ro_inode_operations = { + .lookup = btrfs_lookup, +@@ -7432,6 +7447,7 @@ static const struct inode_operations btr + .removexattr = btrfs_removexattr, + .permission = btrfs_permission, + .fiemap = btrfs_fiemap, ++ .sync_flags = btrfs_sync_flags, + }; + static const struct inode_operations btrfs_special_inode_operations = { + .getattr = btrfs_getattr, +diff -NurpP --minimal linux-3.0.9/fs/btrfs/ioctl.c linux-3.0.9-vs2.3.2.1/fs/btrfs/ioctl.c +--- linux-3.0.9/fs/btrfs/ioctl.c 2011-07-22 11:18:05.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/fs/btrfs/ioctl.c 2011-06-22 12:39:15.000000000 +0200 +@@ -70,10 +70,13 @@ static unsigned int btrfs_flags_to_ioctl + { + unsigned int iflags = 0; + +- if (flags & BTRFS_INODE_SYNC) +- iflags |= FS_SYNC_FL; + if (flags & BTRFS_INODE_IMMUTABLE) + iflags |= FS_IMMUTABLE_FL; ++ if (flags & BTRFS_INODE_IXUNLINK) ++ iflags |= FS_IXUNLINK_FL; ++ ++ if (flags & BTRFS_INODE_SYNC) ++ iflags |= FS_SYNC_FL; + if (flags & BTRFS_INODE_APPEND) + iflags |= FS_APPEND_FL; + if (flags & BTRFS_INODE_NODUMP) +@@ -90,28 +93,78 @@ static unsigned int btrfs_flags_to_ioctl + else if (flags & BTRFS_INODE_NOCOMPRESS) + iflags |= FS_NOCOMP_FL; + ++ if (flags & BTRFS_INODE_BARRIER) ++ iflags |= FS_BARRIER_FL; ++ if (flags & BTRFS_INODE_COW) ++ iflags |= FS_COW_FL; + return iflags; + } + + /* +- * Update inode->i_flags based on the btrfs internal flags. ++ * Update inode->i_(v)flags based on the btrfs internal flags. + */ + void btrfs_update_iflags(struct inode *inode) + { + struct btrfs_inode *ip = BTRFS_I(inode); + +- inode->i_flags &= ~(S_SYNC|S_APPEND|S_IMMUTABLE|S_NOATIME|S_DIRSYNC); ++ inode->i_flags &= ~(S_IMMUTABLE | S_IXUNLINK | ++ S_SYNC | S_APPEND | S_NOATIME | S_DIRSYNC); + +- if (ip->flags & BTRFS_INODE_SYNC) +- inode->i_flags |= S_SYNC; + if (ip->flags & BTRFS_INODE_IMMUTABLE) + inode->i_flags |= S_IMMUTABLE; ++ if (ip->flags & BTRFS_INODE_IXUNLINK) ++ inode->i_flags |= S_IXUNLINK; ++ ++ if (ip->flags & BTRFS_INODE_SYNC) ++ inode->i_flags |= S_SYNC; + if (ip->flags & BTRFS_INODE_APPEND) + inode->i_flags |= S_APPEND; + if (ip->flags & BTRFS_INODE_NOATIME) + inode->i_flags |= S_NOATIME; + if (ip->flags & BTRFS_INODE_DIRSYNC) + inode->i_flags |= S_DIRSYNC; ++ ++ inode->i_vflags &= ~(V_BARRIER | V_COW); ++ ++ if (ip->flags & BTRFS_INODE_BARRIER) ++ inode->i_vflags |= V_BARRIER; ++ if (ip->flags & BTRFS_INODE_COW) ++ inode->i_vflags |= V_COW; ++} ++ ++/* ++ * Update btrfs internal flags from inode->i_(v)flags. ++ */ ++void btrfs_update_flags(struct inode *inode) ++{ ++ struct btrfs_inode *ip = BTRFS_I(inode); ++ ++ unsigned int flags = inode->i_flags; ++ unsigned int vflags = inode->i_vflags; ++ ++ ip->flags &= ~(BTRFS_INODE_SYNC | BTRFS_INODE_APPEND | ++ BTRFS_INODE_IMMUTABLE | BTRFS_INODE_IXUNLINK | ++ BTRFS_INODE_NOATIME | BTRFS_INODE_DIRSYNC | ++ BTRFS_INODE_BARRIER | BTRFS_INODE_COW); ++ ++ if (flags & S_IMMUTABLE) ++ ip->flags |= BTRFS_INODE_IMMUTABLE; ++ if (flags & S_IXUNLINK) ++ ip->flags |= BTRFS_INODE_IXUNLINK; ++ ++ if (flags & S_SYNC) ++ ip->flags |= BTRFS_INODE_SYNC; ++ if (flags & S_APPEND) ++ ip->flags |= BTRFS_INODE_APPEND; ++ if (flags & S_NOATIME) ++ ip->flags |= BTRFS_INODE_NOATIME; ++ if (flags & S_DIRSYNC) ++ ip->flags |= BTRFS_INODE_DIRSYNC; ++ ++ if (vflags & V_BARRIER) ++ ip->flags |= BTRFS_INODE_BARRIER; ++ if (vflags & V_COW) ++ ip->flags |= BTRFS_INODE_COW; + } + + /* +@@ -129,7 +182,7 @@ void btrfs_inherit_iflags(struct inode * + flags = BTRFS_I(dir)->flags; + + if (S_ISREG(inode->i_mode)) +- flags &= ~BTRFS_INODE_DIRSYNC; ++ flags &= ~(BTRFS_INODE_DIRSYNC | BTRFS_INODE_BARRIER); + else if (!S_ISDIR(inode->i_mode)) + flags &= (BTRFS_INODE_NODUMP | BTRFS_INODE_NOATIME); + +@@ -137,6 +190,30 @@ void btrfs_inherit_iflags(struct inode * + btrfs_update_iflags(inode); + } + ++int btrfs_sync_flags(struct inode *inode, int flags, int vflags) ++{ ++ struct btrfs_inode *ip = BTRFS_I(inode); ++ struct btrfs_root *root = ip->root; ++ struct btrfs_trans_handle *trans; ++ int ret; ++ ++ trans = btrfs_join_transaction(root); ++ BUG_ON(!trans); ++ ++ inode->i_flags = flags; ++ inode->i_vflags = vflags; ++ btrfs_update_flags(inode); ++ ++ ret = btrfs_update_inode(trans, root, inode); ++ BUG_ON(ret); ++ ++ btrfs_update_iflags(inode); ++ inode->i_ctime = CURRENT_TIME; ++ btrfs_end_transaction(trans, root); ++ ++ return 0; ++} ++ + static int btrfs_ioctl_getflags(struct file *file, void __user *arg) + { + struct btrfs_inode *ip = BTRFS_I(file->f_path.dentry->d_inode); +@@ -188,7 +265,8 @@ static int btrfs_ioctl_setflags(struct f + + flags = btrfs_mask_flags(inode->i_mode, flags); + oldflags = btrfs_flags_to_ioctl(ip->flags); +- if ((flags ^ oldflags) & (FS_APPEND_FL | FS_IMMUTABLE_FL)) { ++ if ((flags ^ oldflags) & (FS_APPEND_FL | ++ FS_IMMUTABLE_FL | FS_IXUNLINK_FL)) { + if (!capable(CAP_LINUX_IMMUTABLE)) { + ret = -EPERM; + goto out_unlock; +@@ -199,14 +277,19 @@ static int btrfs_ioctl_setflags(struct f + if (ret) + goto out_unlock; + +- if (flags & FS_SYNC_FL) +- ip->flags |= BTRFS_INODE_SYNC; +- else +- ip->flags &= ~BTRFS_INODE_SYNC; + if (flags & FS_IMMUTABLE_FL) + ip->flags |= BTRFS_INODE_IMMUTABLE; + else + ip->flags &= ~BTRFS_INODE_IMMUTABLE; ++ if (flags & FS_IXUNLINK_FL) ++ ip->flags |= BTRFS_INODE_IXUNLINK; ++ else ++ ip->flags &= ~BTRFS_INODE_IXUNLINK; ++ ++ if (flags & FS_SYNC_FL) ++ ip->flags |= BTRFS_INODE_SYNC; ++ else ++ ip->flags &= ~BTRFS_INODE_SYNC; + if (flags & FS_APPEND_FL) + ip->flags |= BTRFS_INODE_APPEND; + else +diff -NurpP --minimal linux-3.0.9/fs/btrfs/super.c linux-3.0.9-vs2.3.2.1/fs/btrfs/super.c +--- linux-3.0.9/fs/btrfs/super.c 2011-07-22 11:18:05.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/fs/btrfs/super.c 2011-07-19 00:44:39.000000000 +0200 +@@ -162,7 +162,7 @@ enum { + Opt_notreelog, Opt_ratio, Opt_flushoncommit, Opt_discard, + Opt_space_cache, Opt_clear_cache, Opt_user_subvol_rm_allowed, + Opt_enospc_debug, Opt_subvolrootid, Opt_defrag, +- Opt_inode_cache, Opt_err, ++ Opt_inode_cache, Opt_tag, Opt_notag, Opt_tagid, Opt_err, + }; + + static match_table_t tokens = { +@@ -195,6 +195,9 @@ static match_table_t tokens = { + {Opt_subvolrootid, "subvolrootid=%d"}, + {Opt_defrag, "autodefrag"}, + {Opt_inode_cache, "inode_cache"}, ++ {Opt_tag, "tag"}, ++ {Opt_notag, "notag"}, ++ {Opt_tagid, "tagid=%u"}, + {Opt_err, NULL}, + }; + +@@ -381,6 +384,22 @@ int btrfs_parse_options(struct btrfs_roo + printk(KERN_INFO "btrfs: enabling auto defrag"); + btrfs_set_opt(info->mount_opt, AUTO_DEFRAG); + break; ++#ifndef CONFIG_TAGGING_NONE ++ case Opt_tag: ++ printk(KERN_INFO "btrfs: use tagging\n"); ++ btrfs_set_opt(info->mount_opt, TAGGED); ++ break; ++ case Opt_notag: ++ printk(KERN_INFO "btrfs: disabled tagging\n"); ++ btrfs_clear_opt(info->mount_opt, TAGGED); ++ break; ++#endif ++#ifdef CONFIG_PROPAGATE ++ case Opt_tagid: ++ /* use args[0] */ ++ btrfs_set_opt(info->mount_opt, TAGGED); ++ break; ++#endif + case Opt_err: + printk(KERN_INFO "btrfs: unrecognized mount option " + "'%s'\n", p); +@@ -907,6 +926,12 @@ static int btrfs_remount(struct super_bl + if (ret) + return -EINVAL; + ++ if (btrfs_test_opt(root, TAGGED) && !(sb->s_flags & MS_TAGGED)) { ++ printk("btrfs: %s: tagging not permitted on remount.\n", ++ sb->s_id); ++ return -EINVAL; ++ } ++ + if ((*flags & MS_RDONLY) == (sb->s_flags & MS_RDONLY)) + return 0; + +diff -NurpP --minimal linux-3.0.9/fs/char_dev.c linux-3.0.9-vs2.3.2.1/fs/char_dev.c +--- linux-3.0.9/fs/char_dev.c 2011-03-15 18:07:31.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/fs/char_dev.c 2011-06-10 22:11:24.000000000 +0200 +@@ -21,6 +21,8 @@ + #include <linux/mutex.h> + #include <linux/backing-dev.h> + #include <linux/tty.h> ++#include <linux/vs_context.h> ++#include <linux/vs_device.h> + + #include "internal.h" + +@@ -371,14 +373,21 @@ static int chrdev_open(struct inode *ino + struct cdev *p; + struct cdev *new = NULL; + int ret = 0; ++ dev_t mdev; ++ ++ if (!vs_map_chrdev(inode->i_rdev, &mdev, DATTR_OPEN)) ++ return -EPERM; ++ inode->i_mdev = mdev; + + spin_lock(&cdev_lock); + p = inode->i_cdev; + if (!p) { + struct kobject *kobj; + int idx; ++ + spin_unlock(&cdev_lock); +- kobj = kobj_lookup(cdev_map, inode->i_rdev, &idx); ++ ++ kobj = kobj_lookup(cdev_map, mdev, &idx); + if (!kobj) + return -ENXIO; + new = container_of(kobj, struct cdev, kobj); +diff -NurpP --minimal linux-3.0.9/fs/dcache.c linux-3.0.9-vs2.3.2.1/fs/dcache.c +--- linux-3.0.9/fs/dcache.c 2011-07-22 11:18:05.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/fs/dcache.c 2011-07-22 11:20:39.000000000 +0200 +@@ -36,6 +36,7 @@ + #include <linux/bit_spinlock.h> + #include <linux/rculist_bl.h> + #include <linux/prefetch.h> ++#include <linux/vs_limit.h> + #include "internal.h" + + /* +@@ -479,6 +480,8 @@ int d_invalidate(struct dentry * dentry) + spin_lock(&dentry->d_lock); + } + ++ vx_dentry_dec(dentry); ++ + /* + * Somebody else still using it? + * +@@ -506,6 +509,7 @@ EXPORT_SYMBOL(d_invalidate); + static inline void __dget_dlock(struct dentry *dentry) + { + dentry->d_count++; ++ vx_dentry_inc(dentry); + } + + static inline void __dget(struct dentry *dentry) +@@ -1266,6 +1270,9 @@ struct dentry *d_alloc(struct dentry * p + struct dentry *dentry; + char *dname; + ++ if (!vx_dentry_avail(1)) ++ return NULL; ++ + dentry = kmem_cache_alloc(dentry_cache, GFP_KERNEL); + if (!dentry) + return NULL; +@@ -1288,6 +1295,7 @@ struct dentry *d_alloc(struct dentry * p + + dentry->d_count = 1; + dentry->d_flags = 0; ++ vx_dentry_inc(dentry); + spin_lock_init(&dentry->d_lock); + seqcount_init(&dentry->d_seq); + dentry->d_inode = NULL; +@@ -1945,6 +1953,7 @@ struct dentry *__d_lookup(struct dentry + } + + dentry->d_count++; ++ vx_dentry_inc(dentry); + found = dentry; + spin_unlock(&dentry->d_lock); + break; +diff -NurpP --minimal linux-3.0.9/fs/devpts/inode.c linux-3.0.9-vs2.3.2.1/fs/devpts/inode.c +--- linux-3.0.9/fs/devpts/inode.c 2011-05-22 16:17:50.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/fs/devpts/inode.c 2011-06-10 22:11:24.000000000 +0200 +@@ -25,6 +25,7 @@ + #include <linux/parser.h> + #include <linux/fsnotify.h> + #include <linux/seq_file.h> ++#include <linux/vs_base.h> + + #define DEVPTS_DEFAULT_MODE 0600 + /* +@@ -36,6 +37,20 @@ + #define DEVPTS_DEFAULT_PTMX_MODE 0000 + #define PTMX_MINOR 2 + ++static int devpts_permission(struct inode *inode, int mask, unsigned int flags) ++{ ++ int ret = -EACCES; ++ ++ /* devpts is xid tagged */ ++ if (vx_check((xid_t)inode->i_tag, VS_WATCH_P | VS_IDENT)) ++ ret = generic_permission(inode, mask, flags, NULL); ++ return ret; ++} ++ ++static struct inode_operations devpts_file_inode_operations = { ++ .permission = devpts_permission, ++}; ++ + extern int pty_limit; /* Config limit on Unix98 ptys */ + static DEFINE_MUTEX(allocated_ptys_lock); + +@@ -263,6 +278,34 @@ static int devpts_show_options(struct se + return 0; + } + ++static int devpts_filter(struct dentry *de) ++{ ++ xid_t xid = 0; ++ ++ /* devpts is xid tagged */ ++ if (de && de->d_inode) ++ xid = (xid_t)de->d_inode->i_tag; ++#ifdef CONFIG_VSERVER_WARN_DEVPTS ++ else ++ vxwprintk_task(1, "devpts " VS_Q("%.*s") " without inode.", ++ de->d_name.len, de->d_name.name); ++#endif ++ return vx_check(xid, VS_WATCH_P | VS_IDENT); ++} ++ ++static int devpts_readdir(struct file * filp, void * dirent, filldir_t filldir) ++{ ++ return dcache_readdir_filter(filp, dirent, filldir, devpts_filter); ++} ++ ++static struct file_operations devpts_dir_operations = { ++ .open = dcache_dir_open, ++ .release = dcache_dir_close, ++ .llseek = dcache_dir_lseek, ++ .read = generic_read_dir, ++ .readdir = devpts_readdir, ++}; ++ + static const struct super_operations devpts_sops = { + .statfs = simple_statfs, + .remount_fs = devpts_remount, +@@ -302,12 +345,15 @@ devpts_fill_super(struct super_block *s, + inode = new_inode(s); + if (!inode) + goto free_fsi; ++ + inode->i_ino = 1; + inode->i_mtime = inode->i_atime = inode->i_ctime = CURRENT_TIME; + inode->i_mode = S_IFDIR | S_IRUGO | S_IXUGO | S_IWUSR; + inode->i_op = &simple_dir_inode_operations; +- inode->i_fop = &simple_dir_operations; ++ inode->i_fop = &devpts_dir_operations; + inode->i_nlink = 2; ++ /* devpts is xid tagged */ ++ inode->i_tag = (tag_t)vx_current_xid(); + + s->s_root = d_alloc_root(inode); + if (s->s_root) +@@ -494,6 +540,9 @@ int devpts_pty_new(struct inode *ptmx_in + inode->i_gid = opts->setgid ? opts->gid : current_fsgid(); + inode->i_mtime = inode->i_atime = inode->i_ctime = CURRENT_TIME; + init_special_inode(inode, S_IFCHR|opts->mode, device); ++ /* devpts is xid tagged */ ++ inode->i_tag = (tag_t)vx_current_xid(); ++ inode->i_op = &devpts_file_inode_operations; + inode->i_private = tty; + tty->driver_data = inode; + +diff -NurpP --minimal linux-3.0.9/fs/ext2/balloc.c linux-3.0.9-vs2.3.2.1/fs/ext2/balloc.c +--- linux-3.0.9/fs/ext2/balloc.c 2011-05-22 16:17:51.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/fs/ext2/balloc.c 2011-06-10 22:11:24.000000000 +0200 +@@ -701,7 +701,6 @@ ext2_try_to_allocate(struct super_block + start = 0; + end = EXT2_BLOCKS_PER_GROUP(sb); + } +- + BUG_ON(start > EXT2_BLOCKS_PER_GROUP(sb)); + + repeat: +diff -NurpP --minimal linux-3.0.9/fs/ext2/ext2.h linux-3.0.9-vs2.3.2.1/fs/ext2/ext2.h +--- linux-3.0.9/fs/ext2/ext2.h 2011-05-22 16:17:51.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/fs/ext2/ext2.h 2011-06-10 22:11:24.000000000 +0200 +@@ -126,6 +126,7 @@ extern void ext2_set_inode_flags(struct + extern void ext2_get_inode_flags(struct ext2_inode_info *); + extern int ext2_fiemap(struct inode *inode, struct fiemap_extent_info *fieinfo, + u64 start, u64 len); ++extern int ext2_sync_flags(struct inode *, int, int); + + /* ioctl.c */ + extern long ext2_ioctl(struct file *, unsigned int, unsigned long); +diff -NurpP --minimal linux-3.0.9/fs/ext2/file.c linux-3.0.9-vs2.3.2.1/fs/ext2/file.c +--- linux-3.0.9/fs/ext2/file.c 2010-08-02 16:52:48.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/fs/ext2/file.c 2011-06-10 22:11:24.000000000 +0200 +@@ -104,4 +104,5 @@ const struct inode_operations ext2_file_ + .setattr = ext2_setattr, + .check_acl = ext2_check_acl, + .fiemap = ext2_fiemap, ++ .sync_flags = ext2_sync_flags, + }; +diff -NurpP --minimal linux-3.0.9/fs/ext2/ialloc.c linux-3.0.9-vs2.3.2.1/fs/ext2/ialloc.c +--- linux-3.0.9/fs/ext2/ialloc.c 2011-05-22 16:17:51.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/fs/ext2/ialloc.c 2011-06-10 22:11:24.000000000 +0200 +@@ -17,6 +17,7 @@ + #include <linux/backing-dev.h> + #include <linux/buffer_head.h> + #include <linux/random.h> ++#include <linux/vs_tag.h> + #include "ext2.h" + #include "xattr.h" + #include "acl.h" +@@ -549,6 +550,7 @@ got: + inode->i_mode = mode; + inode->i_uid = current_fsuid(); + inode->i_gid = dir->i_gid; ++ inode->i_tag = dx_current_fstag(sb); + } else + inode_init_owner(inode, dir, mode); + +diff -NurpP --minimal linux-3.0.9/fs/ext2/inode.c linux-3.0.9-vs2.3.2.1/fs/ext2/inode.c +--- linux-3.0.9/fs/ext2/inode.c 2011-05-22 16:17:51.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/fs/ext2/inode.c 2011-06-10 22:11:24.000000000 +0200 +@@ -32,6 +32,7 @@ + #include <linux/mpage.h> + #include <linux/fiemap.h> + #include <linux/namei.h> ++#include <linux/vs_tag.h> + #include "ext2.h" + #include "acl.h" + #include "xip.h" +@@ -1167,7 +1168,7 @@ static void ext2_truncate_blocks(struct + return; + if (ext2_inode_is_fast_symlink(inode)) + return; +- if (IS_APPEND(inode) || IS_IMMUTABLE(inode)) ++ if (IS_APPEND(inode) || IS_IXORUNLINK(inode)) + return; + __ext2_truncate_blocks(inode, offset); + } +@@ -1256,36 +1257,61 @@ void ext2_set_inode_flags(struct inode * + { + unsigned int flags = EXT2_I(inode)->i_flags; + +- inode->i_flags &= ~(S_SYNC|S_APPEND|S_IMMUTABLE|S_NOATIME|S_DIRSYNC); ++ inode->i_flags &= ~(S_IMMUTABLE | S_IXUNLINK | ++ S_SYNC | S_APPEND | S_NOATIME | S_DIRSYNC); ++ ++ ++ if (flags & EXT2_IMMUTABLE_FL) ++ inode->i_flags |= S_IMMUTABLE; ++ if (flags & EXT2_IXUNLINK_FL) ++ inode->i_flags |= S_IXUNLINK; ++ + if (flags & EXT2_SYNC_FL) + inode->i_flags |= S_SYNC; + if (flags & EXT2_APPEND_FL) + inode->i_flags |= S_APPEND; +- if (flags & EXT2_IMMUTABLE_FL) +- inode->i_flags |= S_IMMUTABLE; + if (flags & EXT2_NOATIME_FL) + inode->i_flags |= S_NOATIME; + if (flags & EXT2_DIRSYNC_FL) + inode->i_flags |= S_DIRSYNC; ++ ++ inode->i_vflags &= ~(V_BARRIER | V_COW); ++ ++ if (flags & EXT2_BARRIER_FL) ++ inode->i_vflags |= V_BARRIER; ++ if (flags & EXT2_COW_FL) ++ inode->i_vflags |= V_COW; + } + + /* Propagate flags from i_flags to EXT2_I(inode)->i_flags */ + void ext2_get_inode_flags(struct ext2_inode_info *ei) + { + unsigned int flags = ei->vfs_inode.i_flags; ++ unsigned int vflags = ei->vfs_inode.i_vflags; ++ ++ ei->i_flags &= ~(EXT2_SYNC_FL | EXT2_APPEND_FL | ++ EXT2_IMMUTABLE_FL | EXT2_IXUNLINK_FL | ++ EXT2_NOATIME_FL | EXT2_DIRSYNC_FL | ++ EXT2_BARRIER_FL | EXT2_COW_FL); ++ ++ if (flags & S_IMMUTABLE) ++ ei->i_flags |= EXT2_IMMUTABLE_FL; ++ if (flags & S_IXUNLINK) ++ ei->i_flags |= EXT2_IXUNLINK_FL; + +- ei->i_flags &= ~(EXT2_SYNC_FL|EXT2_APPEND_FL| +- EXT2_IMMUTABLE_FL|EXT2_NOATIME_FL|EXT2_DIRSYNC_FL); + if (flags & S_SYNC) + ei->i_flags |= EXT2_SYNC_FL; + if (flags & S_APPEND) + ei->i_flags |= EXT2_APPEND_FL; +- if (flags & S_IMMUTABLE) +- ei->i_flags |= EXT2_IMMUTABLE_FL; + if (flags & S_NOATIME) + ei->i_flags |= EXT2_NOATIME_FL; + if (flags & S_DIRSYNC) + ei->i_flags |= EXT2_DIRSYNC_FL; ++ ++ if (vflags & V_BARRIER) ++ ei->i_flags |= EXT2_BARRIER_FL; ++ if (vflags & V_COW) ++ ei->i_flags |= EXT2_COW_FL; + } + + struct inode *ext2_iget (struct super_block *sb, unsigned long ino) +@@ -1295,6 +1321,8 @@ struct inode *ext2_iget (struct super_bl + struct ext2_inode *raw_inode; + struct inode *inode; + long ret = -EIO; ++ uid_t uid; ++ gid_t gid; + int n; + + inode = iget_locked(sb, ino); +@@ -1313,12 +1341,17 @@ struct inode *ext2_iget (struct super_bl + } + + inode->i_mode = le16_to_cpu(raw_inode->i_mode); +- inode->i_uid = (uid_t)le16_to_cpu(raw_inode->i_uid_low); +- inode->i_gid = (gid_t)le16_to_cpu(raw_inode->i_gid_low); ++ uid = (uid_t)le16_to_cpu(raw_inode->i_uid_low); ++ gid = (gid_t)le16_to_cpu(raw_inode->i_gid_low); + if (!(test_opt (inode->i_sb, NO_UID32))) { +- inode->i_uid |= le16_to_cpu(raw_inode->i_uid_high) << 16; +- inode->i_gid |= le16_to_cpu(raw_inode->i_gid_high) << 16; ++ uid |= le16_to_cpu(raw_inode->i_uid_high) << 16; ++ gid |= le16_to_cpu(raw_inode->i_gid_high) << 16; + } ++ inode->i_uid = INOTAG_UID(DX_TAG(inode), uid, gid); ++ inode->i_gid = INOTAG_GID(DX_TAG(inode), uid, gid); ++ inode->i_tag = INOTAG_TAG(DX_TAG(inode), uid, gid, ++ le16_to_cpu(raw_inode->i_raw_tag)); ++ + inode->i_nlink = le16_to_cpu(raw_inode->i_links_count); + inode->i_size = le32_to_cpu(raw_inode->i_size); + inode->i_atime.tv_sec = (signed)le32_to_cpu(raw_inode->i_atime); +@@ -1416,8 +1449,8 @@ static int __ext2_write_inode(struct ino + struct ext2_inode_info *ei = EXT2_I(inode); + struct super_block *sb = inode->i_sb; + ino_t ino = inode->i_ino; +- uid_t uid = inode->i_uid; +- gid_t gid = inode->i_gid; ++ uid_t uid = TAGINO_UID(DX_TAG(inode), inode->i_uid, inode->i_tag); ++ gid_t gid = TAGINO_GID(DX_TAG(inode), inode->i_gid, inode->i_tag); + struct buffer_head * bh; + struct ext2_inode * raw_inode = ext2_get_inode(sb, ino, &bh); + int n; +@@ -1453,6 +1486,9 @@ static int __ext2_write_inode(struct ino + raw_inode->i_uid_high = 0; + raw_inode->i_gid_high = 0; + } ++#ifdef CONFIG_TAGGING_INTERN ++ raw_inode->i_raw_tag = cpu_to_le16(inode->i_tag); ++#endif + raw_inode->i_links_count = cpu_to_le16(inode->i_nlink); + raw_inode->i_size = cpu_to_le32(inode->i_size); + raw_inode->i_atime = cpu_to_le32(inode->i_atime.tv_sec); +@@ -1533,7 +1569,8 @@ int ext2_setattr(struct dentry *dentry, + if (is_quota_modification(inode, iattr)) + dquot_initialize(inode); + if ((iattr->ia_valid & ATTR_UID && iattr->ia_uid != inode->i_uid) || +- (iattr->ia_valid & ATTR_GID && iattr->ia_gid != inode->i_gid)) { ++ (iattr->ia_valid & ATTR_GID && iattr->ia_gid != inode->i_gid) || ++ (iattr->ia_valid & ATTR_TAG && iattr->ia_tag != inode->i_tag)) { + error = dquot_transfer(inode, iattr); + if (error) + return error; +diff -NurpP --minimal linux-3.0.9/fs/ext2/ioctl.c linux-3.0.9-vs2.3.2.1/fs/ext2/ioctl.c +--- linux-3.0.9/fs/ext2/ioctl.c 2011-05-22 16:17:51.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/fs/ext2/ioctl.c 2011-06-10 22:11:24.000000000 +0200 +@@ -17,6 +17,16 @@ + #include <asm/uaccess.h> + + ++int ext2_sync_flags(struct inode *inode, int flags, int vflags) ++{ ++ inode->i_flags = flags; ++ inode->i_vflags = vflags; ++ ext2_get_inode_flags(EXT2_I(inode)); ++ inode->i_ctime = CURRENT_TIME_SEC; ++ mark_inode_dirty(inode); ++ return 0; ++} ++ + long ext2_ioctl(struct file *filp, unsigned int cmd, unsigned long arg) + { + struct inode *inode = filp->f_dentry->d_inode; +@@ -51,6 +61,11 @@ long ext2_ioctl(struct file *filp, unsig + + flags = ext2_mask_flags(inode->i_mode, flags); + ++ if (IS_BARRIER(inode)) { ++ vxwprintk_task(1, "messing with the barrier."); ++ return -EACCES; ++ } ++ + mutex_lock(&inode->i_mutex); + /* Is it quota file? Do not allow user to mess with it */ + if (IS_NOQUOTA(inode)) { +@@ -66,7 +81,9 @@ long ext2_ioctl(struct file *filp, unsig + * + * This test looks nicer. Thanks to Pauline Middelink + */ +- if ((flags ^ oldflags) & (EXT2_APPEND_FL | EXT2_IMMUTABLE_FL)) { ++ if ((oldflags & EXT2_IMMUTABLE_FL) || ++ ((flags ^ oldflags) & (EXT2_APPEND_FL | ++ EXT2_IMMUTABLE_FL | EXT2_IXUNLINK_FL))) { + if (!capable(CAP_LINUX_IMMUTABLE)) { + mutex_unlock(&inode->i_mutex); + ret = -EPERM; +@@ -74,7 +91,7 @@ long ext2_ioctl(struct file *filp, unsig + } + } + +- flags = flags & EXT2_FL_USER_MODIFIABLE; ++ flags &= EXT2_FL_USER_MODIFIABLE; + flags |= oldflags & ~EXT2_FL_USER_MODIFIABLE; + ei->i_flags = flags; + mutex_unlock(&inode->i_mutex); +diff -NurpP --minimal linux-3.0.9/fs/ext2/namei.c linux-3.0.9-vs2.3.2.1/fs/ext2/namei.c +--- linux-3.0.9/fs/ext2/namei.c 2011-05-22 16:17:51.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/fs/ext2/namei.c 2011-06-10 22:11:24.000000000 +0200 +@@ -32,6 +32,7 @@ + + #include <linux/pagemap.h> + #include <linux/quotaops.h> ++#include <linux/vs_tag.h> + #include "ext2.h" + #include "xattr.h" + #include "acl.h" +@@ -75,6 +76,7 @@ static struct dentry *ext2_lookup(struct + return ERR_PTR(-EIO); + } else { + return ERR_CAST(inode); ++ dx_propagate_tag(nd, inode); + } + } + } +@@ -413,6 +415,7 @@ const struct inode_operations ext2_dir_i + #endif + .setattr = ext2_setattr, + .check_acl = ext2_check_acl, ++ .sync_flags = ext2_sync_flags, + }; + + const struct inode_operations ext2_special_inode_operations = { +diff -NurpP --minimal linux-3.0.9/fs/ext2/super.c linux-3.0.9-vs2.3.2.1/fs/ext2/super.c +--- linux-3.0.9/fs/ext2/super.c 2011-07-22 11:18:05.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/fs/ext2/super.c 2011-06-10 22:11:24.000000000 +0200 +@@ -394,7 +394,8 @@ enum { + Opt_err_ro, Opt_nouid32, Opt_nocheck, Opt_debug, + Opt_oldalloc, Opt_orlov, Opt_nobh, Opt_user_xattr, Opt_nouser_xattr, + Opt_acl, Opt_noacl, Opt_xip, Opt_ignore, Opt_err, Opt_quota, +- Opt_usrquota, Opt_grpquota, Opt_reservation, Opt_noreservation ++ Opt_usrquota, Opt_grpquota, Opt_reservation, Opt_noreservation, ++ Opt_tag, Opt_notag, Opt_tagid + }; + + static const match_table_t tokens = { +@@ -422,6 +423,9 @@ static const match_table_t tokens = { + {Opt_acl, "acl"}, + {Opt_noacl, "noacl"}, + {Opt_xip, "xip"}, ++ {Opt_tag, "tag"}, ++ {Opt_notag, "notag"}, ++ {Opt_tagid, "tagid=%u"}, + {Opt_grpquota, "grpquota"}, + {Opt_ignore, "noquota"}, + {Opt_quota, "quota"}, +@@ -492,6 +496,20 @@ static int parse_options(char *options, + case Opt_nouid32: + set_opt (sbi->s_mount_opt, NO_UID32); + break; ++#ifndef CONFIG_TAGGING_NONE ++ case Opt_tag: ++ set_opt (sbi->s_mount_opt, TAGGED); ++ break; ++ case Opt_notag: ++ clear_opt (sbi->s_mount_opt, TAGGED); ++ break; ++#endif ++#ifdef CONFIG_PROPAGATE ++ case Opt_tagid: ++ /* use args[0] */ ++ set_opt (sbi->s_mount_opt, TAGGED); ++ break; ++#endif + case Opt_nocheck: + clear_opt (sbi->s_mount_opt, CHECK); + break; +@@ -850,6 +868,8 @@ static int ext2_fill_super(struct super_ + if (!parse_options((char *) data, sb)) + goto failed_mount; + ++ if (EXT2_SB(sb)->s_mount_opt & EXT2_MOUNT_TAGGED) ++ sb->s_flags |= MS_TAGGED; + sb->s_flags = (sb->s_flags & ~MS_POSIXACL) | + ((EXT2_SB(sb)->s_mount_opt & EXT2_MOUNT_POSIX_ACL) ? + MS_POSIXACL : 0); +@@ -1224,6 +1244,14 @@ static int ext2_remount (struct super_bl + goto restore_opts; + } + ++ if ((sbi->s_mount_opt & EXT2_MOUNT_TAGGED) && ++ !(sb->s_flags & MS_TAGGED)) { ++ printk("EXT2-fs: %s: tagging not permitted on remount.\n", ++ sb->s_id); ++ err = -EINVAL; ++ goto restore_opts; ++ } ++ + sb->s_flags = (sb->s_flags & ~MS_POSIXACL) | + ((sbi->s_mount_opt & EXT2_MOUNT_POSIX_ACL) ? MS_POSIXACL : 0); + +diff -NurpP --minimal linux-3.0.9/fs/ext3/file.c linux-3.0.9-vs2.3.2.1/fs/ext3/file.c +--- linux-3.0.9/fs/ext3/file.c 2010-07-07 18:31:51.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/fs/ext3/file.c 2011-06-10 22:11:24.000000000 +0200 +@@ -81,5 +81,6 @@ const struct inode_operations ext3_file_ + #endif + .check_acl = ext3_check_acl, + .fiemap = ext3_fiemap, ++ .sync_flags = ext3_sync_flags, + }; + +diff -NurpP --minimal linux-3.0.9/fs/ext3/ialloc.c linux-3.0.9-vs2.3.2.1/fs/ext3/ialloc.c +--- linux-3.0.9/fs/ext3/ialloc.c 2011-05-22 16:17:52.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/fs/ext3/ialloc.c 2011-06-10 22:11:24.000000000 +0200 +@@ -23,6 +23,7 @@ + #include <linux/buffer_head.h> + #include <linux/random.h> + #include <linux/bitops.h> ++#include <linux/vs_tag.h> + + #include <asm/byteorder.h> + +@@ -532,6 +533,7 @@ got: + inode->i_mode = mode; + inode->i_uid = current_fsuid(); + inode->i_gid = dir->i_gid; ++ inode->i_tag = dx_current_fstag(sb); + } else + inode_init_owner(inode, dir, mode); + +diff -NurpP --minimal linux-3.0.9/fs/ext3/inode.c linux-3.0.9-vs2.3.2.1/fs/ext3/inode.c +--- linux-3.0.9/fs/ext3/inode.c 2011-07-22 11:18:05.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/fs/ext3/inode.c 2011-06-10 22:11:24.000000000 +0200 +@@ -38,6 +38,7 @@ + #include <linux/bio.h> + #include <linux/fiemap.h> + #include <linux/namei.h> ++#include <linux/vs_tag.h> + #include "xattr.h" + #include "acl.h" + +@@ -2391,7 +2392,7 @@ static void ext3_free_branches(handle_t + + int ext3_can_truncate(struct inode *inode) + { +- if (IS_APPEND(inode) || IS_IMMUTABLE(inode)) ++ if (IS_APPEND(inode) || IS_IXORUNLINK(inode)) + return 0; + if (S_ISREG(inode->i_mode)) + return 1; +@@ -2775,36 +2776,60 @@ void ext3_set_inode_flags(struct inode * + { + unsigned int flags = EXT3_I(inode)->i_flags; + +- inode->i_flags &= ~(S_SYNC|S_APPEND|S_IMMUTABLE|S_NOATIME|S_DIRSYNC); ++ inode->i_flags &= ~(S_IMMUTABLE | S_IXUNLINK | ++ S_SYNC | S_APPEND | S_NOATIME | S_DIRSYNC); ++ ++ if (flags & EXT3_IMMUTABLE_FL) ++ inode->i_flags |= S_IMMUTABLE; ++ if (flags & EXT3_IXUNLINK_FL) ++ inode->i_flags |= S_IXUNLINK; ++ + if (flags & EXT3_SYNC_FL) + inode->i_flags |= S_SYNC; + if (flags & EXT3_APPEND_FL) + inode->i_flags |= S_APPEND; +- if (flags & EXT3_IMMUTABLE_FL) +- inode->i_flags |= S_IMMUTABLE; + if (flags & EXT3_NOATIME_FL) + inode->i_flags |= S_NOATIME; + if (flags & EXT3_DIRSYNC_FL) + inode->i_flags |= S_DIRSYNC; ++ ++ inode->i_vflags &= ~(V_BARRIER | V_COW); ++ ++ if (flags & EXT3_BARRIER_FL) ++ inode->i_vflags |= V_BARRIER; ++ if (flags & EXT3_COW_FL) ++ inode->i_vflags |= V_COW; + } + + /* Propagate flags from i_flags to EXT3_I(inode)->i_flags */ + void ext3_get_inode_flags(struct ext3_inode_info *ei) + { + unsigned int flags = ei->vfs_inode.i_flags; ++ unsigned int vflags = ei->vfs_inode.i_vflags; ++ ++ ei->i_flags &= ~(EXT3_SYNC_FL | EXT3_APPEND_FL | ++ EXT3_IMMUTABLE_FL | EXT3_IXUNLINK_FL | ++ EXT3_NOATIME_FL | EXT3_DIRSYNC_FL | ++ EXT3_BARRIER_FL | EXT3_COW_FL); ++ ++ if (flags & S_IMMUTABLE) ++ ei->i_flags |= EXT3_IMMUTABLE_FL; ++ if (flags & S_IXUNLINK) ++ ei->i_flags |= EXT3_IXUNLINK_FL; + +- ei->i_flags &= ~(EXT3_SYNC_FL|EXT3_APPEND_FL| +- EXT3_IMMUTABLE_FL|EXT3_NOATIME_FL|EXT3_DIRSYNC_FL); + if (flags & S_SYNC) + ei->i_flags |= EXT3_SYNC_FL; + if (flags & S_APPEND) + ei->i_flags |= EXT3_APPEND_FL; +- if (flags & S_IMMUTABLE) +- ei->i_flags |= EXT3_IMMUTABLE_FL; + if (flags & S_NOATIME) + ei->i_flags |= EXT3_NOATIME_FL; + if (flags & S_DIRSYNC) + ei->i_flags |= EXT3_DIRSYNC_FL; ++ ++ if (vflags & V_BARRIER) ++ ei->i_flags |= EXT3_BARRIER_FL; ++ if (vflags & V_COW) ++ ei->i_flags |= EXT3_COW_FL; + } + + struct inode *ext3_iget(struct super_block *sb, unsigned long ino) +@@ -2818,6 +2843,8 @@ struct inode *ext3_iget(struct super_blo + transaction_t *transaction; + long ret; + int block; ++ uid_t uid; ++ gid_t gid; + + inode = iget_locked(sb, ino); + if (!inode) +@@ -2834,12 +2861,17 @@ struct inode *ext3_iget(struct super_blo + bh = iloc.bh; + raw_inode = ext3_raw_inode(&iloc); + inode->i_mode = le16_to_cpu(raw_inode->i_mode); +- inode->i_uid = (uid_t)le16_to_cpu(raw_inode->i_uid_low); +- inode->i_gid = (gid_t)le16_to_cpu(raw_inode->i_gid_low); ++ uid = (uid_t)le16_to_cpu(raw_inode->i_uid_low); ++ gid = (gid_t)le16_to_cpu(raw_inode->i_gid_low); + if(!(test_opt (inode->i_sb, NO_UID32))) { +- inode->i_uid |= le16_to_cpu(raw_inode->i_uid_high) << 16; +- inode->i_gid |= le16_to_cpu(raw_inode->i_gid_high) << 16; ++ uid |= le16_to_cpu(raw_inode->i_uid_high) << 16; ++ gid |= le16_to_cpu(raw_inode->i_gid_high) << 16; + } ++ inode->i_uid = INOTAG_UID(DX_TAG(inode), uid, gid); ++ inode->i_gid = INOTAG_GID(DX_TAG(inode), uid, gid); ++ inode->i_tag = INOTAG_TAG(DX_TAG(inode), uid, gid, ++ le16_to_cpu(raw_inode->i_raw_tag)); ++ + inode->i_nlink = le16_to_cpu(raw_inode->i_links_count); + inode->i_size = le32_to_cpu(raw_inode->i_size); + inode->i_atime.tv_sec = (signed)le32_to_cpu(raw_inode->i_atime); +@@ -2994,6 +3026,8 @@ static int ext3_do_update_inode(handle_t + struct ext3_inode *raw_inode = ext3_raw_inode(iloc); + struct ext3_inode_info *ei = EXT3_I(inode); + struct buffer_head *bh = iloc->bh; ++ uid_t uid = TAGINO_UID(DX_TAG(inode), inode->i_uid, inode->i_tag); ++ gid_t gid = TAGINO_GID(DX_TAG(inode), inode->i_gid, inode->i_tag); + int err = 0, rc, block; + + again: +@@ -3008,29 +3042,32 @@ again: + ext3_get_inode_flags(ei); + raw_inode->i_mode = cpu_to_le16(inode->i_mode); + if(!(test_opt(inode->i_sb, NO_UID32))) { +- raw_inode->i_uid_low = cpu_to_le16(low_16_bits(inode->i_uid)); +- raw_inode->i_gid_low = cpu_to_le16(low_16_bits(inode->i_gid)); ++ raw_inode->i_uid_low = cpu_to_le16(low_16_bits(uid)); ++ raw_inode->i_gid_low = cpu_to_le16(low_16_bits(gid)); + /* + * Fix up interoperability with old kernels. Otherwise, old inodes get + * re-used with the upper 16 bits of the uid/gid intact + */ + if(!ei->i_dtime) { + raw_inode->i_uid_high = +- cpu_to_le16(high_16_bits(inode->i_uid)); ++ cpu_to_le16(high_16_bits(uid)); + raw_inode->i_gid_high = +- cpu_to_le16(high_16_bits(inode->i_gid)); ++ cpu_to_le16(high_16_bits(gid)); + } else { + raw_inode->i_uid_high = 0; + raw_inode->i_gid_high = 0; + } + } else { + raw_inode->i_uid_low = +- cpu_to_le16(fs_high2lowuid(inode->i_uid)); ++ cpu_to_le16(fs_high2lowuid(uid)); + raw_inode->i_gid_low = +- cpu_to_le16(fs_high2lowgid(inode->i_gid)); ++ cpu_to_le16(fs_high2lowgid(gid)); + raw_inode->i_uid_high = 0; + raw_inode->i_gid_high = 0; + } ++#ifdef CONFIG_TAGGING_INTERN ++ raw_inode->i_raw_tag = cpu_to_le16(inode->i_tag); ++#endif + raw_inode->i_links_count = cpu_to_le16(inode->i_nlink); + raw_inode->i_size = cpu_to_le32(ei->i_disksize); + raw_inode->i_atime = cpu_to_le32(inode->i_atime.tv_sec); +@@ -3190,7 +3227,8 @@ int ext3_setattr(struct dentry *dentry, + if (is_quota_modification(inode, attr)) + dquot_initialize(inode); + if ((ia_valid & ATTR_UID && attr->ia_uid != inode->i_uid) || +- (ia_valid & ATTR_GID && attr->ia_gid != inode->i_gid)) { ++ (ia_valid & ATTR_GID && attr->ia_gid != inode->i_gid) || ++ (ia_valid & ATTR_TAG && attr->ia_tag != inode->i_tag)) { + handle_t *handle; + + /* (user+group)*(old+new) structure, inode write (sb, +@@ -3212,6 +3250,8 @@ int ext3_setattr(struct dentry *dentry, + inode->i_uid = attr->ia_uid; + if (attr->ia_valid & ATTR_GID) + inode->i_gid = attr->ia_gid; ++ if ((attr->ia_valid & ATTR_TAG) && IS_TAGGED(inode)) ++ inode->i_tag = attr->ia_tag; + error = ext3_mark_inode_dirty(handle, inode); + ext3_journal_stop(handle); + } +diff -NurpP --minimal linux-3.0.9/fs/ext3/ioctl.c linux-3.0.9-vs2.3.2.1/fs/ext3/ioctl.c +--- linux-3.0.9/fs/ext3/ioctl.c 2011-05-22 16:17:52.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/fs/ext3/ioctl.c 2011-06-10 22:11:24.000000000 +0200 +@@ -8,6 +8,7 @@ + */ + + #include <linux/fs.h> ++#include <linux/mount.h> + #include <linux/jbd.h> + #include <linux/capability.h> + #include <linux/ext3_fs.h> +@@ -17,6 +18,34 @@ + #include <linux/compat.h> + #include <asm/uaccess.h> + ++ ++int ext3_sync_flags(struct inode *inode, int flags, int vflags) ++{ ++ handle_t *handle = NULL; ++ struct ext3_iloc iloc; ++ int err; ++ ++ handle = ext3_journal_start(inode, 1); ++ if (IS_ERR(handle)) ++ return PTR_ERR(handle); ++ ++ if (IS_SYNC(inode)) ++ handle->h_sync = 1; ++ err = ext3_reserve_inode_write(handle, inode, &iloc); ++ if (err) ++ goto flags_err; ++ ++ inode->i_flags = flags; ++ inode->i_vflags = vflags; ++ ext3_get_inode_flags(EXT3_I(inode)); ++ inode->i_ctime = CURRENT_TIME_SEC; ++ ++ err = ext3_mark_iloc_dirty(handle, inode, &iloc); ++flags_err: ++ ext3_journal_stop(handle); ++ return err; ++} ++ + long ext3_ioctl(struct file *filp, unsigned int cmd, unsigned long arg) + { + struct inode *inode = filp->f_dentry->d_inode; +@@ -50,6 +79,11 @@ long ext3_ioctl(struct file *filp, unsig + + flags = ext3_mask_flags(inode->i_mode, flags); + ++ if (IS_BARRIER(inode)) { ++ vxwprintk_task(1, "messing with the barrier."); ++ return -EACCES; ++ } ++ + mutex_lock(&inode->i_mutex); + + /* Is it quota file? Do not allow user to mess with it */ +@@ -68,7 +102,9 @@ long ext3_ioctl(struct file *filp, unsig + * + * This test looks nicer. Thanks to Pauline Middelink + */ +- if ((flags ^ oldflags) & (EXT3_APPEND_FL | EXT3_IMMUTABLE_FL)) { ++ if ((oldflags & EXT3_IMMUTABLE_FL) || ++ ((flags ^ oldflags) & (EXT3_APPEND_FL | ++ EXT3_IMMUTABLE_FL | EXT3_IXUNLINK_FL))) { + if (!capable(CAP_LINUX_IMMUTABLE)) + goto flags_out; + } +@@ -93,7 +129,7 @@ long ext3_ioctl(struct file *filp, unsig + if (err) + goto flags_err; + +- flags = flags & EXT3_FL_USER_MODIFIABLE; ++ flags &= EXT3_FL_USER_MODIFIABLE; + flags |= oldflags & ~EXT3_FL_USER_MODIFIABLE; + ei->i_flags = flags; + +diff -NurpP --minimal linux-3.0.9/fs/ext3/namei.c linux-3.0.9-vs2.3.2.1/fs/ext3/namei.c +--- linux-3.0.9/fs/ext3/namei.c 2011-11-15 16:40:46.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/fs/ext3/namei.c 2011-08-29 03:45:09.000000000 +0200 +@@ -36,6 +36,7 @@ + #include <linux/quotaops.h> + #include <linux/buffer_head.h> + #include <linux/bio.h> ++#include <linux/vs_tag.h> + + #include "namei.h" + #include "xattr.h" +@@ -923,6 +924,7 @@ restart: + if (bh) + ll_rw_block(READ_META, 1, &bh); + } ++ dx_propagate_tag(nd, inode); + } + if ((bh = bh_use[ra_ptr++]) == NULL) + goto next; +@@ -2536,6 +2538,7 @@ const struct inode_operations ext3_dir_i + .removexattr = generic_removexattr, + #endif + .check_acl = ext3_check_acl, ++ .sync_flags = ext3_sync_flags, + }; + + const struct inode_operations ext3_special_inode_operations = { +diff -NurpP --minimal linux-3.0.9/fs/ext3/super.c linux-3.0.9-vs2.3.2.1/fs/ext3/super.c +--- linux-3.0.9/fs/ext3/super.c 2011-07-22 11:18:05.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/fs/ext3/super.c 2011-06-10 22:11:24.000000000 +0200 +@@ -821,7 +821,8 @@ enum { + Opt_usrjquota, Opt_grpjquota, Opt_offusrjquota, Opt_offgrpjquota, + Opt_jqfmt_vfsold, Opt_jqfmt_vfsv0, Opt_jqfmt_vfsv1, Opt_quota, + Opt_noquota, Opt_ignore, Opt_barrier, Opt_nobarrier, Opt_err, +- Opt_resize, Opt_usrquota, Opt_grpquota ++ Opt_resize, Opt_usrquota, Opt_grpquota, ++ Opt_tag, Opt_notag, Opt_tagid + }; + + static const match_table_t tokens = { +@@ -878,6 +879,9 @@ static const match_table_t tokens = { + {Opt_barrier, "barrier"}, + {Opt_nobarrier, "nobarrier"}, + {Opt_resize, "resize"}, ++ {Opt_tag, "tag"}, ++ {Opt_notag, "notag"}, ++ {Opt_tagid, "tagid=%u"}, + {Opt_err, NULL}, + }; + +@@ -1030,6 +1034,20 @@ static int parse_options (char *options, + case Opt_nouid32: + set_opt (sbi->s_mount_opt, NO_UID32); + break; ++#ifndef CONFIG_TAGGING_NONE ++ case Opt_tag: ++ set_opt (sbi->s_mount_opt, TAGGED); ++ break; ++ case Opt_notag: ++ clear_opt (sbi->s_mount_opt, TAGGED); ++ break; ++#endif ++#ifdef CONFIG_PROPAGATE ++ case Opt_tagid: ++ /* use args[0] */ ++ set_opt (sbi->s_mount_opt, TAGGED); ++ break; ++#endif + case Opt_nocheck: + clear_opt (sbi->s_mount_opt, CHECK); + break; +@@ -1724,6 +1742,9 @@ static int ext3_fill_super (struct super + NULL, 0)) + goto failed_mount; + ++ if (EXT3_SB(sb)->s_mount_opt & EXT3_MOUNT_TAGGED) ++ sb->s_flags |= MS_TAGGED; ++ + sb->s_flags = (sb->s_flags & ~MS_POSIXACL) | + (test_opt(sb, POSIX_ACL) ? MS_POSIXACL : 0); + +@@ -2604,6 +2625,14 @@ static int ext3_remount (struct super_bl + if (test_opt(sb, ABORT)) + ext3_abort(sb, __func__, "Abort forced by user"); + ++ if ((sbi->s_mount_opt & EXT3_MOUNT_TAGGED) && ++ !(sb->s_flags & MS_TAGGED)) { ++ printk("EXT3-fs: %s: tagging not permitted on remount.\n", ++ sb->s_id); ++ err = -EINVAL; ++ goto restore_opts; ++ } ++ + sb->s_flags = (sb->s_flags & ~MS_POSIXACL) | + (test_opt(sb, POSIX_ACL) ? MS_POSIXACL : 0); + +diff -NurpP --minimal linux-3.0.9/fs/ext4/ext4.h linux-3.0.9-vs2.3.2.1/fs/ext4/ext4.h +--- linux-3.0.9/fs/ext4/ext4.h 2011-11-15 16:40:46.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/fs/ext4/ext4.h 2011-11-15 17:37:06.000000000 +0100 +@@ -351,8 +351,12 @@ struct flex_groups { + #define EXT4_EXTENTS_FL 0x00080000 /* Inode uses extents */ + #define EXT4_EA_INODE_FL 0x00200000 /* Inode used for large EA */ + #define EXT4_EOFBLOCKS_FL 0x00400000 /* Blocks allocated beyond EOF */ ++#define EXT4_IXUNLINK_FL 0x08000000 /* Immutable invert on unlink */ + #define EXT4_RESERVED_FL 0x80000000 /* reserved for ext4 lib */ + ++#define EXT4_BARRIER_FL 0x04000000 /* Barrier for chroot() */ ++#define EXT4_COW_FL 0x20000000 /* Copy on Write marker */ ++ + #define EXT4_FL_USER_VISIBLE 0x004BDFFF /* User visible flags */ + #define EXT4_FL_USER_MODIFIABLE 0x004B80FF /* User modifiable flags */ + +@@ -609,7 +613,8 @@ struct ext4_inode { + __le16 l_i_file_acl_high; + __le16 l_i_uid_high; /* these 2 fields */ + __le16 l_i_gid_high; /* were reserved2[0] */ +- __u32 l_i_reserved2; ++ __le16 l_i_tag; /* Context Tag */ ++ __u16 l_i_reserved2; + } linux2; + struct { + __le16 h_i_reserved1; /* Obsoleted fragment number/size which are removed in ext4 */ +@@ -727,6 +732,7 @@ do { \ + #define i_gid_low i_gid + #define i_uid_high osd2.linux2.l_i_uid_high + #define i_gid_high osd2.linux2.l_i_gid_high ++#define i_raw_tag osd2.linux2.l_i_tag + #define i_reserved2 osd2.linux2.l_i_reserved2 + + #elif defined(__GNU__) +@@ -903,6 +909,7 @@ struct ext4_inode_info { + #define EXT4_MOUNT_POSIX_ACL 0x08000 /* POSIX Access Control Lists */ + #define EXT4_MOUNT_NO_AUTO_DA_ALLOC 0x10000 /* No auto delalloc mapping */ + #define EXT4_MOUNT_BARRIER 0x20000 /* Use block barriers */ ++#define EXT4_MOUNT_TAGGED 0x40000 /* Enable Context Tags */ + #define EXT4_MOUNT_QUOTA 0x80000 /* Some quota option set */ + #define EXT4_MOUNT_USRQUOTA 0x100000 /* "old" user quota */ + #define EXT4_MOUNT_GRPQUOTA 0x200000 /* "old" group quota */ +@@ -2173,6 +2180,7 @@ extern int ext4_map_blocks(handle_t *han + struct ext4_map_blocks *map, int flags); + extern int ext4_fiemap(struct inode *inode, struct fiemap_extent_info *fieinfo, + __u64 start, __u64 len); ++extern int ext4_sync_flags(struct inode *, int, int); + /* move_extent.c */ + extern int ext4_move_extents(struct file *o_filp, struct file *d_filp, + __u64 start_orig, __u64 start_donor, +diff -NurpP --minimal linux-3.0.9/fs/ext4/file.c linux-3.0.9-vs2.3.2.1/fs/ext4/file.c +--- linux-3.0.9/fs/ext4/file.c 2011-07-22 11:18:05.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/fs/ext4/file.c 2011-06-10 22:11:24.000000000 +0200 +@@ -282,5 +282,6 @@ const struct inode_operations ext4_file_ + #endif + .check_acl = ext4_check_acl, + .fiemap = ext4_fiemap, ++ .sync_flags = ext4_sync_flags, + }; + +diff -NurpP --minimal linux-3.0.9/fs/ext4/ialloc.c linux-3.0.9-vs2.3.2.1/fs/ext4/ialloc.c +--- linux-3.0.9/fs/ext4/ialloc.c 2011-05-22 16:17:52.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/fs/ext4/ialloc.c 2011-06-10 22:11:24.000000000 +0200 +@@ -22,6 +22,7 @@ + #include <linux/random.h> + #include <linux/bitops.h> + #include <linux/blkdev.h> ++#include <linux/vs_tag.h> + #include <asm/byteorder.h> + + #include "ext4.h" +@@ -992,6 +993,7 @@ got: + inode->i_mode = mode; + inode->i_uid = current_fsuid(); + inode->i_gid = dir->i_gid; ++ inode->i_tag = dx_current_fstag(sb); + } else + inode_init_owner(inode, dir, mode); + +diff -NurpP --minimal linux-3.0.9/fs/ext4/inode.c linux-3.0.9-vs2.3.2.1/fs/ext4/inode.c +--- linux-3.0.9/fs/ext4/inode.c 2011-11-15 16:40:46.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/fs/ext4/inode.c 2011-11-15 17:37:06.000000000 +0100 +@@ -42,6 +42,7 @@ + #include <linux/printk.h> + #include <linux/slab.h> + #include <linux/ratelimit.h> ++#include <linux/vs_tag.h> + + #include "ext4_jbd2.h" + #include "xattr.h" +@@ -4816,41 +4817,64 @@ void ext4_set_inode_flags(struct inode * + { + unsigned int flags = EXT4_I(inode)->i_flags; + +- inode->i_flags &= ~(S_SYNC|S_APPEND|S_IMMUTABLE|S_NOATIME|S_DIRSYNC); ++ inode->i_flags &= ~(S_IMMUTABLE | S_IXUNLINK | ++ S_SYNC | S_APPEND | S_NOATIME | S_DIRSYNC); ++ ++ if (flags & EXT4_IMMUTABLE_FL) ++ inode->i_flags |= S_IMMUTABLE; ++ if (flags & EXT4_IXUNLINK_FL) ++ inode->i_flags |= S_IXUNLINK; ++ + if (flags & EXT4_SYNC_FL) + inode->i_flags |= S_SYNC; + if (flags & EXT4_APPEND_FL) + inode->i_flags |= S_APPEND; +- if (flags & EXT4_IMMUTABLE_FL) +- inode->i_flags |= S_IMMUTABLE; + if (flags & EXT4_NOATIME_FL) + inode->i_flags |= S_NOATIME; + if (flags & EXT4_DIRSYNC_FL) + inode->i_flags |= S_DIRSYNC; ++ ++ inode->i_vflags &= ~(V_BARRIER | V_COW); ++ ++ if (flags & EXT4_BARRIER_FL) ++ inode->i_vflags |= V_BARRIER; ++ if (flags & EXT4_COW_FL) ++ inode->i_vflags |= V_COW; + } + + /* Propagate flags from i_flags to EXT4_I(inode)->i_flags */ + void ext4_get_inode_flags(struct ext4_inode_info *ei) + { +- unsigned int vfs_fl; ++ unsigned int vfs_fl, vfs_vf; + unsigned long old_fl, new_fl; + + do { + vfs_fl = ei->vfs_inode.i_flags; ++ vfs_vf = ei->vfs_inode.i_vflags; + old_fl = ei->i_flags; + new_fl = old_fl & ~(EXT4_SYNC_FL|EXT4_APPEND_FL| + EXT4_IMMUTABLE_FL|EXT4_NOATIME_FL| +- EXT4_DIRSYNC_FL); ++ EXT4_DIRSYNC_FL|EXT4_BARRIER_FL| ++ EXT4_COW_FL); ++ ++ if (vfs_fl & S_IMMUTABLE) ++ new_fl |= EXT4_IMMUTABLE_FL; ++ if (vfs_fl & S_IXUNLINK) ++ new_fl |= EXT4_IXUNLINK_FL; ++ + if (vfs_fl & S_SYNC) + new_fl |= EXT4_SYNC_FL; + if (vfs_fl & S_APPEND) + new_fl |= EXT4_APPEND_FL; +- if (vfs_fl & S_IMMUTABLE) +- new_fl |= EXT4_IMMUTABLE_FL; + if (vfs_fl & S_NOATIME) + new_fl |= EXT4_NOATIME_FL; + if (vfs_fl & S_DIRSYNC) + new_fl |= EXT4_DIRSYNC_FL; ++ ++ if (vfs_vf & V_BARRIER) ++ new_fl |= EXT4_BARRIER_FL; ++ if (vfs_vf & V_COW) ++ new_fl |= EXT4_COW_FL; + } while (cmpxchg(&ei->i_flags, old_fl, new_fl) != old_fl); + } + +@@ -4886,6 +4910,8 @@ struct inode *ext4_iget(struct super_blo + journal_t *journal = EXT4_SB(sb)->s_journal; + long ret; + int block; ++ uid_t uid; ++ gid_t gid; + + inode = iget_locked(sb, ino); + if (!inode) +@@ -4901,12 +4927,16 @@ struct inode *ext4_iget(struct super_blo + goto bad_inode; + raw_inode = ext4_raw_inode(&iloc); + inode->i_mode = le16_to_cpu(raw_inode->i_mode); +- inode->i_uid = (uid_t)le16_to_cpu(raw_inode->i_uid_low); +- inode->i_gid = (gid_t)le16_to_cpu(raw_inode->i_gid_low); ++ uid = (uid_t)le16_to_cpu(raw_inode->i_uid_low); ++ gid = (gid_t)le16_to_cpu(raw_inode->i_gid_low); + if (!(test_opt(inode->i_sb, NO_UID32))) { +- inode->i_uid |= le16_to_cpu(raw_inode->i_uid_high) << 16; +- inode->i_gid |= le16_to_cpu(raw_inode->i_gid_high) << 16; ++ uid |= le16_to_cpu(raw_inode->i_uid_high) << 16; ++ gid |= le16_to_cpu(raw_inode->i_gid_high) << 16; + } ++ inode->i_uid = INOTAG_UID(DX_TAG(inode), uid, gid); ++ inode->i_gid = INOTAG_GID(DX_TAG(inode), uid, gid); ++ inode->i_tag = INOTAG_TAG(DX_TAG(inode), uid, gid, ++ le16_to_cpu(raw_inode->i_raw_tag)); + inode->i_nlink = le16_to_cpu(raw_inode->i_links_count); + + ext4_clear_state_flags(ei); /* Only relevant on 32-bit archs */ +@@ -5125,6 +5155,8 @@ static int ext4_do_update_inode(handle_t + struct ext4_inode *raw_inode = ext4_raw_inode(iloc); + struct ext4_inode_info *ei = EXT4_I(inode); + struct buffer_head *bh = iloc->bh; ++ uid_t uid = TAGINO_UID(DX_TAG(inode), inode->i_uid, inode->i_tag); ++ gid_t gid = TAGINO_GID(DX_TAG(inode), inode->i_gid, inode->i_tag); + int err = 0, rc, block; + + /* For fields not not tracking in the in-memory inode, +@@ -5135,29 +5167,32 @@ static int ext4_do_update_inode(handle_t + ext4_get_inode_flags(ei); + raw_inode->i_mode = cpu_to_le16(inode->i_mode); + if (!(test_opt(inode->i_sb, NO_UID32))) { +- raw_inode->i_uid_low = cpu_to_le16(low_16_bits(inode->i_uid)); +- raw_inode->i_gid_low = cpu_to_le16(low_16_bits(inode->i_gid)); ++ raw_inode->i_uid_low = cpu_to_le16(low_16_bits(uid)); ++ raw_inode->i_gid_low = cpu_to_le16(low_16_bits(gid)); + /* + * Fix up interoperability with old kernels. Otherwise, old inodes get + * re-used with the upper 16 bits of the uid/gid intact + */ + if (!ei->i_dtime) { + raw_inode->i_uid_high = +- cpu_to_le16(high_16_bits(inode->i_uid)); ++ cpu_to_le16(high_16_bits(uid)); + raw_inode->i_gid_high = +- cpu_to_le16(high_16_bits(inode->i_gid)); ++ cpu_to_le16(high_16_bits(gid)); + } else { + raw_inode->i_uid_high = 0; + raw_inode->i_gid_high = 0; + } + } else { + raw_inode->i_uid_low = +- cpu_to_le16(fs_high2lowuid(inode->i_uid)); ++ cpu_to_le16(fs_high2lowuid(uid)); + raw_inode->i_gid_low = +- cpu_to_le16(fs_high2lowgid(inode->i_gid)); ++ cpu_to_le16(fs_high2lowgid(gid)); + raw_inode->i_uid_high = 0; + raw_inode->i_gid_high = 0; + } ++#ifdef CONFIG_TAGGING_INTERN ++ raw_inode->i_raw_tag = cpu_to_le16(inode->i_tag); ++#endif + raw_inode->i_links_count = cpu_to_le16(inode->i_nlink); + + EXT4_INODE_SET_XTIME(i_ctime, inode, raw_inode); +@@ -5343,7 +5378,8 @@ int ext4_setattr(struct dentry *dentry, + if (is_quota_modification(inode, attr)) + dquot_initialize(inode); + if ((ia_valid & ATTR_UID && attr->ia_uid != inode->i_uid) || +- (ia_valid & ATTR_GID && attr->ia_gid != inode->i_gid)) { ++ (ia_valid & ATTR_GID && attr->ia_gid != inode->i_gid) || ++ (ia_valid & ATTR_TAG && attr->ia_tag != inode->i_tag)) { + handle_t *handle; + + /* (user+group)*(old+new) structure, inode write (sb, +@@ -5365,6 +5401,8 @@ int ext4_setattr(struct dentry *dentry, + inode->i_uid = attr->ia_uid; + if (attr->ia_valid & ATTR_GID) + inode->i_gid = attr->ia_gid; ++ if ((attr->ia_valid & ATTR_TAG) && IS_TAGGED(inode)) ++ inode->i_tag = attr->ia_tag; + error = ext4_mark_inode_dirty(handle, inode); + ext4_journal_stop(handle); + } +diff -NurpP --minimal linux-3.0.9/fs/ext4/ioctl.c linux-3.0.9-vs2.3.2.1/fs/ext4/ioctl.c +--- linux-3.0.9/fs/ext4/ioctl.c 2011-05-22 16:17:52.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/fs/ext4/ioctl.c 2011-06-10 22:11:24.000000000 +0200 +@@ -14,10 +14,39 @@ + #include <linux/compat.h> + #include <linux/mount.h> + #include <linux/file.h> ++#include <linux/vs_tag.h> + #include <asm/uaccess.h> + #include "ext4_jbd2.h" + #include "ext4.h" + ++ ++int ext4_sync_flags(struct inode *inode, int flags, int vflags) ++{ ++ handle_t *handle = NULL; ++ struct ext4_iloc iloc; ++ int err; ++ ++ handle = ext4_journal_start(inode, 1); ++ if (IS_ERR(handle)) ++ return PTR_ERR(handle); ++ ++ if (IS_SYNC(inode)) ++ ext4_handle_sync(handle); ++ err = ext4_reserve_inode_write(handle, inode, &iloc); ++ if (err) ++ goto flags_err; ++ ++ inode->i_flags = flags; ++ inode->i_vflags = vflags; ++ ext4_get_inode_flags(EXT4_I(inode)); ++ inode->i_ctime = ext4_current_time(inode); ++ ++ err = ext4_mark_iloc_dirty(handle, inode, &iloc); ++flags_err: ++ ext4_journal_stop(handle); ++ return err; ++} ++ + long ext4_ioctl(struct file *filp, unsigned int cmd, unsigned long arg) + { + struct inode *inode = filp->f_dentry->d_inode; +@@ -50,6 +79,11 @@ long ext4_ioctl(struct file *filp, unsig + + flags = ext4_mask_flags(inode->i_mode, flags); + ++ if (IS_BARRIER(inode)) { ++ vxwprintk_task(1, "messing with the barrier."); ++ return -EACCES; ++ } ++ + err = -EPERM; + mutex_lock(&inode->i_mutex); + /* Is it quota file? Do not allow user to mess with it */ +@@ -67,7 +101,9 @@ long ext4_ioctl(struct file *filp, unsig + * + * This test looks nicer. Thanks to Pauline Middelink + */ +- if ((flags ^ oldflags) & (EXT4_APPEND_FL | EXT4_IMMUTABLE_FL)) { ++ if ((oldflags & EXT4_IMMUTABLE_FL) || ++ ((flags ^ oldflags) & (EXT4_APPEND_FL | ++ EXT4_IMMUTABLE_FL | EXT4_IXUNLINK_FL))) { + if (!capable(CAP_LINUX_IMMUTABLE)) + goto flags_out; + } +diff -NurpP --minimal linux-3.0.9/fs/ext4/namei.c linux-3.0.9-vs2.3.2.1/fs/ext4/namei.c +--- linux-3.0.9/fs/ext4/namei.c 2011-11-15 16:40:46.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/fs/ext4/namei.c 2011-11-15 17:37:06.000000000 +0100 +@@ -34,6 +34,7 @@ + #include <linux/quotaops.h> + #include <linux/buffer_head.h> + #include <linux/bio.h> ++#include <linux/vs_tag.h> + #include "ext4.h" + #include "ext4_jbd2.h" + +@@ -924,6 +925,7 @@ restart: + if (bh) + ll_rw_block(READ_META, 1, &bh); + } ++ dx_propagate_tag(nd, inode); + } + if ((bh = bh_use[ra_ptr++]) == NULL) + goto next; +@@ -2598,6 +2600,7 @@ const struct inode_operations ext4_dir_i + #endif + .check_acl = ext4_check_acl, + .fiemap = ext4_fiemap, ++ .sync_flags = ext4_sync_flags, + }; + + const struct inode_operations ext4_special_inode_operations = { +diff -NurpP --minimal linux-3.0.9/fs/ext4/super.c linux-3.0.9-vs2.3.2.1/fs/ext4/super.c +--- linux-3.0.9/fs/ext4/super.c 2011-11-15 16:40:46.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/fs/ext4/super.c 2011-08-31 19:37:44.000000000 +0200 +@@ -1293,6 +1293,7 @@ enum { + Opt_dioread_nolock, Opt_dioread_lock, + Opt_discard, Opt_nodiscard, + Opt_init_inode_table, Opt_noinit_inode_table, ++ Opt_tag, Opt_notag, Opt_tagid + }; + + static const match_table_t tokens = { +@@ -1368,6 +1369,9 @@ static const match_table_t tokens = { + {Opt_init_inode_table, "init_itable=%u"}, + {Opt_init_inode_table, "init_itable"}, + {Opt_noinit_inode_table, "noinit_itable"}, ++ {Opt_tag, "tag"}, ++ {Opt_notag, "notag"}, ++ {Opt_tagid, "tagid=%u"}, + {Opt_err, NULL}, + }; + +@@ -1536,6 +1540,20 @@ static int parse_options(char *options, + case Opt_nouid32: + set_opt(sb, NO_UID32); + break; ++#ifndef CONFIG_TAGGING_NONE ++ case Opt_tag: ++ set_opt(sb, TAGGED); ++ break; ++ case Opt_notag: ++ clear_opt(sb, TAGGED); ++ break; ++#endif ++#ifdef CONFIG_PROPAGATE ++ case Opt_tagid: ++ /* use args[0] */ ++ set_opt(sb, TAGGED); ++ break; ++#endif + case Opt_debug: + set_opt(sb, DEBUG); + break; +@@ -3193,6 +3211,9 @@ static int ext4_fill_super(struct super_ + &journal_ioprio, NULL, 0)) + goto failed_mount; + ++ if (EXT4_SB(sb)->s_mount_opt & EXT4_MOUNT_TAGGED) ++ sb->s_flags |= MS_TAGGED; ++ + sb->s_flags = (sb->s_flags & ~MS_POSIXACL) | + (test_opt(sb, POSIX_ACL) ? MS_POSIXACL : 0); + +@@ -4323,6 +4344,14 @@ static int ext4_remount(struct super_blo + if (sbi->s_mount_flags & EXT4_MF_FS_ABORTED) + ext4_abort(sb, "Abort forced by user"); + ++ if ((sbi->s_mount_opt & EXT4_MOUNT_TAGGED) && ++ !(sb->s_flags & MS_TAGGED)) { ++ printk("EXT4-fs: %s: tagging not permitted on remount.\n", ++ sb->s_id); ++ err = -EINVAL; ++ goto restore_opts; ++ } ++ + sb->s_flags = (sb->s_flags & ~MS_POSIXACL) | + (test_opt(sb, POSIX_ACL) ? MS_POSIXACL : 0); + +diff -NurpP --minimal linux-3.0.9/fs/fcntl.c linux-3.0.9-vs2.3.2.1/fs/fcntl.c +--- linux-3.0.9/fs/fcntl.c 2011-05-22 16:17:52.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/fs/fcntl.c 2011-06-10 22:11:24.000000000 +0200 +@@ -20,6 +20,7 @@ + #include <linux/signal.h> + #include <linux/rcupdate.h> + #include <linux/pid_namespace.h> ++#include <linux/vs_limit.h> + + #include <asm/poll.h> + #include <asm/siginfo.h> +@@ -103,6 +104,8 @@ SYSCALL_DEFINE3(dup3, unsigned int, oldf + + if (tofree) + filp_close(tofree, files); ++ else ++ vx_openfd_inc(newfd); /* fd was unused */ + + return newfd; + +@@ -447,6 +450,8 @@ SYSCALL_DEFINE3(fcntl, unsigned int, fd, + filp = fget_raw(fd); + if (!filp) + goto out; ++ if (!vx_files_avail(1)) ++ goto out; + + if (unlikely(filp->f_mode & FMODE_PATH)) { + if (!check_fcntl_cmd(cmd)) { +diff -NurpP --minimal linux-3.0.9/fs/file.c linux-3.0.9-vs2.3.2.1/fs/file.c +--- linux-3.0.9/fs/file.c 2011-05-22 16:17:52.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/fs/file.c 2011-06-10 22:11:24.000000000 +0200 +@@ -21,6 +21,7 @@ + #include <linux/spinlock.h> + #include <linux/rcupdate.h> + #include <linux/workqueue.h> ++#include <linux/vs_limit.h> + + struct fdtable_defer { + spinlock_t lock; +@@ -359,6 +360,8 @@ struct files_struct *dup_fd(struct files + struct file *f = *old_fds++; + if (f) { + get_file(f); ++ /* TODO: sum it first for check and performance */ ++ vx_openfd_inc(open_files - i); + } else { + /* + * The fd may be claimed in the fd bitmap but not yet +@@ -466,6 +469,7 @@ repeat: + else + FD_CLR(fd, fdt->close_on_exec); + error = fd; ++ vx_openfd_inc(fd); + #if 1 + /* Sanity check */ + if (rcu_dereference_raw(fdt->fd[fd]) != NULL) { +diff -NurpP --minimal linux-3.0.9/fs/file_table.c linux-3.0.9-vs2.3.2.1/fs/file_table.c +--- linux-3.0.9/fs/file_table.c 2011-05-22 16:17:52.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/fs/file_table.c 2011-06-10 22:11:24.000000000 +0200 +@@ -24,6 +24,8 @@ + #include <linux/percpu_counter.h> + #include <linux/percpu.h> + #include <linux/ima.h> ++#include <linux/vs_limit.h> ++#include <linux/vs_context.h> + + #include <asm/atomic.h> + +@@ -135,6 +137,8 @@ struct file *get_empty_filp(void) + spin_lock_init(&f->f_lock); + eventpoll_init_file(f); + /* f->f_version: 0 */ ++ f->f_xid = vx_current_xid(); ++ vx_files_inc(f); + return f; + + over: +@@ -253,6 +257,8 @@ static void __fput(struct file *file) + } + fops_put(file->f_op); + put_pid(file->f_owner.pid); ++ vx_files_dec(file); ++ file->f_xid = 0; + file_sb_list_del(file); + if ((file->f_mode & (FMODE_READ | FMODE_WRITE)) == FMODE_READ) + i_readcount_dec(inode); +@@ -383,6 +389,8 @@ void put_filp(struct file *file) + { + if (atomic_long_dec_and_test(&file->f_count)) { + security_file_free(file); ++ vx_files_dec(file); ++ file->f_xid = 0; + file_sb_list_del(file); + file_free(file); + } +diff -NurpP --minimal linux-3.0.9/fs/fs_struct.c linux-3.0.9-vs2.3.2.1/fs/fs_struct.c +--- linux-3.0.9/fs/fs_struct.c 2011-03-15 18:07:31.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/fs/fs_struct.c 2011-06-10 22:11:24.000000000 +0200 +@@ -4,6 +4,7 @@ + #include <linux/path.h> + #include <linux/slab.h> + #include <linux/fs_struct.h> ++#include <linux/vserver/global.h> + #include "internal.h" + + static inline void path_get_longterm(struct path *path) +@@ -96,6 +97,7 @@ void free_fs_struct(struct fs_struct *fs + { + path_put_longterm(&fs->root); + path_put_longterm(&fs->pwd); ++ atomic_dec(&vs_global_fs); + kmem_cache_free(fs_cachep, fs); + } + +@@ -135,6 +137,7 @@ struct fs_struct *copy_fs_struct(struct + fs->pwd = old->pwd; + path_get_longterm(&fs->pwd); + spin_unlock(&old->lock); ++ atomic_inc(&vs_global_fs); + } + return fs; + } +diff -NurpP --minimal linux-3.0.9/fs/gfs2/file.c linux-3.0.9-vs2.3.2.1/fs/gfs2/file.c +--- linux-3.0.9/fs/gfs2/file.c 2011-07-22 11:18:05.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/fs/gfs2/file.c 2011-06-10 22:11:24.000000000 +0200 +@@ -134,6 +134,9 @@ static const u32 fsflags_to_gfs2[32] = { + [7] = GFS2_DIF_NOATIME, + [12] = GFS2_DIF_EXHASH, + [14] = GFS2_DIF_INHERIT_JDATA, ++ [27] = GFS2_DIF_IXUNLINK, ++ [26] = GFS2_DIF_BARRIER, ++ [29] = GFS2_DIF_COW, + }; + + static const u32 gfs2_to_fsflags[32] = { +@@ -143,6 +146,9 @@ static const u32 gfs2_to_fsflags[32] = { + [gfs2fl_NoAtime] = FS_NOATIME_FL, + [gfs2fl_ExHash] = FS_INDEX_FL, + [gfs2fl_InheritJdata] = FS_JOURNAL_DATA_FL, ++ [gfs2fl_IXUnlink] = FS_IXUNLINK_FL, ++ [gfs2fl_Barrier] = FS_BARRIER_FL, ++ [gfs2fl_Cow] = FS_COW_FL, + }; + + static int gfs2_get_flags(struct file *filp, u32 __user *ptr) +@@ -173,10 +179,16 @@ void gfs2_set_inode_flags(struct inode * + { + struct gfs2_inode *ip = GFS2_I(inode); + unsigned int flags = inode->i_flags; ++ unsigned int vflags = inode->i_vflags; ++ ++ flags &= ~(S_IMMUTABLE | S_IXUNLINK | ++ S_SYNC | S_APPEND | S_NOATIME | S_DIRSYNC); + +- flags &= ~(S_SYNC|S_APPEND|S_IMMUTABLE|S_NOATIME|S_DIRSYNC); + if (ip->i_diskflags & GFS2_DIF_IMMUTABLE) + flags |= S_IMMUTABLE; ++ if (ip->i_diskflags & GFS2_DIF_IXUNLINK) ++ flags |= S_IXUNLINK; ++ + if (ip->i_diskflags & GFS2_DIF_APPENDONLY) + flags |= S_APPEND; + if (ip->i_diskflags & GFS2_DIF_NOATIME) +@@ -184,6 +196,43 @@ void gfs2_set_inode_flags(struct inode * + if (ip->i_diskflags & GFS2_DIF_SYNC) + flags |= S_SYNC; + inode->i_flags = flags; ++ ++ vflags &= ~(V_BARRIER | V_COW); ++ ++ if (ip->i_diskflags & GFS2_DIF_BARRIER) ++ vflags |= V_BARRIER; ++ if (ip->i_diskflags & GFS2_DIF_COW) ++ vflags |= V_COW; ++ inode->i_vflags = vflags; ++} ++ ++void gfs2_get_inode_flags(struct inode *inode) ++{ ++ struct gfs2_inode *ip = GFS2_I(inode); ++ unsigned int flags = inode->i_flags; ++ unsigned int vflags = inode->i_vflags; ++ ++ ip->i_diskflags &= ~(GFS2_DIF_APPENDONLY | ++ GFS2_DIF_NOATIME | GFS2_DIF_SYNC | ++ GFS2_DIF_IMMUTABLE | GFS2_DIF_IXUNLINK | ++ GFS2_DIF_BARRIER | GFS2_DIF_COW); ++ ++ if (flags & S_IMMUTABLE) ++ ip->i_diskflags |= GFS2_DIF_IMMUTABLE; ++ if (flags & S_IXUNLINK) ++ ip->i_diskflags |= GFS2_DIF_IXUNLINK; ++ ++ if (flags & S_APPEND) ++ ip->i_diskflags |= GFS2_DIF_APPENDONLY; ++ if (flags & S_NOATIME) ++ ip->i_diskflags |= GFS2_DIF_NOATIME; ++ if (flags & S_SYNC) ++ ip->i_diskflags |= GFS2_DIF_SYNC; ++ ++ if (vflags & V_BARRIER) ++ ip->i_diskflags |= GFS2_DIF_BARRIER; ++ if (vflags & V_COW) ++ ip->i_diskflags |= GFS2_DIF_COW; + } + + /* Flags that can be set by user space */ +@@ -295,6 +344,37 @@ static int gfs2_set_flags(struct file *f + return do_gfs2_set_flags(filp, gfsflags, ~GFS2_DIF_JDATA); + } + ++int gfs2_sync_flags(struct inode *inode, int flags, int vflags) ++{ ++ struct gfs2_inode *ip = GFS2_I(inode); ++ struct gfs2_sbd *sdp = GFS2_SB(inode); ++ struct buffer_head *bh; ++ struct gfs2_holder gh; ++ int error; ++ ++ error = gfs2_glock_nq_init(ip->i_gl, LM_ST_EXCLUSIVE, 0, &gh); ++ if (error) ++ return error; ++ error = gfs2_trans_begin(sdp, RES_DINODE, 0); ++ if (error) ++ goto out; ++ error = gfs2_meta_inode_buffer(ip, &bh); ++ if (error) ++ goto out_trans_end; ++ gfs2_trans_add_bh(ip->i_gl, bh, 1); ++ inode->i_flags = flags; ++ inode->i_vflags = vflags; ++ gfs2_get_inode_flags(inode); ++ gfs2_dinode_out(ip, bh->b_data); ++ brelse(bh); ++ gfs2_set_aops(inode); ++out_trans_end: ++ gfs2_trans_end(sdp); ++out: ++ gfs2_glock_dq_uninit(&gh); ++ return error; ++} ++ + static long gfs2_ioctl(struct file *filp, unsigned int cmd, unsigned long arg) + { + switch(cmd) { +diff -NurpP --minimal linux-3.0.9/fs/gfs2/inode.h linux-3.0.9-vs2.3.2.1/fs/gfs2/inode.h +--- linux-3.0.9/fs/gfs2/inode.h 2011-07-22 11:18:05.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/fs/gfs2/inode.h 2011-06-10 22:11:24.000000000 +0200 +@@ -120,6 +120,7 @@ extern const struct file_operations gfs2 + extern const struct file_operations gfs2_dir_fops_nolock; + + extern void gfs2_set_inode_flags(struct inode *inode); ++extern int gfs2_sync_flags(struct inode *inode, int flags, int vflags); + + #ifdef CONFIG_GFS2_FS_LOCKING_DLM + extern const struct file_operations gfs2_file_fops; +diff -NurpP --minimal linux-3.0.9/fs/inode.c linux-3.0.9-vs2.3.2.1/fs/inode.c +--- linux-3.0.9/fs/inode.c 2011-07-22 11:18:05.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/fs/inode.c 2011-07-01 11:35:34.000000000 +0200 +@@ -26,6 +26,7 @@ + #include <linux/ima.h> + #include <linux/cred.h> + #include <linux/buffer_head.h> /* for inode_has_buffers */ ++#include <linux/vs_tag.h> + #include "internal.h" + + /* +@@ -146,6 +147,9 @@ int inode_init_always(struct super_block + struct address_space *const mapping = &inode->i_data; + + inode->i_sb = sb; ++ ++ /* essential because of inode slab reuse */ ++ inode->i_tag = 0; + inode->i_blkbits = sb->s_blocksize_bits; + inode->i_flags = 0; + atomic_set(&inode->i_count, 1); +@@ -166,6 +170,7 @@ int inode_init_always(struct super_block + inode->i_bdev = NULL; + inode->i_cdev = NULL; + inode->i_rdev = 0; ++ inode->i_mdev = 0; + inode->dirtied_when = 0; + + if (security_inode_alloc(inode)) +@@ -404,6 +409,8 @@ void __insert_inode_hash(struct inode *i + } + EXPORT_SYMBOL(__insert_inode_hash); + ++EXPORT_SYMBOL_GPL(__iget); ++ + /** + * remove_inode_hash - remove an inode from the hash + * @inode: inode to unhash +@@ -1643,9 +1650,11 @@ void init_special_inode(struct inode *in + if (S_ISCHR(mode)) { + inode->i_fop = &def_chr_fops; + inode->i_rdev = rdev; ++ inode->i_mdev = rdev; + } else if (S_ISBLK(mode)) { + inode->i_fop = &def_blk_fops; + inode->i_rdev = rdev; ++ inode->i_mdev = rdev; + } else if (S_ISFIFO(mode)) + inode->i_fop = &def_fifo_fops; + else if (S_ISSOCK(mode)) +@@ -1674,6 +1683,7 @@ void inode_init_owner(struct inode *inod + } else + inode->i_gid = current_fsgid(); + inode->i_mode = mode; ++ inode->i_tag = dx_current_fstag(inode->i_sb); + } + EXPORT_SYMBOL(inode_init_owner); + +diff -NurpP --minimal linux-3.0.9/fs/ioctl.c linux-3.0.9-vs2.3.2.1/fs/ioctl.c +--- linux-3.0.9/fs/ioctl.c 2011-05-22 16:17:52.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/fs/ioctl.c 2011-06-10 22:11:24.000000000 +0200 +@@ -15,6 +15,9 @@ + #include <linux/writeback.h> + #include <linux/buffer_head.h> + #include <linux/falloc.h> ++#include <linux/proc_fs.h> ++#include <linux/vserver/inode.h> ++#include <linux/vs_tag.h> + + #include <asm/ioctls.h> + +diff -NurpP --minimal linux-3.0.9/fs/ioprio.c linux-3.0.9-vs2.3.2.1/fs/ioprio.c +--- linux-3.0.9/fs/ioprio.c 2011-01-05 21:50:24.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/fs/ioprio.c 2011-06-10 22:11:24.000000000 +0200 +@@ -27,6 +27,7 @@ + #include <linux/syscalls.h> + #include <linux/security.h> + #include <linux/pid_namespace.h> ++#include <linux/vs_base.h> + + int set_task_ioprio(struct task_struct *task, int ioprio) + { +@@ -119,6 +120,8 @@ SYSCALL_DEFINE3(ioprio_set, int, which, + else + pgrp = find_vpid(who); + do_each_pid_thread(pgrp, PIDTYPE_PGID, p) { ++ if (!vx_check(p->xid, VS_ADMIN_P | VS_IDENT)) ++ continue; + ret = set_task_ioprio(p, ioprio); + if (ret) + break; +@@ -208,6 +211,8 @@ SYSCALL_DEFINE2(ioprio_get, int, which, + else + pgrp = find_vpid(who); + do_each_pid_thread(pgrp, PIDTYPE_PGID, p) { ++ if (!vx_check(p->xid, VS_ADMIN_P | VS_IDENT)) ++ continue; + tmpio = get_task_ioprio(p); + if (tmpio < 0) + continue; +diff -NurpP --minimal linux-3.0.9/fs/jfs/file.c linux-3.0.9-vs2.3.2.1/fs/jfs/file.c +--- linux-3.0.9/fs/jfs/file.c 2011-07-22 11:18:05.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/fs/jfs/file.c 2011-07-01 11:35:34.000000000 +0200 +@@ -102,7 +102,8 @@ int jfs_setattr(struct dentry *dentry, s + if (is_quota_modification(inode, iattr)) + dquot_initialize(inode); + if ((iattr->ia_valid & ATTR_UID && iattr->ia_uid != inode->i_uid) || +- (iattr->ia_valid & ATTR_GID && iattr->ia_gid != inode->i_gid)) { ++ (iattr->ia_valid & ATTR_GID && iattr->ia_gid != inode->i_gid) || ++ (iattr->ia_valid & ATTR_TAG && iattr->ia_tag != inode->i_tag)) { + rc = dquot_transfer(inode, iattr); + if (rc) + return rc; +@@ -133,6 +134,7 @@ const struct inode_operations jfs_file_i + #ifdef CONFIG_JFS_POSIX_ACL + .check_acl = jfs_check_acl, + #endif ++ .sync_flags = jfs_sync_flags, + }; + + const struct file_operations jfs_file_operations = { +diff -NurpP --minimal linux-3.0.9/fs/jfs/ioctl.c linux-3.0.9-vs2.3.2.1/fs/jfs/ioctl.c +--- linux-3.0.9/fs/jfs/ioctl.c 2011-05-22 16:17:52.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/fs/jfs/ioctl.c 2011-06-10 22:11:24.000000000 +0200 +@@ -11,6 +11,7 @@ + #include <linux/mount.h> + #include <linux/time.h> + #include <linux/sched.h> ++#include <linux/mount.h> + #include <asm/current.h> + #include <asm/uaccess.h> + +@@ -52,6 +53,16 @@ static long jfs_map_ext2(unsigned long f + } + + ++int jfs_sync_flags(struct inode *inode, int flags, int vflags) ++{ ++ inode->i_flags = flags; ++ inode->i_vflags = vflags; ++ jfs_get_inode_flags(JFS_IP(inode)); ++ inode->i_ctime = CURRENT_TIME_SEC; ++ mark_inode_dirty(inode); ++ return 0; ++} ++ + long jfs_ioctl(struct file *filp, unsigned int cmd, unsigned long arg) + { + struct inode *inode = filp->f_dentry->d_inode; +@@ -85,6 +96,11 @@ long jfs_ioctl(struct file *filp, unsign + if (!S_ISDIR(inode->i_mode)) + flags &= ~JFS_DIRSYNC_FL; + ++ if (IS_BARRIER(inode)) { ++ vxwprintk_task(1, "messing with the barrier."); ++ return -EACCES; ++ } ++ + /* Is it quota file? Do not allow user to mess with it */ + if (IS_NOQUOTA(inode)) { + err = -EPERM; +@@ -102,8 +118,8 @@ long jfs_ioctl(struct file *filp, unsign + * the relevant capability. + */ + if ((oldflags & JFS_IMMUTABLE_FL) || +- ((flags ^ oldflags) & +- (JFS_APPEND_FL | JFS_IMMUTABLE_FL))) { ++ ((flags ^ oldflags) & (JFS_APPEND_FL | ++ JFS_IMMUTABLE_FL | JFS_IXUNLINK_FL))) { + if (!capable(CAP_LINUX_IMMUTABLE)) { + mutex_unlock(&inode->i_mutex); + err = -EPERM; +@@ -111,7 +127,7 @@ long jfs_ioctl(struct file *filp, unsign + } + } + +- flags = flags & JFS_FL_USER_MODIFIABLE; ++ flags &= JFS_FL_USER_MODIFIABLE; + flags |= oldflags & ~JFS_FL_USER_MODIFIABLE; + jfs_inode->mode2 = flags; + +diff -NurpP --minimal linux-3.0.9/fs/jfs/jfs_dinode.h linux-3.0.9-vs2.3.2.1/fs/jfs/jfs_dinode.h +--- linux-3.0.9/fs/jfs/jfs_dinode.h 2008-12-25 00:26:37.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/fs/jfs/jfs_dinode.h 2011-06-10 22:11:24.000000000 +0200 +@@ -161,9 +161,13 @@ struct dinode { + + #define JFS_APPEND_FL 0x01000000 /* writes to file may only append */ + #define JFS_IMMUTABLE_FL 0x02000000 /* Immutable file */ ++#define JFS_IXUNLINK_FL 0x08000000 /* Immutable invert on unlink */ + +-#define JFS_FL_USER_VISIBLE 0x03F80000 +-#define JFS_FL_USER_MODIFIABLE 0x03F80000 ++#define JFS_BARRIER_FL 0x04000000 /* Barrier for chroot() */ ++#define JFS_COW_FL 0x20000000 /* Copy on Write marker */ ++ ++#define JFS_FL_USER_VISIBLE 0x07F80000 ++#define JFS_FL_USER_MODIFIABLE 0x07F80000 + #define JFS_FL_INHERIT 0x03C80000 + + /* These are identical to EXT[23]_IOC_GETFLAGS/SETFLAGS */ +diff -NurpP --minimal linux-3.0.9/fs/jfs/jfs_filsys.h linux-3.0.9-vs2.3.2.1/fs/jfs/jfs_filsys.h +--- linux-3.0.9/fs/jfs/jfs_filsys.h 2008-12-25 00:26:37.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/fs/jfs/jfs_filsys.h 2011-06-10 22:11:24.000000000 +0200 +@@ -263,6 +263,7 @@ + #define JFS_NAME_MAX 255 + #define JFS_PATH_MAX BPSIZE + ++#define JFS_TAGGED 0x00800000 /* Context Tagging */ + + /* + * file system state (superblock state) +diff -NurpP --minimal linux-3.0.9/fs/jfs/jfs_imap.c linux-3.0.9-vs2.3.2.1/fs/jfs/jfs_imap.c +--- linux-3.0.9/fs/jfs/jfs_imap.c 2011-07-22 11:18:05.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/fs/jfs/jfs_imap.c 2011-07-01 11:35:34.000000000 +0200 +@@ -46,6 +46,7 @@ + #include <linux/pagemap.h> + #include <linux/quotaops.h> + #include <linux/slab.h> ++#include <linux/vs_tag.h> + + #include "jfs_incore.h" + #include "jfs_inode.h" +@@ -3058,6 +3059,8 @@ static int copy_from_dinode(struct dinod + { + struct jfs_inode_info *jfs_ip = JFS_IP(ip); + struct jfs_sb_info *sbi = JFS_SBI(ip->i_sb); ++ uid_t uid; ++ gid_t gid; + + jfs_ip->fileset = le32_to_cpu(dip->di_fileset); + jfs_ip->mode2 = le32_to_cpu(dip->di_mode); +@@ -3078,14 +3081,18 @@ static int copy_from_dinode(struct dinod + } + ip->i_nlink = le32_to_cpu(dip->di_nlink); + +- jfs_ip->saved_uid = le32_to_cpu(dip->di_uid); ++ uid = le32_to_cpu(dip->di_uid); ++ gid = le32_to_cpu(dip->di_gid); ++ ip->i_tag = INOTAG_TAG(DX_TAG(ip), uid, gid, 0); ++ ++ jfs_ip->saved_uid = INOTAG_UID(DX_TAG(ip), uid, gid); + if (sbi->uid == -1) + ip->i_uid = jfs_ip->saved_uid; + else { + ip->i_uid = sbi->uid; + } + +- jfs_ip->saved_gid = le32_to_cpu(dip->di_gid); ++ jfs_ip->saved_gid = INOTAG_GID(DX_TAG(ip), uid, gid); + if (sbi->gid == -1) + ip->i_gid = jfs_ip->saved_gid; + else { +@@ -3150,14 +3157,12 @@ static void copy_to_dinode(struct dinode + dip->di_size = cpu_to_le64(ip->i_size); + dip->di_nblocks = cpu_to_le64(PBLK2LBLK(ip->i_sb, ip->i_blocks)); + dip->di_nlink = cpu_to_le32(ip->i_nlink); +- if (sbi->uid == -1) +- dip->di_uid = cpu_to_le32(ip->i_uid); +- else +- dip->di_uid = cpu_to_le32(jfs_ip->saved_uid); +- if (sbi->gid == -1) +- dip->di_gid = cpu_to_le32(ip->i_gid); +- else +- dip->di_gid = cpu_to_le32(jfs_ip->saved_gid); ++ ++ dip->di_uid = cpu_to_le32(TAGINO_UID(DX_TAG(ip), ++ (sbi->uid == -1) ? ip->i_uid : jfs_ip->saved_uid, ip->i_tag)); ++ dip->di_gid = cpu_to_le32(TAGINO_GID(DX_TAG(ip), ++ (sbi->gid == -1) ? ip->i_gid : jfs_ip->saved_gid, ip->i_tag)); ++ + jfs_get_inode_flags(jfs_ip); + /* + * mode2 is only needed for storing the higher order bits. +diff -NurpP --minimal linux-3.0.9/fs/jfs/jfs_inode.c linux-3.0.9-vs2.3.2.1/fs/jfs/jfs_inode.c +--- linux-3.0.9/fs/jfs/jfs_inode.c 2010-08-02 16:52:49.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/fs/jfs/jfs_inode.c 2011-06-10 22:11:24.000000000 +0200 +@@ -18,6 +18,7 @@ + + #include <linux/fs.h> + #include <linux/quotaops.h> ++#include <linux/vs_tag.h> + #include "jfs_incore.h" + #include "jfs_inode.h" + #include "jfs_filsys.h" +@@ -30,29 +31,46 @@ void jfs_set_inode_flags(struct inode *i + { + unsigned int flags = JFS_IP(inode)->mode2; + +- inode->i_flags &= ~(S_IMMUTABLE | S_APPEND | +- S_NOATIME | S_DIRSYNC | S_SYNC); ++ inode->i_flags &= ~(S_IMMUTABLE | S_IXUNLINK | ++ S_SYNC | S_APPEND | S_NOATIME | S_DIRSYNC); + + if (flags & JFS_IMMUTABLE_FL) + inode->i_flags |= S_IMMUTABLE; ++ if (flags & JFS_IXUNLINK_FL) ++ inode->i_flags |= S_IXUNLINK; ++ ++ if (flags & JFS_SYNC_FL) ++ inode->i_flags |= S_SYNC; + if (flags & JFS_APPEND_FL) + inode->i_flags |= S_APPEND; + if (flags & JFS_NOATIME_FL) + inode->i_flags |= S_NOATIME; + if (flags & JFS_DIRSYNC_FL) + inode->i_flags |= S_DIRSYNC; +- if (flags & JFS_SYNC_FL) +- inode->i_flags |= S_SYNC; ++ ++ inode->i_vflags &= ~(V_BARRIER | V_COW); ++ ++ if (flags & JFS_BARRIER_FL) ++ inode->i_vflags |= V_BARRIER; ++ if (flags & JFS_COW_FL) ++ inode->i_vflags |= V_COW; + } + + void jfs_get_inode_flags(struct jfs_inode_info *jfs_ip) + { + unsigned int flags = jfs_ip->vfs_inode.i_flags; ++ unsigned int vflags = jfs_ip->vfs_inode.i_vflags; ++ ++ jfs_ip->mode2 &= ~(JFS_IMMUTABLE_FL | JFS_IXUNLINK_FL | ++ JFS_APPEND_FL | JFS_NOATIME_FL | ++ JFS_DIRSYNC_FL | JFS_SYNC_FL | ++ JFS_BARRIER_FL | JFS_COW_FL); + +- jfs_ip->mode2 &= ~(JFS_IMMUTABLE_FL | JFS_APPEND_FL | JFS_NOATIME_FL | +- JFS_DIRSYNC_FL | JFS_SYNC_FL); + if (flags & S_IMMUTABLE) + jfs_ip->mode2 |= JFS_IMMUTABLE_FL; ++ if (flags & S_IXUNLINK) ++ jfs_ip->mode2 |= JFS_IXUNLINK_FL; ++ + if (flags & S_APPEND) + jfs_ip->mode2 |= JFS_APPEND_FL; + if (flags & S_NOATIME) +@@ -61,6 +79,11 @@ void jfs_get_inode_flags(struct jfs_inod + jfs_ip->mode2 |= JFS_DIRSYNC_FL; + if (flags & S_SYNC) + jfs_ip->mode2 |= JFS_SYNC_FL; ++ ++ if (vflags & V_BARRIER) ++ jfs_ip->mode2 |= JFS_BARRIER_FL; ++ if (vflags & V_COW) ++ jfs_ip->mode2 |= JFS_COW_FL; + } + + /* +diff -NurpP --minimal linux-3.0.9/fs/jfs/jfs_inode.h linux-3.0.9-vs2.3.2.1/fs/jfs/jfs_inode.h +--- linux-3.0.9/fs/jfs/jfs_inode.h 2011-07-22 11:18:05.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/fs/jfs/jfs_inode.h 2011-06-10 22:11:24.000000000 +0200 +@@ -39,6 +39,7 @@ extern struct dentry *jfs_fh_to_dentry(s + extern struct dentry *jfs_fh_to_parent(struct super_block *sb, struct fid *fid, + int fh_len, int fh_type); + extern void jfs_set_inode_flags(struct inode *); ++extern int jfs_sync_flags(struct inode *, int, int); + extern int jfs_get_block(struct inode *, sector_t, struct buffer_head *, int); + extern int jfs_setattr(struct dentry *, struct iattr *); + +diff -NurpP --minimal linux-3.0.9/fs/jfs/namei.c linux-3.0.9-vs2.3.2.1/fs/jfs/namei.c +--- linux-3.0.9/fs/jfs/namei.c 2011-05-22 16:17:53.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/fs/jfs/namei.c 2011-06-10 22:11:24.000000000 +0200 +@@ -22,6 +22,7 @@ + #include <linux/ctype.h> + #include <linux/quotaops.h> + #include <linux/exportfs.h> ++#include <linux/vs_tag.h> + #include "jfs_incore.h" + #include "jfs_superblock.h" + #include "jfs_inode.h" +@@ -1486,6 +1487,7 @@ static struct dentry *jfs_lookup(struct + return ERR_CAST(ip); + } + ++ dx_propagate_tag(nd, ip); + return d_splice_alias(ip, dentry); + } + +@@ -1550,6 +1552,7 @@ const struct inode_operations jfs_dir_in + #ifdef CONFIG_JFS_POSIX_ACL + .check_acl = jfs_check_acl, + #endif ++ .sync_flags = jfs_sync_flags, + }; + + const struct file_operations jfs_dir_operations = { +diff -NurpP --minimal linux-3.0.9/fs/jfs/super.c linux-3.0.9-vs2.3.2.1/fs/jfs/super.c +--- linux-3.0.9/fs/jfs/super.c 2011-05-22 16:17:53.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/fs/jfs/super.c 2011-06-10 22:11:24.000000000 +0200 +@@ -198,7 +198,8 @@ static void jfs_put_super(struct super_b + enum { + Opt_integrity, Opt_nointegrity, Opt_iocharset, Opt_resize, + Opt_resize_nosize, Opt_errors, Opt_ignore, Opt_err, Opt_quota, +- Opt_usrquota, Opt_grpquota, Opt_uid, Opt_gid, Opt_umask ++ Opt_usrquota, Opt_grpquota, Opt_uid, Opt_gid, Opt_umask, ++ Opt_tag, Opt_notag, Opt_tagid + }; + + static const match_table_t tokens = { +@@ -208,6 +209,10 @@ static const match_table_t tokens = { + {Opt_resize, "resize=%u"}, + {Opt_resize_nosize, "resize"}, + {Opt_errors, "errors=%s"}, ++ {Opt_tag, "tag"}, ++ {Opt_notag, "notag"}, ++ {Opt_tagid, "tagid=%u"}, ++ {Opt_tag, "tagxid"}, + {Opt_ignore, "noquota"}, + {Opt_ignore, "quota"}, + {Opt_usrquota, "usrquota"}, +@@ -342,6 +347,20 @@ static int parse_options(char *options, + } + break; + } ++#ifndef CONFIG_TAGGING_NONE ++ case Opt_tag: ++ *flag |= JFS_TAGGED; ++ break; ++ case Opt_notag: ++ *flag &= JFS_TAGGED; ++ break; ++#endif ++#ifdef CONFIG_PROPAGATE ++ case Opt_tagid: ++ /* use args[0] */ ++ *flag |= JFS_TAGGED; ++ break; ++#endif + default: + printk("jfs: Unrecognized mount option \"%s\" " + " or missing value\n", p); +@@ -373,6 +392,12 @@ static int jfs_remount(struct super_bloc + return -EINVAL; + } + ++ if ((flag & JFS_TAGGED) && !(sb->s_flags & MS_TAGGED)) { ++ printk(KERN_ERR "JFS: %s: tagging not permitted on remount.\n", ++ sb->s_id); ++ return -EINVAL; ++ } ++ + if (newLVSize) { + if (sb->s_flags & MS_RDONLY) { + printk(KERN_ERR +@@ -455,6 +480,9 @@ static int jfs_fill_super(struct super_b + #ifdef CONFIG_JFS_POSIX_ACL + sb->s_flags |= MS_POSIXACL; + #endif ++ /* map mount option tagxid */ ++ if (sbi->flag & JFS_TAGGED) ++ sb->s_flags |= MS_TAGGED; + + if (newLVSize) { + printk(KERN_ERR "resize option for remount only\n"); +diff -NurpP --minimal linux-3.0.9/fs/libfs.c linux-3.0.9-vs2.3.2.1/fs/libfs.c +--- linux-3.0.9/fs/libfs.c 2011-07-22 11:18:05.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/fs/libfs.c 2011-07-22 11:20:39.000000000 +0200 +@@ -133,7 +133,8 @@ static inline unsigned char dt_type(stru + * both impossible due to the lock on directory. + */ + +-int dcache_readdir(struct file * filp, void * dirent, filldir_t filldir) ++static inline int do_dcache_readdir_filter(struct file *filp, ++ void *dirent, filldir_t filldir, int (*filter)(struct dentry *dentry)) + { + struct dentry *dentry = filp->f_path.dentry; + struct dentry *cursor = filp->private_data; +@@ -164,6 +165,8 @@ int dcache_readdir(struct file * filp, v + for (p=q->next; p != &dentry->d_subdirs; p=p->next) { + struct dentry *next; + next = list_entry(p, struct dentry, d_u.d_child); ++ if (filter && !filter(next)) ++ continue; + spin_lock_nested(&next->d_lock, DENTRY_D_LOCK_NESTED); + if (!simple_positive(next)) { + spin_unlock(&next->d_lock); +@@ -190,6 +193,17 @@ int dcache_readdir(struct file * filp, v + return 0; + } + ++int dcache_readdir(struct file *filp, void *dirent, filldir_t filldir) ++{ ++ return do_dcache_readdir_filter(filp, dirent, filldir, NULL); ++} ++ ++int dcache_readdir_filter(struct file *filp, void *dirent, filldir_t filldir, ++ int (*filter)(struct dentry *)) ++{ ++ return do_dcache_readdir_filter(filp, dirent, filldir, filter); ++} ++ + ssize_t generic_read_dir(struct file *filp, char __user *buf, size_t siz, loff_t *ppos) + { + return -EISDIR; +@@ -965,6 +979,7 @@ EXPORT_SYMBOL(dcache_dir_close); + EXPORT_SYMBOL(dcache_dir_lseek); + EXPORT_SYMBOL(dcache_dir_open); + EXPORT_SYMBOL(dcache_readdir); ++EXPORT_SYMBOL(dcache_readdir_filter); + EXPORT_SYMBOL(generic_read_dir); + EXPORT_SYMBOL(mount_pseudo); + EXPORT_SYMBOL(simple_write_begin); +diff -NurpP --minimal linux-3.0.9/fs/locks.c linux-3.0.9-vs2.3.2.1/fs/locks.c +--- linux-3.0.9/fs/locks.c 2011-07-22 11:18:05.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/fs/locks.c 2011-07-19 00:51:58.000000000 +0200 +@@ -126,6 +126,8 @@ + #include <linux/time.h> + #include <linux/rcupdate.h> + #include <linux/pid_namespace.h> ++#include <linux/vs_base.h> ++#include <linux/vs_limit.h> + + #include <asm/uaccess.h> + +@@ -171,13 +173,19 @@ static void locks_init_lock_always(struc + fl->fl_flags = 0; + fl->fl_type = 0; + fl->fl_start = fl->fl_end = 0; ++ fl->fl_xid = -1; + } + ++ + /* Allocate an empty lock structure. */ + struct file_lock *locks_alloc_lock(void) + { +- struct file_lock *fl = kmem_cache_alloc(filelock_cache, GFP_KERNEL); ++ struct file_lock *fl; ++ ++ if (!vx_locks_avail(1)) ++ return NULL; + ++ fl = kmem_cache_alloc(filelock_cache, GFP_KERNEL); + if (fl) + locks_init_lock_always(fl); + +@@ -208,6 +216,7 @@ void locks_free_lock(struct file_lock *f + BUG_ON(!list_empty(&fl->fl_block)); + BUG_ON(!list_empty(&fl->fl_link)); + ++ vx_locks_dec(fl); + locks_release_private(fl); + kmem_cache_free(filelock_cache, fl); + } +@@ -272,6 +281,7 @@ void locks_copy_lock(struct file_lock *n + new->fl_file = fl->fl_file; + new->fl_ops = fl->fl_ops; + new->fl_lmops = fl->fl_lmops; ++ new->fl_xid = fl->fl_xid; + + locks_copy_private(new, fl); + } +@@ -310,6 +320,11 @@ static int flock_make_lock(struct file * + fl->fl_flags = FL_FLOCK; + fl->fl_type = type; + fl->fl_end = OFFSET_MAX; ++ ++ vxd_assert(filp->f_xid == vx_current_xid(), ++ "f_xid(%d) == current(%d)", filp->f_xid, vx_current_xid()); ++ fl->fl_xid = filp->f_xid; ++ vx_locks_inc(fl); + + *lock = fl; + return 0; +@@ -459,6 +474,7 @@ static int lease_init(struct file *filp, + + fl->fl_owner = current->files; + fl->fl_pid = current->tgid; ++ fl->fl_xid = vx_current_xid(); + + fl->fl_file = filp; + fl->fl_flags = FL_LEASE; +@@ -478,6 +494,11 @@ static struct file_lock *lease_alloc(str + if (fl == NULL) + return ERR_PTR(error); + ++ fl->fl_xid = vx_current_xid(); ++ if (filp) ++ vxd_assert(filp->f_xid == fl->fl_xid, ++ "f_xid(%d) == fl_xid(%d)", filp->f_xid, fl->fl_xid); ++ vx_locks_inc(fl); + error = lease_init(filp, type, fl); + if (error) { + locks_free_lock(fl); +@@ -779,6 +800,7 @@ static int flock_lock_file(struct file * + lock_flocks(); + } + ++ new_fl->fl_xid = -1; + find_conflict: + for_each_lock(inode, before) { + struct file_lock *fl = *before; +@@ -799,6 +821,7 @@ find_conflict: + goto out; + locks_copy_lock(new_fl, request); + locks_insert_lock(before, new_fl); ++ vx_locks_inc(new_fl); + new_fl = NULL; + error = 0; + +@@ -809,7 +832,8 @@ out: + return error; + } + +-static int __posix_lock_file(struct inode *inode, struct file_lock *request, struct file_lock *conflock) ++static int __posix_lock_file(struct inode *inode, struct file_lock *request, ++ struct file_lock *conflock, xid_t xid) + { + struct file_lock *fl; + struct file_lock *new_fl = NULL; +@@ -819,6 +843,8 @@ static int __posix_lock_file(struct inod + struct file_lock **before; + int error, added = 0; + ++ vxd_assert(xid == vx_current_xid(), ++ "xid(%d) == current(%d)", xid, vx_current_xid()); + /* + * We may need two file_lock structures for this operation, + * so we get them in advance to avoid races. +@@ -829,7 +855,11 @@ static int __posix_lock_file(struct inod + (request->fl_type != F_UNLCK || + request->fl_start != 0 || request->fl_end != OFFSET_MAX)) { + new_fl = locks_alloc_lock(); ++ new_fl->fl_xid = xid; ++ vx_locks_inc(new_fl); + new_fl2 = locks_alloc_lock(); ++ new_fl2->fl_xid = xid; ++ vx_locks_inc(new_fl2); + } + + lock_flocks(); +@@ -1028,7 +1058,8 @@ static int __posix_lock_file(struct inod + int posix_lock_file(struct file *filp, struct file_lock *fl, + struct file_lock *conflock) + { +- return __posix_lock_file(filp->f_path.dentry->d_inode, fl, conflock); ++ return __posix_lock_file(filp->f_path.dentry->d_inode, ++ fl, conflock, filp->f_xid); + } + EXPORT_SYMBOL(posix_lock_file); + +@@ -1118,7 +1149,7 @@ int locks_mandatory_area(int read_write, + fl.fl_end = offset + count - 1; + + for (;;) { +- error = __posix_lock_file(inode, &fl, NULL); ++ error = __posix_lock_file(inode, &fl, NULL, filp->f_xid); + if (error != FILE_LOCK_DEFERRED) + break; + error = wait_event_interruptible(fl.fl_wait, !fl.fl_next); +@@ -1431,6 +1462,7 @@ int generic_setlease(struct file *filp, + goto out; + + locks_insert_lock(before, lease); ++ vx_locks_inc(lease); + return 0; + + out: +@@ -1815,6 +1847,11 @@ int fcntl_setlk(unsigned int fd, struct + if (file_lock == NULL) + return -ENOLCK; + ++ vxd_assert(filp->f_xid == vx_current_xid(), ++ "f_xid(%d) == current(%d)", filp->f_xid, vx_current_xid()); ++ file_lock->fl_xid = filp->f_xid; ++ vx_locks_inc(file_lock); ++ + /* + * This might block, so we do it before checking the inode. + */ +@@ -1933,6 +1970,11 @@ int fcntl_setlk64(unsigned int fd, struc + if (file_lock == NULL) + return -ENOLCK; + ++ vxd_assert(filp->f_xid == vx_current_xid(), ++ "f_xid(%d) == current(%d)", filp->f_xid, vx_current_xid()); ++ file_lock->fl_xid = filp->f_xid; ++ vx_locks_inc(file_lock); ++ + /* + * This might block, so we do it before checking the inode. + */ +@@ -2198,8 +2240,11 @@ static int locks_show(struct seq_file *f + + lock_get_status(f, fl, *((loff_t *)f->private), ""); + +- list_for_each_entry(bfl, &fl->fl_block, fl_block) ++ list_for_each_entry(bfl, &fl->fl_block, fl_block) { ++ if (!vx_check(fl->fl_xid, VS_WATCH_P | VS_IDENT)) ++ continue; + lock_get_status(f, bfl, *((loff_t *)f->private), " ->"); ++ } + + return 0; + } +diff -NurpP --minimal linux-3.0.9/fs/namei.c linux-3.0.9-vs2.3.2.1/fs/namei.c +--- linux-3.0.9/fs/namei.c 2011-11-15 16:40:47.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/fs/namei.c 2011-11-16 14:27:47.000000000 +0100 +@@ -32,6 +32,14 @@ + #include <linux/fcntl.h> + #include <linux/device_cgroup.h> + #include <linux/fs_struct.h> ++#include <linux/proc_fs.h> ++#include <linux/vserver/inode.h> ++#include <linux/vs_base.h> ++#include <linux/vs_tag.h> ++#include <linux/vs_cowbl.h> ++#include <linux/vs_device.h> ++#include <linux/vs_context.h> ++#include <linux/pid_namespace.h> + #include <asm/uaccess.h> + + #include "internal.h" +@@ -175,6 +183,89 @@ void putname(const char *name) + EXPORT_SYMBOL(putname); + #endif + ++static inline int dx_barrier(const struct inode *inode) ++{ ++ if (IS_BARRIER(inode) && !vx_check(0, VS_ADMIN | VS_WATCH)) { ++ vxwprintk_task(1, "did hit the barrier."); ++ return 1; ++ } ++ return 0; ++} ++ ++static int __dx_permission(const struct inode *inode, int mask) ++{ ++ if (dx_barrier(inode)) ++ return -EACCES; ++ ++ if (inode->i_sb->s_magic == DEVPTS_SUPER_MAGIC) { ++ /* devpts is xid tagged */ ++ if (S_ISDIR(inode->i_mode) || ++ vx_check((xid_t)inode->i_tag, VS_IDENT | VS_WATCH_P)) ++ return 0; ++ ++ /* just pretend we didn't find anything */ ++ return -ENOENT; ++ } ++ else if (inode->i_sb->s_magic == PROC_SUPER_MAGIC) { ++ struct proc_dir_entry *de = PDE(inode); ++ ++ if (de && !vx_hide_check(0, de->vx_flags)) ++ goto out; ++ ++ if ((mask & (MAY_WRITE | MAY_APPEND))) { ++ struct pid *pid; ++ struct task_struct *tsk; ++ ++ if (vx_check(0, VS_ADMIN | VS_WATCH_P) || ++ vx_flags(VXF_STATE_SETUP, 0)) ++ return 0; ++ ++ pid = PROC_I(inode)->pid; ++ if (!pid) ++ goto out; ++ ++ rcu_read_lock(); ++ tsk = pid_task(pid, PIDTYPE_PID); ++ vxdprintk(VXD_CBIT(tag, 0), "accessing %p[#%u]", ++ tsk, (tsk ? vx_task_xid(tsk) : 0)); ++ if (tsk && ++ vx_check(vx_task_xid(tsk), VS_IDENT | VS_WATCH_P)) { ++ rcu_read_unlock(); ++ return 0; ++ } ++ rcu_read_unlock(); ++ } ++ else { ++ /* FIXME: Should we block some entries here? */ ++ return 0; ++ } ++ } ++ else { ++ if (dx_notagcheck(inode->i_sb) || ++ dx_check(inode->i_tag, DX_HOSTID | DX_ADMIN | DX_WATCH | ++ DX_IDENT)) ++ return 0; ++ } ++ ++out: ++ return -EACCES; ++} ++ ++int dx_permission(const struct inode *inode, int mask) ++{ ++ int ret = __dx_permission(inode, mask); ++ if (unlikely(ret)) { ++#ifndef CONFIG_VSERVER_WARN_DEVPTS ++ if (inode->i_sb->s_magic != DEVPTS_SUPER_MAGIC) ++#endif ++ vxwprintk_task(1, ++ "denied [0x%x] access to inode %s:%p[#%d,%lu]", ++ mask, inode->i_sb->s_id, inode, inode->i_tag, ++ inode->i_ino); ++ } ++ return ret; ++} ++ + /* + * This does basic POSIX ACL permission checking + */ +@@ -285,10 +376,14 @@ int inode_permission(struct inode *inode + /* + * Nobody gets write access to an immutable file. + */ +- if (IS_IMMUTABLE(inode)) ++ if (IS_IMMUTABLE(inode) && !IS_COW(inode)) + return -EACCES; + } + ++ retval = dx_permission(inode, mask); ++ if (retval) ++ return retval; ++ + if (inode->i_op->permission) + retval = inode->i_op->permission(inode, mask, 0); + else +@@ -584,6 +679,9 @@ static inline int exec_permission(struct + int ret; + struct user_namespace *ns = inode_userns(inode); + ++ if (dx_barrier(inode)) ++ return -EACCES; ++ + if (inode->i_op->permission) { + ret = inode->i_op->permission(inode, MAY_EXEC, flags); + } else { +@@ -1095,7 +1193,8 @@ static void follow_dotdot(struct nameida + + if (nd->path.dentry == nd->root.dentry && + nd->path.mnt == nd->root.mnt) { +- break; ++ /* for sane '/' avoid follow_mount() */ ++ return; + } + if (nd->path.dentry != nd->path.mnt->mnt_root) { + /* rare case of legitimate dget_parent()... */ +@@ -1178,6 +1277,9 @@ static int do_lookup(struct nameidata *n + goto unlazy; + } + } ++ ++ /* FIXME: check dx permission */ ++ + path->mnt = mnt; + path->dentry = dentry; + if (unlikely(!__follow_mount_rcu(nd, path, inode))) +@@ -1226,6 +1328,8 @@ retry: + } + } + ++ /* FIXME: check dx permission */ ++ + path->mnt = mnt; + path->dentry = dentry; + err = follow_managed(path, nd->flags); +@@ -1853,7 +1957,7 @@ static int may_delete(struct inode *dir, + if (IS_APPEND(dir)) + return -EPERM; + if (check_sticky(dir, victim->d_inode)||IS_APPEND(victim->d_inode)|| +- IS_IMMUTABLE(victim->d_inode) || IS_SWAPFILE(victim->d_inode)) ++ IS_IXORUNLINK(victim->d_inode) || IS_SWAPFILE(victim->d_inode)) + return -EPERM; + if (isdir) { + if (!S_ISDIR(victim->d_inode->i_mode)) +@@ -1980,6 +2084,15 @@ static int may_open(struct path *path, i + break; + } + ++#ifdef CONFIG_VSERVER_COWBL ++ if (IS_COW(inode) && ++ ((flag & O_ACCMODE) != O_RDONLY)) { ++ if (IS_COW_LINK(inode)) ++ return -EMLINK; ++ inode->i_flags &= ~(S_IXUNLINK|S_IMMUTABLE); ++ mark_inode_dirty(inode); ++ } ++#endif + error = inode_permission(inode, acc_mode); + if (error) + return error; +@@ -2224,6 +2337,16 @@ ok: + } + common: + error = may_open(&nd->path, acc_mode, open_flag); ++#ifdef CONFIG_VSERVER_COWBL ++ if (error == -EMLINK) { ++ struct dentry *dentry; ++ dentry = cow_break_link(pathname); ++ if (IS_ERR(dentry)) ++ error = PTR_ERR(dentry); ++ else ++ dput(dentry); ++ } ++#endif + if (error) + goto exit; + filp = nameidata_to_filp(nd); +@@ -2266,6 +2389,7 @@ static struct file *path_openat(int dfd, + struct path path; + int error; + ++restart: + filp = get_empty_filp(); + if (!filp) + return ERR_PTR(-ENFILE); +@@ -2303,6 +2427,17 @@ static struct file *path_openat(int dfd, + filp = do_last(nd, &path, op, pathname); + put_link(nd, &link, cookie); + } ++ ++#ifdef CONFIG_VSERVER_COWBL ++ if (filp == ERR_PTR(-EMLINK)) { ++ if (nd->root.mnt && !(nd->flags & LOOKUP_ROOT)) ++ path_put(&nd->root); ++ if (base) ++ fput(base); ++ release_open_intent(nd); ++ goto restart; ++ } ++#endif + out: + if (nd->root.mnt && !(nd->flags & LOOKUP_ROOT)) + path_put(&nd->root); +@@ -2886,7 +3021,7 @@ int vfs_link(struct dentry *old_dentry, + /* + * A link to an append-only or immutable file cannot be created. + */ +- if (IS_APPEND(inode) || IS_IMMUTABLE(inode)) ++ if (IS_APPEND(inode) || IS_IXORUNLINK(inode)) + return -EPERM; + if (!dir->i_op->link) + return -EPERM; +@@ -3274,6 +3409,227 @@ int vfs_follow_link(struct nameidata *nd + return __vfs_follow_link(nd, link); + } + ++ ++#ifdef CONFIG_VSERVER_COWBL ++ ++static inline ++long do_cow_splice(struct file *in, struct file *out, size_t len) ++{ ++ loff_t ppos = 0; ++ ++ return do_splice_direct(in, &ppos, out, len, 0); ++} ++ ++struct dentry *cow_break_link(const char *pathname) ++{ ++ int ret, mode, pathlen, redo = 0; ++ struct nameidata old_nd, dir_nd; ++ struct path old_path; ++ struct dentry *dir, *old_dentry, *new_dentry = NULL; ++ struct file *old_file; ++ struct file *new_file; ++ char *to, *path, pad='\251'; ++ loff_t size; ++ ++ vxdprintk(VXD_CBIT(misc, 1), ++ "cow_break_link(" VS_Q("%s") ")", pathname); ++ path = kmalloc(PATH_MAX, GFP_KERNEL); ++ ret = -ENOMEM; ++ if (!path) ++ goto out; ++ ++ /* old_nd will have refs to dentry and mnt */ ++ ret = do_path_lookup(AT_FDCWD, pathname, LOOKUP_FOLLOW, &old_nd); ++ vxdprintk(VXD_CBIT(misc, 2), ++ "do_path_lookup(old): %d [r=%d]", ++ ret, mnt_get_count(old_nd.path.mnt)); ++ if (ret < 0) ++ goto out_free_path; ++ ++ old_path = old_nd.path; ++ old_dentry = old_path.dentry; ++ mode = old_dentry->d_inode->i_mode; ++ ++ to = d_path(&old_path, path, PATH_MAX-2); ++ pathlen = strlen(to); ++ vxdprintk(VXD_CBIT(misc, 2), ++ "old path " VS_Q("%s") " [%p:" VS_Q("%.*s") ":%d]", to, ++ old_dentry, ++ old_dentry->d_name.len, old_dentry->d_name.name, ++ old_dentry->d_name.len); ++ ++ to[pathlen + 1] = 0; ++retry: ++ new_dentry = NULL; ++ to[pathlen] = pad--; ++ ret = -ELOOP; ++ if (pad <= '\240') ++ goto out_rel_old; ++ ++ vxdprintk(VXD_CBIT(misc, 1), "temp copy " VS_Q("%s"), to); ++ /* dir_nd will have refs to dentry and mnt */ ++ ret = do_path_lookup(AT_FDCWD, to, ++ LOOKUP_PARENT | LOOKUP_OPEN | LOOKUP_CREATE, &dir_nd); ++ vxdprintk(VXD_CBIT(misc, 2), "do_path_lookup(new): %d", ret); ++ if (ret < 0) ++ goto retry; ++ ++ /* this puppy downs the inode mutex */ ++ new_dentry = lookup_create(&dir_nd, 0); ++ if (!new_dentry || IS_ERR(new_dentry)) { ++ path_put(&dir_nd.path); ++ vxdprintk(VXD_CBIT(misc, 2), ++ "lookup_create(new): failed with %ld", ++ PTR_ERR(new_dentry)); ++ mutex_unlock(&dir_nd.path.dentry->d_inode->i_mutex); ++ goto retry; ++ } ++ vxdprintk(VXD_CBIT(misc, 2), ++ "lookup_create(new): %p [" VS_Q("%.*s") ":%d]", ++ new_dentry, ++ new_dentry->d_name.len, new_dentry->d_name.name, ++ new_dentry->d_name.len); ++ ++ dir = dir_nd.path.dentry; ++ ++ ret = vfs_create(dir->d_inode, new_dentry, mode, &dir_nd); ++ vxdprintk(VXD_CBIT(misc, 2), ++ "vfs_create(new): %d", ret); ++ if (ret == -EEXIST) { ++ mutex_unlock(&dir->d_inode->i_mutex); ++ path_put(&dir_nd.path); ++ dput(new_dentry); ++ goto retry; ++ } ++ else if (ret < 0) ++ goto out_unlock_new; ++ ++ /* drop out early, ret passes ENOENT */ ++ ret = -ENOENT; ++ if ((redo = d_unhashed(old_dentry))) ++ goto out_unlock_new; ++ ++ path_get(&old_path); ++ /* this one cleans up the dentry/mnt in case of failure */ ++ old_file = dentry_open(old_dentry, old_path.mnt, ++ O_RDONLY, current_cred()); ++ vxdprintk(VXD_CBIT(misc, 2), ++ "dentry_open(old): %p", old_file); ++ if (IS_ERR(old_file)) { ++ ret = PTR_ERR(old_file); ++ goto out_unlock_new; ++ } ++ ++ dget(new_dentry); ++ mntget(old_path.mnt); ++ /* this one cleans up the dentry/mnt in case of failure */ ++ new_file = dentry_open(new_dentry, old_path.mnt, ++ O_WRONLY, current_cred()); ++ vxdprintk(VXD_CBIT(misc, 2), ++ "dentry_open(new): %p", new_file); ++ if (IS_ERR(new_file)) { ++ ret = PTR_ERR(new_file); ++ goto out_fput_old; ++ } ++ ++ size = i_size_read(old_file->f_dentry->d_inode); ++ ret = do_cow_splice(old_file, new_file, size); ++ vxdprintk(VXD_CBIT(misc, 2), "do_splice_direct: %d", ret); ++ if (ret < 0) { ++ goto out_fput_both; ++ } else if (ret < size) { ++ ret = -ENOSPC; ++ goto out_fput_both; ++ } else { ++ struct inode *old_inode = old_dentry->d_inode; ++ struct inode *new_inode = new_dentry->d_inode; ++ struct iattr attr = { ++ .ia_uid = old_inode->i_uid, ++ .ia_gid = old_inode->i_gid, ++ .ia_valid = ATTR_UID | ATTR_GID ++ }; ++ ++ setattr_copy(new_inode, &attr); ++ mark_inode_dirty(new_inode); ++ } ++ ++ mutex_lock(&old_dentry->d_inode->i_sb->s_vfs_rename_mutex); ++ ++ /* drop out late */ ++ ret = -ENOENT; ++ if ((redo = d_unhashed(old_dentry))) ++ goto out_unlock; ++ ++ vxdprintk(VXD_CBIT(misc, 2), ++ "vfs_rename: [" VS_Q("%*s") ":%d] -> [" VS_Q("%*s") ":%d]", ++ new_dentry->d_name.len, new_dentry->d_name.name, ++ new_dentry->d_name.len, ++ old_dentry->d_name.len, old_dentry->d_name.name, ++ old_dentry->d_name.len); ++ ret = vfs_rename(dir_nd.path.dentry->d_inode, new_dentry, ++ old_dentry->d_parent->d_inode, old_dentry); ++ vxdprintk(VXD_CBIT(misc, 2), "vfs_rename: %d", ret); ++ ++out_unlock: ++ mutex_unlock(&old_dentry->d_inode->i_sb->s_vfs_rename_mutex); ++ ++out_fput_both: ++ vxdprintk(VXD_CBIT(misc, 3), ++ "fput(new_file=%p[#%ld])", new_file, ++ atomic_long_read(&new_file->f_count)); ++ fput(new_file); ++ ++out_fput_old: ++ vxdprintk(VXD_CBIT(misc, 3), ++ "fput(old_file=%p[#%ld])", old_file, ++ atomic_long_read(&old_file->f_count)); ++ fput(old_file); ++ ++out_unlock_new: ++ mutex_unlock(&dir->d_inode->i_mutex); ++ if (!ret) ++ goto out_redo; ++ ++ /* error path cleanup */ ++ vfs_unlink(dir->d_inode, new_dentry); ++ ++out_redo: ++ if (!redo) ++ goto out_rel_both; ++ /* lookup dentry once again */ ++ /* old_nd.path is freed as old_path in out_rel_old */ ++ ret = do_path_lookup(AT_FDCWD, pathname, LOOKUP_FOLLOW, &old_nd); ++ if (ret) ++ goto out_rel_both; ++ ++ dput(new_dentry); ++ new_dentry = old_nd.path.dentry; ++ vxdprintk(VXD_CBIT(misc, 2), ++ "do_path_lookup(redo): %p [" VS_Q("%.*s") ":%d]", ++ new_dentry, ++ new_dentry->d_name.len, new_dentry->d_name.name, ++ new_dentry->d_name.len); ++ dget(new_dentry); ++ ++out_rel_both: ++ path_put(&dir_nd.path); ++out_rel_old: ++ path_put(&old_path); ++out_free_path: ++ kfree(path); ++out: ++ if (ret) { ++ dput(new_dentry); ++ new_dentry = ERR_PTR(ret); ++ } ++ vxdprintk(VXD_CBIT(misc, 3), ++ "cow_break_link returning with %p [r=%d]", ++ new_dentry, mnt_get_count(old_nd.path.mnt)); ++ return new_dentry; ++} ++ ++#endif ++ + /* get the link contents into pagecache */ + static char *page_getlink(struct dentry * dentry, struct page **ppage) + { +diff -NurpP --minimal linux-3.0.9/fs/namespace.c linux-3.0.9-vs2.3.2.1/fs/namespace.c +--- linux-3.0.9/fs/namespace.c 2011-11-15 16:40:47.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/fs/namespace.c 2011-11-15 17:37:39.000000000 +0100 +@@ -31,6 +31,11 @@ + #include <linux/idr.h> + #include <linux/fs_struct.h> + #include <linux/fsnotify.h> ++#include <linux/vs_base.h> ++#include <linux/vs_context.h> ++#include <linux/vs_tag.h> ++#include <linux/vserver/space.h> ++#include <linux/vserver/global.h> + #include <asm/uaccess.h> + #include <asm/unistd.h> + #include "pnode.h" +@@ -679,6 +684,10 @@ vfs_kern_mount(struct file_system_type * + if (!type) + return ERR_PTR(-ENODEV); + ++ if ((type->fs_flags & FS_BINARY_MOUNTDATA) && ++ !vx_capable(CAP_SYS_ADMIN, VXC_BINARY_MOUNT)) ++ return ERR_PTR(-EPERM); ++ + mnt = alloc_vfsmnt(name); + if (!mnt) + return ERR_PTR(-ENOMEM); +@@ -724,6 +733,7 @@ static struct vfsmount *clone_mnt(struct + mnt->mnt_root = dget(root); + mnt->mnt_mountpoint = mnt->mnt_root; + mnt->mnt_parent = mnt; ++ mnt->mnt_tag = old->mnt_tag; + + if (flag & CL_SLAVE) { + list_add(&mnt->mnt_slave, &old->mnt_slave_list); +@@ -852,6 +862,31 @@ static inline void mangle(struct seq_fil + seq_escape(m, s, " \t\n\\"); + } + ++static int mnt_is_reachable(struct vfsmount *mnt) ++{ ++ struct path root; ++ struct dentry *point; ++ int ret; ++ ++ if (mnt == mnt->mnt_ns->root) ++ return 1; ++ ++ br_read_lock(vfsmount_lock); ++ root = current->fs->root; ++ point = root.dentry; ++ ++ while ((mnt != mnt->mnt_parent) && (mnt != root.mnt)) { ++ point = mnt->mnt_mountpoint; ++ mnt = mnt->mnt_parent; ++ } ++ ++ ret = (mnt == root.mnt) && is_subdir(point, root.dentry); ++ ++ br_read_unlock(vfsmount_lock); ++ ++ return ret; ++} ++ + /* + * Simple .show_options callback for filesystems which don't want to + * implement more complex mount option showing. +@@ -954,6 +989,8 @@ static int show_sb_opts(struct seq_file + { MS_SYNCHRONOUS, ",sync" }, + { MS_DIRSYNC, ",dirsync" }, + { MS_MANDLOCK, ",mand" }, ++ { MS_TAGGED, ",tag" }, ++ { MS_NOTAGCHECK, ",notagcheck" }, + { 0, NULL } + }; + const struct proc_fs_info *fs_infop; +@@ -1000,16 +1037,26 @@ static int show_vfsmnt(struct seq_file * + int err = 0; + struct path mnt_path = { .dentry = mnt->mnt_root, .mnt = mnt }; + +- if (mnt->mnt_sb->s_op->show_devname) { +- err = mnt->mnt_sb->s_op->show_devname(m, mnt); +- if (err) +- goto out; ++ if (vx_flags(VXF_HIDE_MOUNT, 0)) ++ return SEQ_SKIP; ++ if (!mnt_is_reachable(mnt) && !vx_check(0, VS_WATCH_P)) ++ return SEQ_SKIP; ++ ++ if (!vx_check(0, VS_ADMIN|VS_WATCH) && ++ mnt == current->fs->root.mnt) { ++ seq_puts(m, "/dev/root / "); + } else { +- mangle(m, mnt->mnt_devname ? mnt->mnt_devname : "none"); ++ if (mnt->mnt_sb->s_op->show_devname) { ++ err = mnt->mnt_sb->s_op->show_devname(m, mnt); ++ if (err) ++ goto out; ++ } else { ++ mangle(m, mnt->mnt_devname ? mnt->mnt_devname : "none"); ++ } ++ seq_putc(m, ' '); ++ seq_path(m, &mnt_path, " \t\n\\"); ++ seq_putc(m, ' '); + } +- seq_putc(m, ' '); +- seq_path(m, &mnt_path, " \t\n\\"); +- seq_putc(m, ' '); + show_type(m, mnt->mnt_sb); + seq_puts(m, __mnt_is_readonly(mnt) ? " ro" : " rw"); + err = show_sb_opts(m, mnt->mnt_sb); +@@ -1039,6 +1086,11 @@ static int show_mountinfo(struct seq_fil + struct path root = p->root; + int err = 0; + ++ if (vx_flags(VXF_HIDE_MOUNT, 0)) ++ return SEQ_SKIP; ++ if (!mnt_is_reachable(mnt) && !vx_check(0, VS_WATCH_P)) ++ return SEQ_SKIP; ++ + seq_printf(m, "%i %i %u:%u ", mnt->mnt_id, mnt->mnt_parent->mnt_id, + MAJOR(sb->s_dev), MINOR(sb->s_dev)); + if (sb->s_op->show_path) +@@ -1107,22 +1159,32 @@ static int show_vfsstat(struct seq_file + struct path mnt_path = { .dentry = mnt->mnt_root, .mnt = mnt }; + int err = 0; + +- /* device */ +- if (mnt->mnt_sb->s_op->show_devname) { +- seq_puts(m, "device "); +- err = mnt->mnt_sb->s_op->show_devname(m, mnt); ++ if (vx_flags(VXF_HIDE_MOUNT, 0)) ++ return SEQ_SKIP; ++ if (!mnt_is_reachable(mnt) && !vx_check(0, VS_WATCH_P)) ++ return SEQ_SKIP; ++ ++ if (!vx_check(0, VS_ADMIN|VS_WATCH) && ++ mnt == current->fs->root.mnt) { ++ seq_puts(m, "device /dev/root mounted on / "); + } else { +- if (mnt->mnt_devname) { ++ /* device */ ++ if (mnt->mnt_sb->s_op->show_devname) { + seq_puts(m, "device "); +- mangle(m, mnt->mnt_devname); +- } else +- seq_puts(m, "no device"); +- } ++ err = mnt->mnt_sb->s_op->show_devname(m, mnt); ++ } else { ++ if (mnt->mnt_devname) { ++ seq_puts(m, "device "); ++ mangle(m, mnt->mnt_devname); ++ } else ++ seq_puts(m, "no device"); ++ } + +- /* mount point */ +- seq_puts(m, " mounted on "); +- seq_path(m, &mnt_path, " \t\n\\"); +- seq_putc(m, ' '); ++ /* mount point */ ++ seq_puts(m, " mounted on "); ++ seq_path(m, &mnt_path, " \t\n\\"); ++ seq_putc(m, ' '); ++ } + + /* file system type */ + seq_puts(m, "with fstype "); +@@ -1381,7 +1443,7 @@ SYSCALL_DEFINE2(umount, char __user *, n + goto dput_and_out; + + retval = -EPERM; +- if (!capable(CAP_SYS_ADMIN)) ++ if (!vx_capable(CAP_SYS_ADMIN, VXC_SECURE_MOUNT)) + goto dput_and_out; + + retval = do_umount(path.mnt, flags); +@@ -1407,7 +1469,7 @@ SYSCALL_DEFINE1(oldumount, char __user * + + static int mount_is_safe(struct path *path) + { +- if (capable(CAP_SYS_ADMIN)) ++ if (vx_capable(CAP_SYS_ADMIN, VXC_SECURE_MOUNT)) + return 0; + return -EPERM; + #ifdef notyet +@@ -1717,7 +1779,7 @@ static int do_change_type(struct path *p + int type; + int err = 0; + +- if (!capable(CAP_SYS_ADMIN)) ++ if (!vx_capable(CAP_SYS_ADMIN, VXC_NAMESPACE)) + return -EPERM; + + if (path->dentry != path->mnt->mnt_root) +@@ -1733,6 +1795,7 @@ static int do_change_type(struct path *p + if (err) + goto out_unlock; + } ++ // mnt->mnt_flags = mnt_flags; + + br_write_lock(vfsmount_lock); + for (m = mnt; m; m = (recurse ? next_mnt(m, mnt) : NULL)) +@@ -1748,12 +1811,14 @@ static int do_change_type(struct path *p + * do loopback mount. + */ + static int do_loopback(struct path *path, char *old_name, +- int recurse) ++ tag_t tag, unsigned long flags, int mnt_flags) + { + LIST_HEAD(umount_list); + struct path old_path; + struct vfsmount *mnt = NULL; + int err = mount_is_safe(path); ++ int recurse = flags & MS_REC; ++ + if (err) + return err; + if (!old_name || !*old_name) +@@ -1819,12 +1884,12 @@ static int change_mount_flags(struct vfs + * on it - tough luck. + */ + static int do_remount(struct path *path, int flags, int mnt_flags, +- void *data) ++ void *data, xid_t xid) + { + int err; + struct super_block *sb = path->mnt->mnt_sb; + +- if (!capable(CAP_SYS_ADMIN)) ++ if (!vx_capable(CAP_SYS_ADMIN, VXC_SECURE_REMOUNT)) + return -EPERM; + + if (!check_mnt(path->mnt)) +@@ -1872,7 +1937,7 @@ static int do_move_mount(struct path *pa + struct path old_path, parent_path; + struct vfsmount *p; + int err = 0; +- if (!capable(CAP_SYS_ADMIN)) ++ if (!vx_capable(CAP_SYS_ADMIN, VXC_SECURE_MOUNT)) + return -EPERM; + if (!old_name || !*old_name) + return -EINVAL; +@@ -2023,7 +2088,7 @@ static int do_new_mount(struct path *pat + return -EINVAL; + + /* we need capabilities... */ +- if (!capable(CAP_SYS_ADMIN)) ++ if (!vx_capable(CAP_SYS_ADMIN, VXC_SECURE_MOUNT)) + return -EPERM; + + mnt = do_kern_mount(type, flags, name, data); +@@ -2292,6 +2357,7 @@ long do_mount(char *dev_name, char *dir_ + struct path path; + int retval = 0; + int mnt_flags = 0; ++ tag_t tag = 0; + + /* Discard magic */ + if ((flags & MS_MGC_MSK) == MS_MGC_VAL) +@@ -2319,6 +2385,12 @@ long do_mount(char *dev_name, char *dir_ + if (!(flags & MS_NOATIME)) + mnt_flags |= MNT_RELATIME; + ++ if (dx_parse_tag(data_page, &tag, 1, &mnt_flags, &flags)) { ++ /* FIXME: bind and re-mounts get the tag flag? */ ++ if (flags & (MS_BIND|MS_REMOUNT)) ++ flags |= MS_TAGID; ++ } ++ + /* Separate the per-mountpoint flags */ + if (flags & MS_NOSUID) + mnt_flags |= MNT_NOSUID; +@@ -2335,15 +2407,17 @@ long do_mount(char *dev_name, char *dir_ + if (flags & MS_RDONLY) + mnt_flags |= MNT_READONLY; + ++ if (!capable(CAP_SYS_ADMIN)) ++ mnt_flags |= MNT_NODEV; + flags &= ~(MS_NOSUID | MS_NOEXEC | MS_NODEV | MS_ACTIVE | MS_BORN | + MS_NOATIME | MS_NODIRATIME | MS_RELATIME| MS_KERNMOUNT | + MS_STRICTATIME); + + if (flags & MS_REMOUNT) + retval = do_remount(&path, flags & ~MS_REMOUNT, mnt_flags, +- data_page); ++ data_page, tag); + else if (flags & MS_BIND) +- retval = do_loopback(&path, dev_name, flags & MS_REC); ++ retval = do_loopback(&path, dev_name, tag, flags, mnt_flags); + else if (flags & (MS_SHARED | MS_PRIVATE | MS_SLAVE | MS_UNBINDABLE)) + retval = do_change_type(&path, flags); + else if (flags & MS_MOVE) +@@ -2443,6 +2517,7 @@ static struct mnt_namespace *dup_mnt_ns( + q = next_mnt(q, new_ns->root); + } + up_write(&namespace_sem); ++ atomic_inc(&vs_global_mnt_ns); + + if (rootmnt) + mntput(rootmnt); +@@ -2582,9 +2657,10 @@ SYSCALL_DEFINE2(pivot_root, const char _ + goto out3; + + error = -EINVAL; +- if (IS_MNT_SHARED(old.mnt) || +- IS_MNT_SHARED(new.mnt->mnt_parent) || +- IS_MNT_SHARED(root.mnt->mnt_parent)) ++ if ((IS_MNT_SHARED(old.mnt) || ++ IS_MNT_SHARED(new.mnt->mnt_parent) || ++ IS_MNT_SHARED(root.mnt->mnt_parent)) && ++ !vx_flags(VXF_STATE_SETUP, 0)) + goto out4; + if (!check_mnt(root.mnt) || !check_mnt(new.mnt)) + goto out4; +@@ -2716,6 +2792,7 @@ void put_mnt_ns(struct mnt_namespace *ns + br_write_unlock(vfsmount_lock); + up_write(&namespace_sem); + release_mounts(&umount_list); ++ atomic_dec(&vs_global_mnt_ns); + kfree(ns); + } + EXPORT_SYMBOL(put_mnt_ns); +diff -NurpP --minimal linux-3.0.9/fs/nfs/client.c linux-3.0.9-vs2.3.2.1/fs/nfs/client.c +--- linux-3.0.9/fs/nfs/client.c 2011-07-22 11:18:05.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/fs/nfs/client.c 2011-06-10 22:11:24.000000000 +0200 +@@ -780,6 +780,9 @@ static int nfs_init_server_rpcclient(str + if (server->flags & NFS_MOUNT_SOFT) + server->client->cl_softrtry = 1; + ++ server->client->cl_tag = 0; ++ if (server->flags & NFS_MOUNT_TAGGED) ++ server->client->cl_tag = 1; + return 0; + } + +@@ -951,6 +954,10 @@ static void nfs_server_set_fsinfo(struct + server->acdirmin = server->acdirmax = 0; + } + ++ /* FIXME: needs fsinfo ++ if (server->flags & NFS_MOUNT_TAGGED) ++ sb->s_flags |= MS_TAGGED; */ ++ + server->maxfilesize = fsinfo->maxfilesize; + + server->time_delta = fsinfo->time_delta; +diff -NurpP --minimal linux-3.0.9/fs/nfs/dir.c linux-3.0.9-vs2.3.2.1/fs/nfs/dir.c +--- linux-3.0.9/fs/nfs/dir.c 2011-11-15 16:40:47.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/fs/nfs/dir.c 2011-08-08 23:04:47.000000000 +0200 +@@ -35,6 +35,7 @@ + #include <linux/sched.h> + #include <linux/kmemleak.h> + #include <linux/xattr.h> ++#include <linux/vs_tag.h> + + #include "delegation.h" + #include "iostat.h" +@@ -1308,6 +1309,7 @@ static struct dentry *nfs_lookup(struct + if (IS_ERR(res)) + goto out_unblock_sillyrename; + ++ dx_propagate_tag(nd, inode); + no_entry: + res = d_materialise_unique(dentry, inode); + if (res != NULL) { +diff -NurpP --minimal linux-3.0.9/fs/nfs/inode.c linux-3.0.9-vs2.3.2.1/fs/nfs/inode.c +--- linux-3.0.9/fs/nfs/inode.c 2011-07-22 11:18:05.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/fs/nfs/inode.c 2011-10-11 21:21:33.000000000 +0200 +@@ -38,6 +38,7 @@ + #include <linux/nfs_xdr.h> + #include <linux/slab.h> + #include <linux/compat.h> ++#include <linux/vs_tag.h> + + #include <asm/system.h> + #include <asm/uaccess.h> +@@ -273,6 +274,8 @@ nfs_fhget(struct super_block *sb, struct + if (inode->i_state & I_NEW) { + struct nfs_inode *nfsi = NFS_I(inode); + unsigned long now = jiffies; ++ uid_t uid; ++ gid_t gid; + + /* We set i_ino for the few things that still rely on it, + * such as stat(2) */ +@@ -321,8 +324,8 @@ nfs_fhget(struct super_block *sb, struct + nfsi->change_attr = 0; + inode->i_size = 0; + inode->i_nlink = 0; +- inode->i_uid = -2; +- inode->i_gid = -2; ++ uid = -2; ++ gid = -2; + inode->i_blocks = 0; + memset(nfsi->cookieverf, 0, sizeof(nfsi->cookieverf)); + +@@ -359,13 +362,13 @@ nfs_fhget(struct super_block *sb, struct + else if (nfs_server_capable(inode, NFS_CAP_NLINK)) + nfsi->cache_validity |= NFS_INO_INVALID_ATTR; + if (fattr->valid & NFS_ATTR_FATTR_OWNER) +- inode->i_uid = fattr->uid; ++ uid = fattr->uid; + else if (nfs_server_capable(inode, NFS_CAP_OWNER)) + nfsi->cache_validity |= NFS_INO_INVALID_ATTR + | NFS_INO_INVALID_ACCESS + | NFS_INO_INVALID_ACL; + if (fattr->valid & NFS_ATTR_FATTR_GROUP) +- inode->i_gid = fattr->gid; ++ gid = fattr->gid; + else if (nfs_server_capable(inode, NFS_CAP_OWNER_GROUP)) + nfsi->cache_validity |= NFS_INO_INVALID_ATTR + | NFS_INO_INVALID_ACCESS +@@ -378,6 +381,11 @@ nfs_fhget(struct super_block *sb, struct + */ + inode->i_blocks = nfs_calc_block_size(fattr->du.nfs3.used); + } ++ inode->i_uid = INOTAG_UID(DX_TAG(inode), uid, gid); ++ inode->i_gid = INOTAG_GID(DX_TAG(inode), uid, gid); ++ inode->i_tag = INOTAG_TAG(DX_TAG(inode), uid, gid, 0); ++ /* maybe fattr->xid someday */ ++ + nfsi->attrtimeo = NFS_MINATTRTIMEO(inode); + nfsi->attrtimeo_timestamp = now; + nfsi->access_cache = RB_ROOT; +@@ -494,6 +502,8 @@ void nfs_setattr_update_inode(struct ino + inode->i_uid = attr->ia_uid; + if ((attr->ia_valid & ATTR_GID) != 0) + inode->i_gid = attr->ia_gid; ++ if ((attr->ia_valid & ATTR_TAG) && IS_TAGGED(inode)) ++ inode->i_tag = attr->ia_tag; + NFS_I(inode)->cache_validity |= NFS_INO_INVALID_ACCESS|NFS_INO_INVALID_ACL; + spin_unlock(&inode->i_lock); + } +@@ -941,6 +951,9 @@ static int nfs_check_inode_attributes(st + struct nfs_inode *nfsi = NFS_I(inode); + loff_t cur_size, new_isize; + unsigned long invalid = 0; ++ uid_t uid; ++ gid_t gid; ++ tag_t tag; + + + /* Has the inode gone and changed behind our back? */ +@@ -964,13 +977,18 @@ static int nfs_check_inode_attributes(st + invalid |= NFS_INO_INVALID_ATTR|NFS_INO_REVAL_PAGECACHE; + } + ++ uid = INOTAG_UID(DX_TAG(inode), fattr->uid, fattr->gid); ++ gid = INOTAG_GID(DX_TAG(inode), fattr->uid, fattr->gid); ++ tag = INOTAG_TAG(DX_TAG(inode), fattr->uid, fattr->gid, 0); ++ + /* Have any file permissions changed? */ + if ((fattr->valid & NFS_ATTR_FATTR_MODE) && (inode->i_mode & S_IALLUGO) != (fattr->mode & S_IALLUGO)) + invalid |= NFS_INO_INVALID_ATTR | NFS_INO_INVALID_ACCESS | NFS_INO_INVALID_ACL; +- if ((fattr->valid & NFS_ATTR_FATTR_OWNER) && inode->i_uid != fattr->uid) ++ if ((fattr->valid & NFS_ATTR_FATTR_OWNER) && uid != fattr->uid) + invalid |= NFS_INO_INVALID_ATTR | NFS_INO_INVALID_ACCESS | NFS_INO_INVALID_ACL; +- if ((fattr->valid & NFS_ATTR_FATTR_GROUP) && inode->i_gid != fattr->gid) ++ if ((fattr->valid & NFS_ATTR_FATTR_GROUP) && gid != fattr->gid) + invalid |= NFS_INO_INVALID_ATTR | NFS_INO_INVALID_ACCESS | NFS_INO_INVALID_ACL; ++ /* maybe check for tag too? */ + + /* Has the link count changed? */ + if ((fattr->valid & NFS_ATTR_FATTR_NLINK) && inode->i_nlink != fattr->nlink) +@@ -1205,6 +1223,9 @@ static int nfs_update_inode(struct inode + unsigned long invalid = 0; + unsigned long now = jiffies; + unsigned long save_cache_validity; ++ uid_t uid; ++ gid_t gid; ++ tag_t tag; + + dfprintk(VFS, "NFS: %s(%s/%ld ct=%d info=0x%x)\n", + __func__, inode->i_sb->s_id, inode->i_ino, +@@ -1312,6 +1333,9 @@ static int nfs_update_inode(struct inode + | NFS_INO_REVAL_PAGECACHE + | NFS_INO_REVAL_FORCED); + ++ uid = TAGINO_UID(DX_TAG(inode), inode->i_uid, inode->i_tag); ++ gid = TAGINO_GID(DX_TAG(inode), inode->i_gid, inode->i_tag); ++ tag = inode->i_tag; + + if (fattr->valid & NFS_ATTR_FATTR_ATIME) + memcpy(&inode->i_atime, &fattr->atime, sizeof(inode->i_atime)); +@@ -1333,9 +1357,9 @@ static int nfs_update_inode(struct inode + | NFS_INO_REVAL_FORCED); + + if (fattr->valid & NFS_ATTR_FATTR_OWNER) { +- if (inode->i_uid != fattr->uid) { ++ if (uid != fattr->uid) { + invalid |= NFS_INO_INVALID_ATTR|NFS_INO_INVALID_ACCESS|NFS_INO_INVALID_ACL; +- inode->i_uid = fattr->uid; ++ uid = fattr->uid; + } + } else if (server->caps & NFS_CAP_OWNER) + invalid |= save_cache_validity & (NFS_INO_INVALID_ATTR +@@ -1344,9 +1368,9 @@ static int nfs_update_inode(struct inode + | NFS_INO_REVAL_FORCED); + + if (fattr->valid & NFS_ATTR_FATTR_GROUP) { +- if (inode->i_gid != fattr->gid) { ++ if (gid != fattr->gid) { + invalid |= NFS_INO_INVALID_ATTR|NFS_INO_INVALID_ACCESS|NFS_INO_INVALID_ACL; +- inode->i_gid = fattr->gid; ++ gid = fattr->gid; + } + } else if (server->caps & NFS_CAP_OWNER_GROUP) + invalid |= save_cache_validity & (NFS_INO_INVALID_ATTR +@@ -1354,6 +1378,10 @@ static int nfs_update_inode(struct inode + | NFS_INO_INVALID_ACL + | NFS_INO_REVAL_FORCED); + ++ inode->i_uid = INOTAG_UID(DX_TAG(inode), uid, gid); ++ inode->i_gid = INOTAG_GID(DX_TAG(inode), uid, gid); ++ inode->i_tag = INOTAG_TAG(DX_TAG(inode), uid, gid, tag); ++ + if (fattr->valid & NFS_ATTR_FATTR_NLINK) { + if (inode->i_nlink != fattr->nlink) { + invalid |= NFS_INO_INVALID_ATTR; +diff -NurpP --minimal linux-3.0.9/fs/nfs/nfs3xdr.c linux-3.0.9-vs2.3.2.1/fs/nfs/nfs3xdr.c +--- linux-3.0.9/fs/nfs/nfs3xdr.c 2011-03-15 18:07:32.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/fs/nfs/nfs3xdr.c 2011-06-10 22:11:24.000000000 +0200 +@@ -20,6 +20,7 @@ + #include <linux/nfs3.h> + #include <linux/nfs_fs.h> + #include <linux/nfsacl.h> ++#include <linux/vs_tag.h> + #include "internal.h" + + #define NFSDBG_FACILITY NFSDBG_XDR +@@ -562,7 +563,8 @@ static __be32 *xdr_decode_nfstime3(__be3 + * set_mtime mtime; + * }; + */ +-static void encode_sattr3(struct xdr_stream *xdr, const struct iattr *attr) ++static void encode_sattr3(struct xdr_stream *xdr, ++ const struct iattr *attr, int tag) + { + u32 nbytes; + __be32 *p; +@@ -594,15 +596,19 @@ static void encode_sattr3(struct xdr_str + } else + *p++ = xdr_zero; + +- if (attr->ia_valid & ATTR_UID) { ++ if (attr->ia_valid & ATTR_UID || ++ (tag && (attr->ia_valid & ATTR_TAG))) { + *p++ = xdr_one; +- *p++ = cpu_to_be32(attr->ia_uid); ++ *p++ = cpu_to_be32(TAGINO_UID(tag, ++ attr->ia_uid, attr->ia_tag)); + } else + *p++ = xdr_zero; + +- if (attr->ia_valid & ATTR_GID) { ++ if (attr->ia_valid & ATTR_GID || ++ (tag && (attr->ia_valid & ATTR_TAG))) { + *p++ = xdr_one; +- *p++ = cpu_to_be32(attr->ia_gid); ++ *p++ = cpu_to_be32(TAGINO_GID(tag, ++ attr->ia_gid, attr->ia_tag)); + } else + *p++ = xdr_zero; + +@@ -878,7 +884,7 @@ static void nfs3_xdr_enc_setattr3args(st + const struct nfs3_sattrargs *args) + { + encode_nfs_fh3(xdr, args->fh); +- encode_sattr3(xdr, args->sattr); ++ encode_sattr3(xdr, args->sattr, req->rq_task->tk_client->cl_tag); + encode_sattrguard3(xdr, args); + } + +@@ -1028,13 +1034,13 @@ static void nfs3_xdr_enc_write3args(stru + * }; + */ + static void encode_createhow3(struct xdr_stream *xdr, +- const struct nfs3_createargs *args) ++ const struct nfs3_createargs *args, int tag) + { + encode_uint32(xdr, args->createmode); + switch (args->createmode) { + case NFS3_CREATE_UNCHECKED: + case NFS3_CREATE_GUARDED: +- encode_sattr3(xdr, args->sattr); ++ encode_sattr3(xdr, args->sattr, tag); + break; + case NFS3_CREATE_EXCLUSIVE: + encode_createverf3(xdr, args->verifier); +@@ -1049,7 +1055,7 @@ static void nfs3_xdr_enc_create3args(str + const struct nfs3_createargs *args) + { + encode_diropargs3(xdr, args->fh, args->name, args->len); +- encode_createhow3(xdr, args); ++ encode_createhow3(xdr, args, req->rq_task->tk_client->cl_tag); + } + + /* +@@ -1065,7 +1071,7 @@ static void nfs3_xdr_enc_mkdir3args(stru + const struct nfs3_mkdirargs *args) + { + encode_diropargs3(xdr, args->fh, args->name, args->len); +- encode_sattr3(xdr, args->sattr); ++ encode_sattr3(xdr, args->sattr, req->rq_task->tk_client->cl_tag); + } + + /* +@@ -1082,9 +1088,9 @@ static void nfs3_xdr_enc_mkdir3args(stru + * }; + */ + static void encode_symlinkdata3(struct xdr_stream *xdr, +- const struct nfs3_symlinkargs *args) ++ const struct nfs3_symlinkargs *args, int tag) + { +- encode_sattr3(xdr, args->sattr); ++ encode_sattr3(xdr, args->sattr, tag); + encode_nfspath3(xdr, args->pages, args->pathlen); + } + +@@ -1093,7 +1099,7 @@ static void nfs3_xdr_enc_symlink3args(st + const struct nfs3_symlinkargs *args) + { + encode_diropargs3(xdr, args->fromfh, args->fromname, args->fromlen); +- encode_symlinkdata3(xdr, args); ++ encode_symlinkdata3(xdr, args, req->rq_task->tk_client->cl_tag); + } + + /* +@@ -1121,24 +1127,24 @@ static void nfs3_xdr_enc_symlink3args(st + * }; + */ + static void encode_devicedata3(struct xdr_stream *xdr, +- const struct nfs3_mknodargs *args) ++ const struct nfs3_mknodargs *args, int tag) + { +- encode_sattr3(xdr, args->sattr); ++ encode_sattr3(xdr, args->sattr, tag); + encode_specdata3(xdr, args->rdev); + } + + static void encode_mknoddata3(struct xdr_stream *xdr, +- const struct nfs3_mknodargs *args) ++ const struct nfs3_mknodargs *args, int tag) + { + encode_ftype3(xdr, args->type); + switch (args->type) { + case NF3CHR: + case NF3BLK: +- encode_devicedata3(xdr, args); ++ encode_devicedata3(xdr, args, tag); + break; + case NF3SOCK: + case NF3FIFO: +- encode_sattr3(xdr, args->sattr); ++ encode_sattr3(xdr, args->sattr, tag); + break; + case NF3REG: + case NF3DIR: +@@ -1153,7 +1159,7 @@ static void nfs3_xdr_enc_mknod3args(stru + const struct nfs3_mknodargs *args) + { + encode_diropargs3(xdr, args->fh, args->name, args->len); +- encode_mknoddata3(xdr, args); ++ encode_mknoddata3(xdr, args, req->rq_task->tk_client->cl_tag); + } + + /* +diff -NurpP --minimal linux-3.0.9/fs/nfs/super.c linux-3.0.9-vs2.3.2.1/fs/nfs/super.c +--- linux-3.0.9/fs/nfs/super.c 2011-11-15 16:40:47.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/fs/nfs/super.c 2011-11-15 17:37:07.000000000 +0100 +@@ -53,6 +53,7 @@ + #include <linux/nfs_xdr.h> + #include <linux/magic.h> + #include <linux/parser.h> ++#include <linux/vs_tag.h> + + #include <asm/system.h> + #include <asm/uaccess.h> +@@ -87,6 +88,7 @@ enum { + Opt_sharecache, Opt_nosharecache, + Opt_resvport, Opt_noresvport, + Opt_fscache, Opt_nofscache, ++ Opt_tag, Opt_notag, + + /* Mount options that take integer arguments */ + Opt_port, +@@ -100,6 +102,7 @@ enum { + Opt_mountvers, + Opt_nfsvers, + Opt_minorversion, ++ Opt_tagid, + + /* Mount options that take string arguments */ + Opt_sec, Opt_proto, Opt_mountproto, Opt_mounthost, +@@ -180,6 +183,10 @@ static const match_table_t nfs_mount_opt + { Opt_fscache_uniq, "fsc=%s" }, + { Opt_local_lock, "local_lock=%s" }, + ++ { Opt_tag, "tag" }, ++ { Opt_notag, "notag" }, ++ { Opt_tagid, "tagid=%u" }, ++ + { Opt_err, NULL } + }; + +@@ -650,6 +657,7 @@ static void nfs_show_mount_options(struc + { NFS_MOUNT_NORDIRPLUS, ",nordirplus", "" }, + { NFS_MOUNT_UNSHARED, ",nosharecache", "" }, + { NFS_MOUNT_NORESVPORT, ",noresvport", "" }, ++ { NFS_MOUNT_TAGGED, ",tag", "" }, + { 0, NULL, NULL } + }; + const struct proc_nfs_info *nfs_infop; +@@ -1198,6 +1206,14 @@ static int nfs_parse_mount_options(char + kfree(mnt->fscache_uniq); + mnt->fscache_uniq = NULL; + break; ++#ifndef CONFIG_TAGGING_NONE ++ case Opt_tag: ++ mnt->flags |= NFS_MOUNT_TAGGED; ++ break; ++ case Opt_notag: ++ mnt->flags &= ~NFS_MOUNT_TAGGED; ++ break; ++#endif + + /* + * options that take numeric values +@@ -1304,6 +1320,12 @@ static int nfs_parse_mount_options(char + goto out_invalid_value; + mnt->minorversion = option; + break; ++#ifdef CONFIG_PROPAGATE ++ case Opt_tagid: ++ /* use args[0] */ ++ nfs_data.flags |= NFS_MOUNT_TAGGED; ++ break; ++#endif + + /* + * options that take text values +diff -NurpP --minimal linux-3.0.9/fs/nfsd/auth.c linux-3.0.9-vs2.3.2.1/fs/nfsd/auth.c +--- linux-3.0.9/fs/nfsd/auth.c 2010-02-25 11:52:05.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/fs/nfsd/auth.c 2011-06-10 22:11:24.000000000 +0200 +@@ -1,6 +1,7 @@ + /* Copyright (C) 1995, 1996 Olaf Kirch <okir@monad.swb.de> */ + + #include <linux/sched.h> ++#include <linux/vs_tag.h> + #include "nfsd.h" + #include "auth.h" + +@@ -36,6 +37,9 @@ int nfsd_setuser(struct svc_rqst *rqstp, + + new->fsuid = rqstp->rq_cred.cr_uid; + new->fsgid = rqstp->rq_cred.cr_gid; ++ /* FIXME: this desperately needs a tag :) ++ new->xid = (xid_t)INOTAG_TAG(DX_TAG_NFSD, cred.cr_uid, cred.cr_gid, 0); ++ */ + + rqgi = rqstp->rq_cred.cr_group_info; + +diff -NurpP --minimal linux-3.0.9/fs/nfsd/nfs3xdr.c linux-3.0.9-vs2.3.2.1/fs/nfsd/nfs3xdr.c +--- linux-3.0.9/fs/nfsd/nfs3xdr.c 2011-07-22 11:18:05.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/fs/nfsd/nfs3xdr.c 2011-06-10 22:11:24.000000000 +0200 +@@ -7,6 +7,7 @@ + */ + + #include <linux/namei.h> ++#include <linux/vs_tag.h> + #include "xdr3.h" + #include "auth.h" + +@@ -95,6 +96,8 @@ static __be32 * + decode_sattr3(__be32 *p, struct iattr *iap) + { + u32 tmp; ++ uid_t uid = 0; ++ gid_t gid = 0; + + iap->ia_valid = 0; + +@@ -104,12 +107,15 @@ decode_sattr3(__be32 *p, struct iattr *i + } + if (*p++) { + iap->ia_valid |= ATTR_UID; +- iap->ia_uid = ntohl(*p++); ++ uid = ntohl(*p++); + } + if (*p++) { + iap->ia_valid |= ATTR_GID; +- iap->ia_gid = ntohl(*p++); ++ gid = ntohl(*p++); + } ++ iap->ia_uid = INOTAG_UID(DX_TAG_NFSD, uid, gid); ++ iap->ia_gid = INOTAG_GID(DX_TAG_NFSD, uid, gid); ++ iap->ia_tag = INOTAG_TAG(DX_TAG_NFSD, uid, gid, 0); + if (*p++) { + u64 newsize; + +@@ -165,8 +171,12 @@ encode_fattr3(struct svc_rqst *rqstp, __ + *p++ = htonl(nfs3_ftypes[(stat->mode & S_IFMT) >> 12]); + *p++ = htonl((u32) stat->mode); + *p++ = htonl((u32) stat->nlink); +- *p++ = htonl((u32) nfsd_ruid(rqstp, stat->uid)); +- *p++ = htonl((u32) nfsd_rgid(rqstp, stat->gid)); ++ *p++ = htonl((u32) nfsd_ruid(rqstp, ++ TAGINO_UID(0 /* FIXME: DX_TAG(dentry->d_inode) */, ++ stat->uid, stat->tag))); ++ *p++ = htonl((u32) nfsd_rgid(rqstp, ++ TAGINO_GID(0 /* FIXME: DX_TAG(dentry->d_inode) */, ++ stat->gid, stat->tag))); + if (S_ISLNK(stat->mode) && stat->size > NFS3_MAXPATHLEN) { + p = xdr_encode_hyper(p, (u64) NFS3_MAXPATHLEN); + } else { +diff -NurpP --minimal linux-3.0.9/fs/nfsd/nfs4xdr.c linux-3.0.9-vs2.3.2.1/fs/nfsd/nfs4xdr.c +--- linux-3.0.9/fs/nfsd/nfs4xdr.c 2011-11-15 16:40:47.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/fs/nfsd/nfs4xdr.c 2011-11-15 17:37:07.000000000 +0100 +@@ -45,6 +45,7 @@ + #include <linux/statfs.h> + #include <linux/utsname.h> + #include <linux/sunrpc/svcauth_gss.h> ++#include <linux/vs_tag.h> + + #include "idmap.h" + #include "acl.h" +@@ -2102,14 +2103,18 @@ out_acl: + WRITE32(stat.nlink); + } + if (bmval1 & FATTR4_WORD1_OWNER) { +- status = nfsd4_encode_user(rqstp, stat.uid, &p, &buflen); ++ status = nfsd4_encode_user(rqstp, ++ TAGINO_UID(DX_TAG(dentry->d_inode), ++ stat.uid, stat.tag), &p, &buflen); + if (status == nfserr_resource) + goto out_resource; + if (status) + goto out; + } + if (bmval1 & FATTR4_WORD1_OWNER_GROUP) { +- status = nfsd4_encode_group(rqstp, stat.gid, &p, &buflen); ++ status = nfsd4_encode_group(rqstp, ++ TAGINO_GID(DX_TAG(dentry->d_inode), ++ stat.gid, stat.tag), &p, &buflen); + if (status == nfserr_resource) + goto out_resource; + if (status) +diff -NurpP --minimal linux-3.0.9/fs/nfsd/nfsxdr.c linux-3.0.9-vs2.3.2.1/fs/nfsd/nfsxdr.c +--- linux-3.0.9/fs/nfsd/nfsxdr.c 2011-05-22 16:17:53.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/fs/nfsd/nfsxdr.c 2011-06-10 22:11:24.000000000 +0200 +@@ -6,6 +6,7 @@ + + #include "xdr.h" + #include "auth.h" ++#include <linux/vs_tag.h> + + #define NFSDDBG_FACILITY NFSDDBG_XDR + +@@ -88,6 +89,8 @@ static __be32 * + decode_sattr(__be32 *p, struct iattr *iap) + { + u32 tmp, tmp1; ++ uid_t uid = 0; ++ gid_t gid = 0; + + iap->ia_valid = 0; + +@@ -101,12 +104,15 @@ decode_sattr(__be32 *p, struct iattr *ia + } + if ((tmp = ntohl(*p++)) != (u32)-1) { + iap->ia_valid |= ATTR_UID; +- iap->ia_uid = tmp; ++ uid = tmp; + } + if ((tmp = ntohl(*p++)) != (u32)-1) { + iap->ia_valid |= ATTR_GID; +- iap->ia_gid = tmp; ++ gid = tmp; + } ++ iap->ia_uid = INOTAG_UID(DX_TAG_NFSD, uid, gid); ++ iap->ia_gid = INOTAG_GID(DX_TAG_NFSD, uid, gid); ++ iap->ia_tag = INOTAG_TAG(DX_TAG_NFSD, uid, gid, 0); + if ((tmp = ntohl(*p++)) != (u32)-1) { + iap->ia_valid |= ATTR_SIZE; + iap->ia_size = tmp; +@@ -151,8 +157,10 @@ encode_fattr(struct svc_rqst *rqstp, __b + *p++ = htonl(nfs_ftypes[type >> 12]); + *p++ = htonl((u32) stat->mode); + *p++ = htonl((u32) stat->nlink); +- *p++ = htonl((u32) nfsd_ruid(rqstp, stat->uid)); +- *p++ = htonl((u32) nfsd_rgid(rqstp, stat->gid)); ++ *p++ = htonl((u32) nfsd_ruid(rqstp, ++ TAGINO_UID(DX_TAG(dentry->d_inode), stat->uid, stat->tag))); ++ *p++ = htonl((u32) nfsd_rgid(rqstp, ++ TAGINO_GID(DX_TAG(dentry->d_inode), stat->gid, stat->tag))); + + if (S_ISLNK(type) && stat->size > NFS_MAXPATHLEN) { + *p++ = htonl(NFS_MAXPATHLEN); +diff -NurpP --minimal linux-3.0.9/fs/ocfs2/dlmglue.c linux-3.0.9-vs2.3.2.1/fs/ocfs2/dlmglue.c +--- linux-3.0.9/fs/ocfs2/dlmglue.c 2011-05-22 16:17:53.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/fs/ocfs2/dlmglue.c 2011-06-10 22:11:24.000000000 +0200 +@@ -2041,6 +2041,7 @@ static void __ocfs2_stuff_meta_lvb(struc + lvb->lvb_iclusters = cpu_to_be32(oi->ip_clusters); + lvb->lvb_iuid = cpu_to_be32(inode->i_uid); + lvb->lvb_igid = cpu_to_be32(inode->i_gid); ++ lvb->lvb_itag = cpu_to_be16(inode->i_tag); + lvb->lvb_imode = cpu_to_be16(inode->i_mode); + lvb->lvb_inlink = cpu_to_be16(inode->i_nlink); + lvb->lvb_iatime_packed = +@@ -2091,6 +2092,7 @@ static void ocfs2_refresh_inode_from_lvb + + inode->i_uid = be32_to_cpu(lvb->lvb_iuid); + inode->i_gid = be32_to_cpu(lvb->lvb_igid); ++ inode->i_tag = be16_to_cpu(lvb->lvb_itag); + inode->i_mode = be16_to_cpu(lvb->lvb_imode); + inode->i_nlink = be16_to_cpu(lvb->lvb_inlink); + ocfs2_unpack_timespec(&inode->i_atime, +diff -NurpP --minimal linux-3.0.9/fs/ocfs2/dlmglue.h linux-3.0.9-vs2.3.2.1/fs/ocfs2/dlmglue.h +--- linux-3.0.9/fs/ocfs2/dlmglue.h 2010-10-21 13:07:50.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/fs/ocfs2/dlmglue.h 2011-06-10 22:11:24.000000000 +0200 +@@ -46,7 +46,8 @@ struct ocfs2_meta_lvb { + __be16 lvb_inlink; + __be32 lvb_iattr; + __be32 lvb_igeneration; +- __be32 lvb_reserved2; ++ __be16 lvb_itag; ++ __be16 lvb_reserved2; + }; + + #define OCFS2_QINFO_LVB_VERSION 1 +diff -NurpP --minimal linux-3.0.9/fs/ocfs2/file.c linux-3.0.9-vs2.3.2.1/fs/ocfs2/file.c +--- linux-3.0.9/fs/ocfs2/file.c 2011-07-22 11:18:06.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/fs/ocfs2/file.c 2011-06-10 22:43:33.000000000 +0200 +@@ -1111,7 +1111,7 @@ int ocfs2_setattr(struct dentry *dentry, + attr->ia_valid &= ~ATTR_SIZE; + + #define OCFS2_VALID_ATTRS (ATTR_ATIME | ATTR_MTIME | ATTR_CTIME | ATTR_SIZE \ +- | ATTR_GID | ATTR_UID | ATTR_MODE) ++ | ATTR_GID | ATTR_UID | ATTR_TAG | ATTR_MODE) + if (!(attr->ia_valid & OCFS2_VALID_ATTRS)) + return 0; + +diff -NurpP --minimal linux-3.0.9/fs/ocfs2/inode.c linux-3.0.9-vs2.3.2.1/fs/ocfs2/inode.c +--- linux-3.0.9/fs/ocfs2/inode.c 2011-05-22 16:17:53.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/fs/ocfs2/inode.c 2011-06-10 22:11:24.000000000 +0200 +@@ -28,6 +28,7 @@ + #include <linux/highmem.h> + #include <linux/pagemap.h> + #include <linux/quotaops.h> ++#include <linux/vs_tag.h> + + #include <asm/byteorder.h> + +@@ -78,11 +79,13 @@ void ocfs2_set_inode_flags(struct inode + { + unsigned int flags = OCFS2_I(inode)->ip_attr; + +- inode->i_flags &= ~(S_IMMUTABLE | ++ inode->i_flags &= ~(S_IMMUTABLE | S_IXUNLINK | + S_SYNC | S_APPEND | S_NOATIME | S_DIRSYNC); + + if (flags & OCFS2_IMMUTABLE_FL) + inode->i_flags |= S_IMMUTABLE; ++ if (flags & OCFS2_IXUNLINK_FL) ++ inode->i_flags |= S_IXUNLINK; + + if (flags & OCFS2_SYNC_FL) + inode->i_flags |= S_SYNC; +@@ -92,25 +95,44 @@ void ocfs2_set_inode_flags(struct inode + inode->i_flags |= S_NOATIME; + if (flags & OCFS2_DIRSYNC_FL) + inode->i_flags |= S_DIRSYNC; ++ ++ inode->i_vflags &= ~(V_BARRIER | V_COW); ++ ++ if (flags & OCFS2_BARRIER_FL) ++ inode->i_vflags |= V_BARRIER; ++ if (flags & OCFS2_COW_FL) ++ inode->i_vflags |= V_COW; + } + + /* Propagate flags from i_flags to OCFS2_I(inode)->ip_attr */ + void ocfs2_get_inode_flags(struct ocfs2_inode_info *oi) + { + unsigned int flags = oi->vfs_inode.i_flags; ++ unsigned int vflags = oi->vfs_inode.i_vflags; ++ ++ oi->ip_attr &= ~(OCFS2_SYNC_FL | OCFS2_APPEND_FL | ++ OCFS2_IMMUTABLE_FL | OCFS2_IXUNLINK_FL | ++ OCFS2_NOATIME_FL | OCFS2_DIRSYNC_FL | ++ OCFS2_BARRIER_FL | OCFS2_COW_FL); ++ ++ if (flags & S_IMMUTABLE) ++ oi->ip_attr |= OCFS2_IMMUTABLE_FL; ++ if (flags & S_IXUNLINK) ++ oi->ip_attr |= OCFS2_IXUNLINK_FL; + +- oi->ip_attr &= ~(OCFS2_SYNC_FL|OCFS2_APPEND_FL| +- OCFS2_IMMUTABLE_FL|OCFS2_NOATIME_FL|OCFS2_DIRSYNC_FL); + if (flags & S_SYNC) + oi->ip_attr |= OCFS2_SYNC_FL; + if (flags & S_APPEND) + oi->ip_attr |= OCFS2_APPEND_FL; +- if (flags & S_IMMUTABLE) +- oi->ip_attr |= OCFS2_IMMUTABLE_FL; + if (flags & S_NOATIME) + oi->ip_attr |= OCFS2_NOATIME_FL; + if (flags & S_DIRSYNC) + oi->ip_attr |= OCFS2_DIRSYNC_FL; ++ ++ if (vflags & V_BARRIER) ++ oi->ip_attr |= OCFS2_BARRIER_FL; ++ if (vflags & V_COW) ++ oi->ip_attr |= OCFS2_COW_FL; + } + + struct inode *ocfs2_ilookup(struct super_block *sb, u64 blkno) +@@ -241,6 +263,8 @@ void ocfs2_populate_inode(struct inode * + struct super_block *sb; + struct ocfs2_super *osb; + int use_plocks = 1; ++ uid_t uid; ++ gid_t gid; + + sb = inode->i_sb; + osb = OCFS2_SB(sb); +@@ -269,8 +293,12 @@ void ocfs2_populate_inode(struct inode * + inode->i_generation = le32_to_cpu(fe->i_generation); + inode->i_rdev = huge_decode_dev(le64_to_cpu(fe->id1.dev1.i_rdev)); + inode->i_mode = le16_to_cpu(fe->i_mode); +- inode->i_uid = le32_to_cpu(fe->i_uid); +- inode->i_gid = le32_to_cpu(fe->i_gid); ++ uid = le32_to_cpu(fe->i_uid); ++ gid = le32_to_cpu(fe->i_gid); ++ inode->i_uid = INOTAG_UID(DX_TAG(inode), uid, gid); ++ inode->i_gid = INOTAG_GID(DX_TAG(inode), uid, gid); ++ inode->i_tag = INOTAG_TAG(DX_TAG(inode), uid, gid, ++ /* le16_to_cpu(raw_inode->i_raw_tag)i */ 0); + + /* Fast symlinks will have i_size but no allocated clusters. */ + if (S_ISLNK(inode->i_mode) && !fe->i_clusters) +diff -NurpP --minimal linux-3.0.9/fs/ocfs2/inode.h linux-3.0.9-vs2.3.2.1/fs/ocfs2/inode.h +--- linux-3.0.9/fs/ocfs2/inode.h 2011-01-05 21:50:26.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/fs/ocfs2/inode.h 2011-06-10 22:11:24.000000000 +0200 +@@ -151,6 +151,7 @@ struct buffer_head *ocfs2_bread(struct i + + void ocfs2_set_inode_flags(struct inode *inode); + void ocfs2_get_inode_flags(struct ocfs2_inode_info *oi); ++int ocfs2_sync_flags(struct inode *inode, int, int); + + static inline blkcnt_t ocfs2_inode_sector_count(struct inode *inode) + { +diff -NurpP --minimal linux-3.0.9/fs/ocfs2/ioctl.c linux-3.0.9-vs2.3.2.1/fs/ocfs2/ioctl.c +--- linux-3.0.9/fs/ocfs2/ioctl.c 2011-07-22 11:18:06.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/fs/ocfs2/ioctl.c 2011-06-10 22:11:24.000000000 +0200 +@@ -78,7 +78,41 @@ static int ocfs2_get_inode_attr(struct i + return status; + } + +-static int ocfs2_set_inode_attr(struct inode *inode, unsigned flags, ++int ocfs2_sync_flags(struct inode *inode, int flags, int vflags) ++{ ++ struct ocfs2_super *osb = OCFS2_SB(inode->i_sb); ++ struct buffer_head *bh = NULL; ++ handle_t *handle = NULL; ++ int status; ++ ++ status = ocfs2_inode_lock(inode, &bh, 1); ++ if (status < 0) { ++ mlog_errno(status); ++ return status; ++ } ++ handle = ocfs2_start_trans(osb, OCFS2_INODE_UPDATE_CREDITS); ++ if (IS_ERR(handle)) { ++ status = PTR_ERR(handle); ++ mlog_errno(status); ++ goto bail_unlock; ++ } ++ ++ inode->i_flags = flags; ++ inode->i_vflags = vflags; ++ ocfs2_get_inode_flags(OCFS2_I(inode)); ++ ++ status = ocfs2_mark_inode_dirty(handle, inode, bh); ++ if (status < 0) ++ mlog_errno(status); ++ ++ ocfs2_commit_trans(osb, handle); ++bail_unlock: ++ ocfs2_inode_unlock(inode, 1); ++ brelse(bh); ++ return status; ++} ++ ++int ocfs2_set_inode_attr(struct inode *inode, unsigned flags, + unsigned mask) + { + struct ocfs2_inode_info *ocfs2_inode = OCFS2_I(inode); +@@ -103,6 +137,11 @@ static int ocfs2_set_inode_attr(struct i + if (!S_ISDIR(inode->i_mode)) + flags &= ~OCFS2_DIRSYNC_FL; + ++ if (IS_BARRIER(inode)) { ++ vxwprintk_task(1, "messing with the barrier."); ++ goto bail_unlock; ++ } ++ + handle = ocfs2_start_trans(osb, OCFS2_INODE_UPDATE_CREDITS); + if (IS_ERR(handle)) { + status = PTR_ERR(handle); +@@ -880,6 +919,7 @@ bail: + return status; + } + ++ + long ocfs2_ioctl(struct file *filp, unsigned int cmd, unsigned long arg) + { + struct inode *inode = filp->f_path.dentry->d_inode; +diff -NurpP --minimal linux-3.0.9/fs/ocfs2/namei.c linux-3.0.9-vs2.3.2.1/fs/ocfs2/namei.c +--- linux-3.0.9/fs/ocfs2/namei.c 2011-05-22 16:17:53.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/fs/ocfs2/namei.c 2011-06-10 22:11:24.000000000 +0200 +@@ -41,6 +41,7 @@ + #include <linux/slab.h> + #include <linux/highmem.h> + #include <linux/quotaops.h> ++#include <linux/vs_tag.h> + + #include <cluster/masklog.h> + +@@ -477,6 +478,7 @@ static int __ocfs2_mknod_locked(struct i + struct ocfs2_dinode *fe = NULL; + struct ocfs2_extent_list *fel; + u16 feat; ++ tag_t tag; + + *new_fe_bh = NULL; + +@@ -514,8 +516,11 @@ static int __ocfs2_mknod_locked(struct i + fe->i_suballoc_loc = cpu_to_le64(suballoc_loc); + fe->i_suballoc_bit = cpu_to_le16(suballoc_bit); + fe->i_suballoc_slot = cpu_to_le16(inode_ac->ac_alloc_slot); +- fe->i_uid = cpu_to_le32(inode->i_uid); +- fe->i_gid = cpu_to_le32(inode->i_gid); ++ ++ tag = dx_current_fstag(osb->sb); ++ fe->i_uid = cpu_to_le32(TAGINO_UID(DX_TAG(inode), inode->i_uid, tag)); ++ fe->i_gid = cpu_to_le32(TAGINO_GID(DX_TAG(inode), inode->i_gid, tag)); ++ inode->i_tag = tag; + fe->i_mode = cpu_to_le16(inode->i_mode); + if (S_ISCHR(inode->i_mode) || S_ISBLK(inode->i_mode)) + fe->id1.dev1.i_rdev = cpu_to_le64(huge_encode_dev(dev)); +diff -NurpP --minimal linux-3.0.9/fs/ocfs2/ocfs2.h linux-3.0.9-vs2.3.2.1/fs/ocfs2/ocfs2.h +--- linux-3.0.9/fs/ocfs2/ocfs2.h 2011-05-22 16:17:53.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/fs/ocfs2/ocfs2.h 2011-06-10 22:11:24.000000000 +0200 +@@ -272,6 +272,7 @@ enum ocfs2_mount_options + writes */ + OCFS2_MOUNT_HB_NONE = 1 << 13, /* No heartbeat */ + OCFS2_MOUNT_HB_GLOBAL = 1 << 14, /* Global heartbeat */ ++ OCFS2_MOUNT_TAGGED = 1 << 15, /* use tagging */ + }; + + #define OCFS2_OSB_SOFT_RO 0x0001 +diff -NurpP --minimal linux-3.0.9/fs/ocfs2/ocfs2_fs.h linux-3.0.9-vs2.3.2.1/fs/ocfs2/ocfs2_fs.h +--- linux-3.0.9/fs/ocfs2/ocfs2_fs.h 2011-05-22 16:17:53.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/fs/ocfs2/ocfs2_fs.h 2011-06-10 22:11:24.000000000 +0200 +@@ -266,6 +266,11 @@ + #define OCFS2_TOPDIR_FL FS_TOPDIR_FL /* Top of directory hierarchies*/ + #define OCFS2_RESERVED_FL FS_RESERVED_FL /* reserved for ext2 lib */ + ++#define OCFS2_IXUNLINK_FL FS_IXUNLINK_FL /* Immutable invert on unlink */ ++ ++#define OCFS2_BARRIER_FL FS_BARRIER_FL /* Barrier for chroot() */ ++#define OCFS2_COW_FL FS_COW_FL /* Copy on Write marker */ ++ + #define OCFS2_FL_VISIBLE FS_FL_USER_VISIBLE /* User visible flags */ + #define OCFS2_FL_MODIFIABLE FS_FL_USER_MODIFIABLE /* User modifiable flags */ + +diff -NurpP --minimal linux-3.0.9/fs/ocfs2/super.c linux-3.0.9-vs2.3.2.1/fs/ocfs2/super.c +--- linux-3.0.9/fs/ocfs2/super.c 2011-07-22 11:18:06.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/fs/ocfs2/super.c 2011-06-15 02:40:14.000000000 +0200 +@@ -184,6 +184,7 @@ enum { + Opt_coherency_full, + Opt_resv_level, + Opt_dir_resv_level, ++ Opt_tag, Opt_notag, Opt_tagid, + Opt_err, + }; + +@@ -215,6 +216,9 @@ static const match_table_t tokens = { + {Opt_coherency_full, "coherency=full"}, + {Opt_resv_level, "resv_level=%u"}, + {Opt_dir_resv_level, "dir_resv_level=%u"}, ++ {Opt_tag, "tag"}, ++ {Opt_notag, "notag"}, ++ {Opt_tagid, "tagid=%u"}, + {Opt_err, NULL} + }; + +@@ -662,6 +666,13 @@ static int ocfs2_remount(struct super_bl + goto out; + } + ++ if ((osb->s_mount_opt & OCFS2_MOUNT_TAGGED) != ++ (parsed_options.mount_opt & OCFS2_MOUNT_TAGGED)) { ++ ret = -EINVAL; ++ mlog(ML_ERROR, "Cannot change tagging on remount\n"); ++ goto out; ++ } ++ + /* We're going to/from readonly mode. */ + if ((*flags & MS_RDONLY) != (sb->s_flags & MS_RDONLY)) { + /* Disable quota accounting before remounting RO */ +@@ -1177,6 +1188,9 @@ static int ocfs2_fill_super(struct super + + ocfs2_complete_mount_recovery(osb); + ++ if (osb->s_mount_opt & OCFS2_MOUNT_TAGGED) ++ sb->s_flags |= MS_TAGGED; ++ + if (ocfs2_mount_local(osb)) + snprintf(nodestr, sizeof(nodestr), "local"); + else +@@ -1506,6 +1520,20 @@ static int ocfs2_parse_options(struct su + option < OCFS2_MAX_RESV_LEVEL) + mopt->dir_resv_level = option; + break; ++#ifndef CONFIG_TAGGING_NONE ++ case Opt_tag: ++ mopt->mount_opt |= OCFS2_MOUNT_TAGGED; ++ break; ++ case Opt_notag: ++ mopt->mount_opt &= ~OCFS2_MOUNT_TAGGED; ++ break; ++#endif ++#ifdef CONFIG_PROPAGATE ++ case Opt_tagid: ++ /* use args[0] */ ++ mopt->mount_opt |= OCFS2_MOUNT_TAGGED; ++ break; ++#endif + default: + mlog(ML_ERROR, + "Unrecognized mount option \"%s\" " +diff -NurpP --minimal linux-3.0.9/fs/open.c linux-3.0.9-vs2.3.2.1/fs/open.c +--- linux-3.0.9/fs/open.c 2011-05-22 16:17:53.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/fs/open.c 2011-06-10 22:11:24.000000000 +0200 +@@ -30,6 +30,11 @@ + #include <linux/fs_struct.h> + #include <linux/ima.h> + #include <linux/dnotify.h> ++#include <linux/vs_base.h> ++#include <linux/vs_limit.h> ++#include <linux/vs_tag.h> ++#include <linux/vs_cowbl.h> ++#include <linux/vserver/dlimit.h> + + #include "internal.h" + +@@ -494,6 +499,12 @@ SYSCALL_DEFINE3(fchmodat, int, dfd, cons + error = user_path_at(dfd, filename, LOOKUP_FOLLOW, &path); + if (error) + goto out; ++ ++#ifdef CONFIG_VSERVER_COWBL ++ error = cow_check_and_break(&path); ++ if (error) ++ goto dput_and_out; ++#endif + inode = path.dentry->d_inode; + + error = mnt_want_write(path.mnt); +@@ -531,11 +542,11 @@ static int chown_common(struct path *pat + newattrs.ia_valid = ATTR_CTIME; + if (user != (uid_t) -1) { + newattrs.ia_valid |= ATTR_UID; +- newattrs.ia_uid = user; ++ newattrs.ia_uid = dx_map_uid(user); + } + if (group != (gid_t) -1) { + newattrs.ia_valid |= ATTR_GID; +- newattrs.ia_gid = group; ++ newattrs.ia_gid = dx_map_gid(group); + } + if (!S_ISDIR(inode->i_mode)) + newattrs.ia_valid |= +@@ -560,6 +571,10 @@ SYSCALL_DEFINE3(chown, const char __user + error = mnt_want_write(path.mnt); + if (error) + goto out_release; ++#ifdef CONFIG_VSERVER_COWBL ++ error = cow_check_and_break(&path); ++ if (!error) ++#endif + error = chown_common(&path, user, group); + mnt_drop_write(path.mnt); + out_release: +@@ -587,6 +602,10 @@ SYSCALL_DEFINE5(fchownat, int, dfd, cons + error = mnt_want_write(path.mnt); + if (error) + goto out_release; ++#ifdef CONFIG_VSERVER_COWBL ++ error = cow_check_and_break(&path); ++ if (!error) ++#endif + error = chown_common(&path, user, group); + mnt_drop_write(path.mnt); + out_release: +@@ -606,6 +625,10 @@ SYSCALL_DEFINE3(lchown, const char __use + error = mnt_want_write(path.mnt); + if (error) + goto out_release; ++#ifdef CONFIG_VSERVER_COWBL ++ error = cow_check_and_break(&path); ++ if (!error) ++#endif + error = chown_common(&path, user, group); + mnt_drop_write(path.mnt); + out_release: +@@ -857,6 +880,7 @@ static void __put_unused_fd(struct files + __FD_CLR(fd, fdt->open_fds); + if (fd < files->next_fd) + files->next_fd = fd; ++ vx_openfd_dec(fd); + } + + void put_unused_fd(unsigned int fd) +diff -NurpP --minimal linux-3.0.9/fs/proc/array.c linux-3.0.9-vs2.3.2.1/fs/proc/array.c +--- linux-3.0.9/fs/proc/array.c 2011-07-22 11:18:06.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/fs/proc/array.c 2011-06-10 22:11:24.000000000 +0200 +@@ -81,6 +81,8 @@ + #include <linux/pid_namespace.h> + #include <linux/ptrace.h> + #include <linux/tracehook.h> ++#include <linux/vs_context.h> ++#include <linux/vs_network.h> + + #include <asm/pgtable.h> + #include <asm/processor.h> +@@ -170,6 +172,9 @@ static inline void task_state(struct seq + rcu_read_lock(); + ppid = pid_alive(p) ? + task_tgid_nr_ns(rcu_dereference(p->real_parent), ns) : 0; ++ if (unlikely(vx_current_initpid(p->pid))) ++ ppid = 0; ++ + tpid = 0; + if (pid_alive(p)) { + struct task_struct *tracer = tracehook_tracer_task(p); +@@ -287,7 +292,7 @@ static inline void task_sig(struct seq_f + } + + static void render_cap_t(struct seq_file *m, const char *header, +- kernel_cap_t *a) ++ struct vx_info *vxi, kernel_cap_t *a) + { + unsigned __capi; + +@@ -312,10 +317,11 @@ static inline void task_cap(struct seq_f + cap_bset = cred->cap_bset; + rcu_read_unlock(); + +- render_cap_t(m, "CapInh:\t", &cap_inheritable); +- render_cap_t(m, "CapPrm:\t", &cap_permitted); +- render_cap_t(m, "CapEff:\t", &cap_effective); +- render_cap_t(m, "CapBnd:\t", &cap_bset); ++ /* FIXME: maybe move the p->vx_info masking to __task_cred() ? */ ++ render_cap_t(m, "CapInh:\t", p->vx_info, &cap_inheritable); ++ render_cap_t(m, "CapPrm:\t", p->vx_info, &cap_permitted); ++ render_cap_t(m, "CapEff:\t", p->vx_info, &cap_effective); ++ render_cap_t(m, "CapBnd:\t", p->vx_info, &cap_bset); + } + + static inline void task_context_switch_counts(struct seq_file *m, +@@ -337,6 +343,42 @@ static void task_cpus_allowed(struct seq + seq_putc(m, '\n'); + } + ++int proc_pid_nsproxy(struct seq_file *m, struct pid_namespace *ns, ++ struct pid *pid, struct task_struct *task) ++{ ++ seq_printf(m, "Proxy:\t%p(%c)\n" ++ "Count:\t%u\n" ++ "uts:\t%p(%c)\n" ++ "ipc:\t%p(%c)\n" ++ "mnt:\t%p(%c)\n" ++ "pid:\t%p(%c)\n" ++ "net:\t%p(%c)\n", ++ task->nsproxy, ++ (task->nsproxy == init_task.nsproxy ? 'I' : '-'), ++ atomic_read(&task->nsproxy->count), ++ task->nsproxy->uts_ns, ++ (task->nsproxy->uts_ns == init_task.nsproxy->uts_ns ? 'I' : '-'), ++ task->nsproxy->ipc_ns, ++ (task->nsproxy->ipc_ns == init_task.nsproxy->ipc_ns ? 'I' : '-'), ++ task->nsproxy->mnt_ns, ++ (task->nsproxy->mnt_ns == init_task.nsproxy->mnt_ns ? 'I' : '-'), ++ task->nsproxy->pid_ns, ++ (task->nsproxy->pid_ns == init_task.nsproxy->pid_ns ? 'I' : '-'), ++ task->nsproxy->net_ns, ++ (task->nsproxy->net_ns == init_task.nsproxy->net_ns ? 'I' : '-')); ++ return 0; ++} ++ ++void task_vs_id(struct seq_file *m, struct task_struct *task) ++{ ++ if (task_vx_flags(task, VXF_HIDE_VINFO, 0)) ++ return; ++ ++ seq_printf(m, "VxID: %d\n", vx_task_xid(task)); ++ seq_printf(m, "NxID: %d\n", nx_task_nid(task)); ++} ++ ++ + int proc_pid_status(struct seq_file *m, struct pid_namespace *ns, + struct pid *pid, struct task_struct *task) + { +@@ -353,6 +395,7 @@ int proc_pid_status(struct seq_file *m, + task_cap(m, task); + task_cpus_allowed(m, task); + cpuset_task_status_allowed(m, task); ++ task_vs_id(m, task); + task_context_switch_counts(m, task); + return 0; + } +@@ -462,6 +505,17 @@ static int do_task_stat(struct seq_file + /* convert nsec -> ticks */ + start_time = nsec_to_clock_t(start_time); + ++ /* fixup start time for virt uptime */ ++ if (vx_flags(VXF_VIRT_UPTIME, 0)) { ++ unsigned long long bias = ++ current->vx_info->cvirt.bias_clock; ++ ++ if (start_time > bias) ++ start_time -= bias; ++ else ++ start_time = 0; ++ } ++ + seq_printf(m, "%d (%s) %c %d %d %d %d %d %u %lu \ + %lu %lu %lu %lu %lu %ld %ld %ld %ld %d 0 %llu %lu %ld %lu %lu %lu %lu %lu \ + %lu %lu %lu %lu %lu %lu %lu %lu %d %d %u %u %llu %lu %ld\n", +diff -NurpP --minimal linux-3.0.9/fs/proc/base.c linux-3.0.9-vs2.3.2.1/fs/proc/base.c +--- linux-3.0.9/fs/proc/base.c 2011-11-15 16:40:47.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/fs/proc/base.c 2011-11-15 17:37:07.000000000 +0100 +@@ -83,6 +83,8 @@ + #include <linux/pid_namespace.h> + #include <linux/fs_struct.h> + #include <linux/slab.h> ++#include <linux/vs_context.h> ++#include <linux/vs_network.h> + #ifdef CONFIG_HARDWALL + #include <asm/hardwall.h> + #endif +@@ -1102,11 +1104,16 @@ static ssize_t oom_adjust_write(struct f + goto err_task_lock; + } + +- if (oom_adjust < task->signal->oom_adj && !capable(CAP_SYS_RESOURCE)) { ++ if (oom_adjust < task->signal->oom_adj && ++ !vx_capable(CAP_SYS_RESOURCE, VXC_OOM_ADJUST)) { + err = -EACCES; + goto err_sighand; + } + ++ /* prevent guest processes from circumventing the oom killer */ ++ if (vx_current_xid() && (oom_adjust == OOM_DISABLE)) ++ oom_adjust = OOM_ADJUST_MIN; ++ + if (oom_adjust != task->signal->oom_adj) { + if (oom_adjust == OOM_DISABLE) + atomic_inc(&task->mm->oom_disable_count); +@@ -1275,7 +1282,7 @@ static ssize_t proc_loginuid_write(struc + ssize_t length; + uid_t loginuid; + +- if (!capable(CAP_AUDIT_CONTROL)) ++ if (!vx_capable(CAP_AUDIT_CONTROL, VXC_AUDIT_CONTROL)) + return -EPERM; + + rcu_read_lock(); +@@ -1722,6 +1729,8 @@ struct inode *proc_pid_make_inode(struct + inode->i_gid = cred->egid; + rcu_read_unlock(); + } ++ /* procfs is xid tagged */ ++ inode->i_tag = (tag_t)vx_task_xid(task); + security_task_to_inode(task, inode); + + out: +@@ -1758,6 +1767,8 @@ int pid_getattr(struct vfsmount *mnt, st + + /* dentry stuff */ + ++static unsigned name_to_int(struct dentry *dentry); ++ + /* + * Exceptional case: normally we are not allowed to unhash a busy + * directory. In this case, however, we can do it - no aliasing problems +@@ -1786,6 +1797,12 @@ int pid_revalidate(struct dentry *dentry + task = get_proc_task(inode); + + if (task) { ++ unsigned pid = name_to_int(dentry); ++ ++ if (pid != ~0U && pid != vx_map_pid(task->pid)) { ++ put_task_struct(task); ++ goto drop; ++ } + if ((inode->i_mode == (S_IFDIR|S_IRUGO|S_IXUGO)) || + task_dumpable(task)) { + rcu_read_lock(); +@@ -1802,6 +1819,7 @@ int pid_revalidate(struct dentry *dentry + put_task_struct(task); + return 1; + } ++drop: + d_drop(dentry); + return 0; + } +@@ -2291,6 +2309,13 @@ static struct dentry *proc_pident_lookup + if (!task) + goto out_no_task; + ++ /* TODO: maybe we can come up with a generic approach? */ ++ if (task_vx_flags(task, VXF_HIDE_VINFO, 0) && ++ (dentry->d_name.len == 5) && ++ (!memcmp(dentry->d_name.name, "vinfo", 5) || ++ !memcmp(dentry->d_name.name, "ninfo", 5))) ++ goto out; ++ + /* + * Yes, it does not scale. And it should not. Don't add + * new entries into /proc/<tgid>/ without very good reasons. +@@ -2676,7 +2701,7 @@ out_iput: + static struct dentry *proc_base_lookup(struct inode *dir, struct dentry *dentry) + { + struct dentry *error; +- struct task_struct *task = get_proc_task(dir); ++ struct task_struct *task = get_proc_task_real(dir); + const struct pid_entry *p, *last; + + error = ERR_PTR(-ENOENT); +@@ -2783,6 +2808,9 @@ static int proc_pid_personality(struct s + static const struct file_operations proc_task_operations; + static const struct inode_operations proc_task_inode_operations; + ++extern int proc_pid_vx_info(struct task_struct *, char *); ++extern int proc_pid_nx_info(struct task_struct *, char *); ++ + static const struct pid_entry tgid_base_stuff[] = { + DIR("task", S_IRUGO|S_IXUGO, proc_task_inode_operations, proc_task_operations), + DIR("fd", S_IRUSR|S_IXUSR, proc_fd_inode_operations, proc_fd_operations), +@@ -2846,6 +2874,8 @@ static const struct pid_entry tgid_base_ + #ifdef CONFIG_CGROUPS + REG("cgroup", S_IRUGO, proc_cgroup_operations), + #endif ++ INF("vinfo", S_IRUGO, proc_pid_vx_info), ++ INF("ninfo", S_IRUGO, proc_pid_nx_info), + INF("oom_score", S_IRUGO, proc_oom_score), + REG("oom_adj", S_IRUGO|S_IWUSR, proc_oom_adjust_operations), + REG("oom_score_adj", S_IRUGO|S_IWUSR, proc_oom_score_adj_operations), +@@ -2865,6 +2895,7 @@ static const struct pid_entry tgid_base_ + #ifdef CONFIG_HARDWALL + INF("hardwall", S_IRUGO, proc_pid_hardwall), + #endif ++ ONE("nsproxy", S_IRUGO, proc_pid_nsproxy), + }; + + static int proc_tgid_base_readdir(struct file * filp, +@@ -3057,7 +3088,7 @@ retry: + iter.task = NULL; + pid = find_ge_pid(iter.tgid, ns); + if (pid) { +- iter.tgid = pid_nr_ns(pid, ns); ++ iter.tgid = pid_unmapped_nr_ns(pid, ns); + iter.task = pid_task(pid, PIDTYPE_PID); + /* What we to know is if the pid we have find is the + * pid of a thread_group_leader. Testing for task +@@ -3087,7 +3118,7 @@ static int proc_pid_fill_cache(struct fi + struct tgid_iter iter) + { + char name[PROC_NUMBUF]; +- int len = snprintf(name, sizeof(name), "%d", iter.tgid); ++ int len = snprintf(name, sizeof(name), "%d", vx_map_tgid(iter.tgid)); + return proc_fill_cache(filp, dirent, filldir, name, len, + proc_pid_instantiate, iter.task, NULL); + } +@@ -3104,7 +3135,7 @@ int proc_pid_readdir(struct file * filp, + goto out_no_task; + nr = filp->f_pos - FIRST_PROCESS_ENTRY; + +- reaper = get_proc_task(filp->f_path.dentry->d_inode); ++ reaper = get_proc_task_real(filp->f_path.dentry->d_inode); + if (!reaper) + goto out_no_task; + +@@ -3121,6 +3152,8 @@ int proc_pid_readdir(struct file * filp, + iter.task; + iter.tgid += 1, iter = next_tgid(ns, iter)) { + filp->f_pos = iter.tgid + TGID_OFFSET; ++ if (!vx_proc_task_visible(iter.task)) ++ continue; + if (proc_pid_fill_cache(filp, dirent, filldir, iter) < 0) { + put_task_struct(iter.task); + goto out; +@@ -3274,6 +3307,8 @@ static struct dentry *proc_task_lookup(s + tid = name_to_int(dentry); + if (tid == ~0U) + goto out; ++ if (vx_current_initpid(tid)) ++ goto out; + + ns = dentry->d_sb->s_fs_info; + rcu_read_lock(); +diff -NurpP --minimal linux-3.0.9/fs/proc/generic.c linux-3.0.9-vs2.3.2.1/fs/proc/generic.c +--- linux-3.0.9/fs/proc/generic.c 2011-07-22 11:18:06.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/fs/proc/generic.c 2011-06-10 22:11:24.000000000 +0200 +@@ -22,6 +22,7 @@ + #include <linux/bitops.h> + #include <linux/spinlock.h> + #include <linux/completion.h> ++#include <linux/vserver/inode.h> + #include <asm/uaccess.h> + + #include "internal.h" +@@ -424,11 +425,15 @@ struct dentry *proc_lookup_de(struct pro + for (de = de->subdir; de ; de = de->next) { + if (de->namelen != dentry->d_name.len) + continue; ++ if (!vx_hide_check(0, de->vx_flags)) ++ continue; + if (!memcmp(dentry->d_name.name, de->name, de->namelen)) { + pde_get(de); + spin_unlock(&proc_subdir_lock); + error = -EINVAL; + inode = proc_get_inode(dir->i_sb, de); ++ /* generic proc entries belong to the host */ ++ inode->i_tag = 0; + goto out_unlock; + } + } +@@ -506,6 +511,8 @@ int proc_readdir_de(struct proc_dir_entr + + /* filldir passes info to user space */ + pde_get(de); ++ if (!vx_hide_check(0, de->vx_flags)) ++ goto skip; + spin_unlock(&proc_subdir_lock); + if (filldir(dirent, de->name, de->namelen, filp->f_pos, + de->low_ino, de->mode >> 12) < 0) { +@@ -513,6 +520,7 @@ int proc_readdir_de(struct proc_dir_entr + goto out; + } + spin_lock(&proc_subdir_lock); ++ skip: + filp->f_pos++; + next = de->next; + pde_put(de); +@@ -627,6 +635,7 @@ static struct proc_dir_entry *__proc_cre + ent->nlink = nlink; + atomic_set(&ent->count, 1); + ent->pde_users = 0; ++ ent->vx_flags = IATTR_PROC_DEFAULT; + spin_lock_init(&ent->pde_unload_lock); + ent->pde_unload_completion = NULL; + INIT_LIST_HEAD(&ent->pde_openers); +@@ -650,7 +659,8 @@ struct proc_dir_entry *proc_symlink(cons + kfree(ent->data); + kfree(ent); + ent = NULL; +- } ++ } else ++ ent->vx_flags = IATTR_PROC_SYMLINK; + } else { + kfree(ent); + ent = NULL; +diff -NurpP --minimal linux-3.0.9/fs/proc/inode.c linux-3.0.9-vs2.3.2.1/fs/proc/inode.c +--- linux-3.0.9/fs/proc/inode.c 2011-07-22 11:18:06.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/fs/proc/inode.c 2011-06-10 22:11:24.000000000 +0200 +@@ -442,6 +442,8 @@ struct inode *proc_get_inode(struct supe + inode->i_uid = de->uid; + inode->i_gid = de->gid; + } ++ if (de->vx_flags) ++ PROC_I(inode)->vx_flags = de->vx_flags; + if (de->size) + inode->i_size = de->size; + if (de->nlink) +diff -NurpP --minimal linux-3.0.9/fs/proc/internal.h linux-3.0.9-vs2.3.2.1/fs/proc/internal.h +--- linux-3.0.9/fs/proc/internal.h 2011-07-22 11:18:06.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/fs/proc/internal.h 2011-06-10 22:11:24.000000000 +0200 +@@ -10,6 +10,7 @@ + */ + + #include <linux/proc_fs.h> ++#include <linux/vs_pid.h> + + extern struct proc_dir_entry proc_root; + #ifdef CONFIG_PROC_SYSCTL +@@ -51,6 +52,9 @@ extern int proc_pid_status(struct seq_fi + struct pid *pid, struct task_struct *task); + extern int proc_pid_statm(struct seq_file *m, struct pid_namespace *ns, + struct pid *pid, struct task_struct *task); ++extern int proc_pid_nsproxy(struct seq_file *m, struct pid_namespace *ns, ++ struct pid *pid, struct task_struct *task); ++ + extern loff_t mem_lseek(struct file *file, loff_t offset, int orig); + + extern const struct file_operations proc_maps_operations; +@@ -76,11 +80,16 @@ static inline struct pid *proc_pid(struc + return PROC_I(inode)->pid; + } + +-static inline struct task_struct *get_proc_task(struct inode *inode) ++static inline struct task_struct *get_proc_task_real(struct inode *inode) + { + return get_pid_task(proc_pid(inode), PIDTYPE_PID); + } + ++static inline struct task_struct *get_proc_task(struct inode *inode) ++{ ++ return vx_get_proc_task(inode, proc_pid(inode)); ++} ++ + static inline int proc_fd(struct inode *inode) + { + return PROC_I(inode)->fd; +diff -NurpP --minimal linux-3.0.9/fs/proc/loadavg.c linux-3.0.9-vs2.3.2.1/fs/proc/loadavg.c +--- linux-3.0.9/fs/proc/loadavg.c 2009-09-10 15:26:23.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/fs/proc/loadavg.c 2011-06-10 22:11:24.000000000 +0200 +@@ -12,15 +12,27 @@ + + static int loadavg_proc_show(struct seq_file *m, void *v) + { ++ unsigned long running; ++ unsigned int threads; + unsigned long avnrun[3]; + + get_avenrun(avnrun, FIXED_1/200, 0); + ++ if (vx_flags(VXF_VIRT_LOAD, 0)) { ++ struct vx_info *vxi = current_vx_info(); ++ ++ running = atomic_read(&vxi->cvirt.nr_running); ++ threads = atomic_read(&vxi->cvirt.nr_threads); ++ } else { ++ running = nr_running(); ++ threads = nr_threads; ++ } ++ + seq_printf(m, "%lu.%02lu %lu.%02lu %lu.%02lu %ld/%d %d\n", + LOAD_INT(avnrun[0]), LOAD_FRAC(avnrun[0]), + LOAD_INT(avnrun[1]), LOAD_FRAC(avnrun[1]), + LOAD_INT(avnrun[2]), LOAD_FRAC(avnrun[2]), +- nr_running(), nr_threads, ++ running, threads, + task_active_pid_ns(current)->last_pid); + return 0; + } +diff -NurpP --minimal linux-3.0.9/fs/proc/meminfo.c linux-3.0.9-vs2.3.2.1/fs/proc/meminfo.c +--- linux-3.0.9/fs/proc/meminfo.c 2011-03-15 18:07:33.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/fs/proc/meminfo.c 2011-06-10 22:11:24.000000000 +0200 +@@ -39,7 +39,8 @@ static int meminfo_proc_show(struct seq_ + allowed = ((totalram_pages - hugetlb_total_pages()) + * sysctl_overcommit_ratio / 100) + total_swap_pages; + +- cached = global_page_state(NR_FILE_PAGES) - ++ cached = vx_flags(VXF_VIRT_MEM, 0) ? ++ vx_vsi_cached(&i) : global_page_state(NR_FILE_PAGES) - + total_swapcache_pages - i.bufferram; + if (cached < 0) + cached = 0; +diff -NurpP --minimal linux-3.0.9/fs/proc/root.c linux-3.0.9-vs2.3.2.1/fs/proc/root.c +--- linux-3.0.9/fs/proc/root.c 2011-07-22 11:18:06.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/fs/proc/root.c 2011-06-22 12:39:15.000000000 +0200 +@@ -18,9 +18,14 @@ + #include <linux/bitops.h> + #include <linux/mount.h> + #include <linux/pid_namespace.h> ++#include <linux/vserver/inode.h> + + #include "internal.h" + ++struct proc_dir_entry *proc_virtual; ++ ++extern void proc_vx_init(void); ++ + static int proc_test_super(struct super_block *sb, void *data) + { + return sb->s_fs_info == data; +@@ -125,6 +130,7 @@ void __init proc_root_init(void) + #endif + proc_mkdir("bus", NULL); + proc_sys_init(); ++ proc_vx_init(); + } + + static int proc_root_getattr(struct vfsmount *mnt, struct dentry *dentry, struct kstat *stat +@@ -193,6 +199,7 @@ struct proc_dir_entry proc_root = { + .proc_iops = &proc_root_inode_operations, + .proc_fops = &proc_root_operations, + .parent = &proc_root, ++ .vx_flags = IATTR_ADMIN | IATTR_WATCH, + }; + + int pid_ns_prepare_proc(struct pid_namespace *ns) +diff -NurpP --minimal linux-3.0.9/fs/proc/uptime.c linux-3.0.9-vs2.3.2.1/fs/proc/uptime.c +--- linux-3.0.9/fs/proc/uptime.c 2009-12-03 20:02:53.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/fs/proc/uptime.c 2011-06-10 22:11:24.000000000 +0200 +@@ -4,22 +4,22 @@ + #include <linux/sched.h> + #include <linux/seq_file.h> + #include <linux/time.h> +-#include <linux/kernel_stat.h> ++#include <linux/vserver/cvirt.h> + #include <asm/cputime.h> + + static int uptime_proc_show(struct seq_file *m, void *v) + { + struct timespec uptime; + struct timespec idle; +- int i; +- cputime_t idletime = cputime_zero; +- +- for_each_possible_cpu(i) +- idletime = cputime64_add(idletime, kstat_cpu(i).cpustat.idle); ++ cputime_t idletime = cputime_add(init_task.utime, init_task.stime); + + do_posix_clock_monotonic_gettime(&uptime); + monotonic_to_bootbased(&uptime); + cputime_to_timespec(idletime, &idle); ++ ++ if (vx_flags(VXF_VIRT_UPTIME, 0)) ++ vx_vsi_uptime(&uptime, &idle); ++ + seq_printf(m, "%lu.%02lu %lu.%02lu\n", + (unsigned long) uptime.tv_sec, + (uptime.tv_nsec / (NSEC_PER_SEC / 100)), +diff -NurpP --minimal linux-3.0.9/fs/quota/dquot.c linux-3.0.9-vs2.3.2.1/fs/quota/dquot.c +--- linux-3.0.9/fs/quota/dquot.c 2011-07-22 11:18:06.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/fs/quota/dquot.c 2011-06-10 22:11:24.000000000 +0200 +@@ -1548,6 +1548,9 @@ int __dquot_alloc_space(struct inode *in + int reserve = flags & DQUOT_SPACE_RESERVE; + int nofail = flags & DQUOT_SPACE_NOFAIL; + ++ if ((ret = dl_alloc_space(inode, number))) ++ return ret; ++ + /* + * First test before acquiring mutex - solves deadlocks when we + * re-enter the quota code and are already holding the mutex +@@ -1602,6 +1605,9 @@ int dquot_alloc_inode(const struct inode + int cnt, ret = 0; + char warntype[MAXQUOTAS]; + ++ if ((ret = dl_alloc_inode(inode))) ++ return ret; ++ + /* First test before acquiring mutex - solves deadlocks when we + * re-enter the quota code and are already holding the mutex */ + if (!dquot_active(inode)) +@@ -1672,6 +1678,8 @@ void __dquot_free_space(struct inode *in + char warntype[MAXQUOTAS]; + int reserve = flags & DQUOT_SPACE_RESERVE; + ++ dl_free_space(inode, number); ++ + /* First test before acquiring mutex - solves deadlocks when we + * re-enter the quota code and are already holding the mutex */ + if (!dquot_active(inode)) { +@@ -1710,6 +1718,8 @@ void dquot_free_inode(const struct inode + unsigned int cnt; + char warntype[MAXQUOTAS]; + ++ dl_free_inode(inode); ++ + /* First test before acquiring mutex - solves deadlocks when we + * re-enter the quota code and are already holding the mutex */ + if (!dquot_active(inode)) +diff -NurpP --minimal linux-3.0.9/fs/quota/quota.c linux-3.0.9-vs2.3.2.1/fs/quota/quota.c +--- linux-3.0.9/fs/quota/quota.c 2011-11-15 16:40:47.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/fs/quota/quota.c 2011-11-15 17:37:07.000000000 +0100 +@@ -8,6 +8,7 @@ + #include <linux/fs.h> + #include <linux/namei.h> + #include <linux/slab.h> ++#include <linux/vs_context.h> + #include <asm/current.h> + #include <asm/uaccess.h> + #include <linux/kernel.h> +@@ -38,7 +39,7 @@ static int check_quotactl_permission(str + break; + /*FALLTHROUGH*/ + default: +- if (!capable(CAP_SYS_ADMIN)) ++ if (!vx_capable(CAP_SYS_ADMIN, VXC_QUOTA_CTL)) + return -EPERM; + } + +@@ -293,6 +294,46 @@ static int do_quotactl(struct super_bloc + } + } + ++#if defined(CONFIG_BLK_DEV_VROOT) || defined(CONFIG_BLK_DEV_VROOT_MODULE) ++ ++#include <linux/vroot.h> ++#include <linux/major.h> ++#include <linux/module.h> ++#include <linux/kallsyms.h> ++#include <linux/vserver/debug.h> ++ ++static vroot_grb_func *vroot_get_real_bdev = NULL; ++ ++static DEFINE_SPINLOCK(vroot_grb_lock); ++ ++int register_vroot_grb(vroot_grb_func *func) { ++ int ret = -EBUSY; ++ ++ spin_lock(&vroot_grb_lock); ++ if (!vroot_get_real_bdev) { ++ vroot_get_real_bdev = func; ++ ret = 0; ++ } ++ spin_unlock(&vroot_grb_lock); ++ return ret; ++} ++EXPORT_SYMBOL(register_vroot_grb); ++ ++int unregister_vroot_grb(vroot_grb_func *func) { ++ int ret = -EINVAL; ++ ++ spin_lock(&vroot_grb_lock); ++ if (vroot_get_real_bdev) { ++ vroot_get_real_bdev = NULL; ++ ret = 0; ++ } ++ spin_unlock(&vroot_grb_lock); ++ return ret; ++} ++EXPORT_SYMBOL(unregister_vroot_grb); ++ ++#endif ++ + /* + * look up a superblock on which quota ops will be performed + * - use the name of a block device to find the superblock thereon +@@ -310,6 +351,22 @@ static struct super_block *quotactl_bloc + putname(tmp); + if (IS_ERR(bdev)) + return ERR_CAST(bdev); ++#if defined(CONFIG_BLK_DEV_VROOT) || defined(CONFIG_BLK_DEV_VROOT_MODULE) ++ if (bdev && bdev->bd_inode && ++ imajor(bdev->bd_inode) == VROOT_MAJOR) { ++ struct block_device *bdnew = (void *)-EINVAL; ++ ++ if (vroot_get_real_bdev) ++ bdnew = vroot_get_real_bdev(bdev); ++ else ++ vxdprintk(VXD_CBIT(misc, 0), ++ "vroot_get_real_bdev not set"); ++ bdput(bdev); ++ if (IS_ERR(bdnew)) ++ return ERR_PTR(PTR_ERR(bdnew)); ++ bdev = bdnew; ++ } ++#endif + sb = get_super(bdev); + bdput(bdev); + if (!sb) +diff -NurpP --minimal linux-3.0.9/fs/reiserfs/file.c linux-3.0.9-vs2.3.2.1/fs/reiserfs/file.c +--- linux-3.0.9/fs/reiserfs/file.c 2011-01-05 21:50:26.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/fs/reiserfs/file.c 2011-06-10 22:11:24.000000000 +0200 +@@ -312,4 +312,5 @@ const struct inode_operations reiserfs_f + .listxattr = reiserfs_listxattr, + .removexattr = reiserfs_removexattr, + .permission = reiserfs_permission, ++ .sync_flags = reiserfs_sync_flags, + }; +diff -NurpP --minimal linux-3.0.9/fs/reiserfs/inode.c linux-3.0.9-vs2.3.2.1/fs/reiserfs/inode.c +--- linux-3.0.9/fs/reiserfs/inode.c 2011-05-22 16:17:53.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/fs/reiserfs/inode.c 2011-06-10 22:11:24.000000000 +0200 +@@ -18,6 +18,7 @@ + #include <linux/writeback.h> + #include <linux/quotaops.h> + #include <linux/swap.h> ++#include <linux/vs_tag.h> + + int reiserfs_commit_write(struct file *f, struct page *page, + unsigned from, unsigned to); +@@ -1131,6 +1132,8 @@ static void init_inode(struct inode *ino + struct buffer_head *bh; + struct item_head *ih; + __u32 rdev; ++ uid_t uid; ++ gid_t gid; + //int version = ITEM_VERSION_1; + + bh = PATH_PLAST_BUFFER(path); +@@ -1151,12 +1154,13 @@ static void init_inode(struct inode *ino + (struct stat_data_v1 *)B_I_PITEM(bh, ih); + unsigned long blocks; + ++ uid = sd_v1_uid(sd); ++ gid = sd_v1_gid(sd); ++ + set_inode_item_key_version(inode, KEY_FORMAT_3_5); + set_inode_sd_version(inode, STAT_DATA_V1); + inode->i_mode = sd_v1_mode(sd); + inode->i_nlink = sd_v1_nlink(sd); +- inode->i_uid = sd_v1_uid(sd); +- inode->i_gid = sd_v1_gid(sd); + inode->i_size = sd_v1_size(sd); + inode->i_atime.tv_sec = sd_v1_atime(sd); + inode->i_mtime.tv_sec = sd_v1_mtime(sd); +@@ -1198,11 +1202,12 @@ static void init_inode(struct inode *ino + // (directories and symlinks) + struct stat_data *sd = (struct stat_data *)B_I_PITEM(bh, ih); + ++ uid = sd_v2_uid(sd); ++ gid = sd_v2_gid(sd); ++ + inode->i_mode = sd_v2_mode(sd); + inode->i_nlink = sd_v2_nlink(sd); +- inode->i_uid = sd_v2_uid(sd); + inode->i_size = sd_v2_size(sd); +- inode->i_gid = sd_v2_gid(sd); + inode->i_mtime.tv_sec = sd_v2_mtime(sd); + inode->i_atime.tv_sec = sd_v2_atime(sd); + inode->i_ctime.tv_sec = sd_v2_ctime(sd); +@@ -1232,6 +1237,10 @@ static void init_inode(struct inode *ino + sd_attrs_to_i_attrs(sd_v2_attrs(sd), inode); + } + ++ inode->i_uid = INOTAG_UID(DX_TAG(inode), uid, gid); ++ inode->i_gid = INOTAG_GID(DX_TAG(inode), uid, gid); ++ inode->i_tag = INOTAG_TAG(DX_TAG(inode), uid, gid, 0); ++ + pathrelse(path); + if (S_ISREG(inode->i_mode)) { + inode->i_op = &reiserfs_file_inode_operations; +@@ -1254,13 +1263,15 @@ static void init_inode(struct inode *ino + static void inode2sd(void *sd, struct inode *inode, loff_t size) + { + struct stat_data *sd_v2 = (struct stat_data *)sd; ++ uid_t uid = TAGINO_UID(DX_TAG(inode), inode->i_uid, inode->i_tag); ++ gid_t gid = TAGINO_GID(DX_TAG(inode), inode->i_gid, inode->i_tag); + __u16 flags; + ++ set_sd_v2_uid(sd_v2, uid); ++ set_sd_v2_gid(sd_v2, gid); + set_sd_v2_mode(sd_v2, inode->i_mode); + set_sd_v2_nlink(sd_v2, inode->i_nlink); +- set_sd_v2_uid(sd_v2, inode->i_uid); + set_sd_v2_size(sd_v2, size); +- set_sd_v2_gid(sd_v2, inode->i_gid); + set_sd_v2_mtime(sd_v2, inode->i_mtime.tv_sec); + set_sd_v2_atime(sd_v2, inode->i_atime.tv_sec); + set_sd_v2_ctime(sd_v2, inode->i_ctime.tv_sec); +@@ -2863,14 +2874,19 @@ int reiserfs_commit_write(struct file *f + void sd_attrs_to_i_attrs(__u16 sd_attrs, struct inode *inode) + { + if (reiserfs_attrs(inode->i_sb)) { +- if (sd_attrs & REISERFS_SYNC_FL) +- inode->i_flags |= S_SYNC; +- else +- inode->i_flags &= ~S_SYNC; + if (sd_attrs & REISERFS_IMMUTABLE_FL) + inode->i_flags |= S_IMMUTABLE; + else + inode->i_flags &= ~S_IMMUTABLE; ++ if (sd_attrs & REISERFS_IXUNLINK_FL) ++ inode->i_flags |= S_IXUNLINK; ++ else ++ inode->i_flags &= ~S_IXUNLINK; ++ ++ if (sd_attrs & REISERFS_SYNC_FL) ++ inode->i_flags |= S_SYNC; ++ else ++ inode->i_flags &= ~S_SYNC; + if (sd_attrs & REISERFS_APPEND_FL) + inode->i_flags |= S_APPEND; + else +@@ -2883,6 +2899,15 @@ void sd_attrs_to_i_attrs(__u16 sd_attrs, + REISERFS_I(inode)->i_flags |= i_nopack_mask; + else + REISERFS_I(inode)->i_flags &= ~i_nopack_mask; ++ ++ if (sd_attrs & REISERFS_BARRIER_FL) ++ inode->i_vflags |= V_BARRIER; ++ else ++ inode->i_vflags &= ~V_BARRIER; ++ if (sd_attrs & REISERFS_COW_FL) ++ inode->i_vflags |= V_COW; ++ else ++ inode->i_vflags &= ~V_COW; + } + } + +@@ -2893,6 +2918,11 @@ void i_attrs_to_sd_attrs(struct inode *i + *sd_attrs |= REISERFS_IMMUTABLE_FL; + else + *sd_attrs &= ~REISERFS_IMMUTABLE_FL; ++ if (inode->i_flags & S_IXUNLINK) ++ *sd_attrs |= REISERFS_IXUNLINK_FL; ++ else ++ *sd_attrs &= ~REISERFS_IXUNLINK_FL; ++ + if (inode->i_flags & S_SYNC) + *sd_attrs |= REISERFS_SYNC_FL; + else +@@ -2905,6 +2935,15 @@ void i_attrs_to_sd_attrs(struct inode *i + *sd_attrs |= REISERFS_NOTAIL_FL; + else + *sd_attrs &= ~REISERFS_NOTAIL_FL; ++ ++ if (inode->i_vflags & V_BARRIER) ++ *sd_attrs |= REISERFS_BARRIER_FL; ++ else ++ *sd_attrs &= ~REISERFS_BARRIER_FL; ++ if (inode->i_vflags & V_COW) ++ *sd_attrs |= REISERFS_COW_FL; ++ else ++ *sd_attrs &= ~REISERFS_COW_FL; + } + } + +@@ -3148,7 +3187,8 @@ int reiserfs_setattr(struct dentry *dent + } + + if ((ia_valid & ATTR_UID && attr->ia_uid != inode->i_uid) || +- (ia_valid & ATTR_GID && attr->ia_gid != inode->i_gid)) { ++ (ia_valid & ATTR_GID && attr->ia_gid != inode->i_gid) || ++ (ia_valid & ATTR_TAG && attr->ia_tag != inode->i_tag)) { + struct reiserfs_transaction_handle th; + int jbegin_count = + 2 * +@@ -3177,6 +3217,9 @@ int reiserfs_setattr(struct dentry *dent + inode->i_uid = attr->ia_uid; + if (attr->ia_valid & ATTR_GID) + inode->i_gid = attr->ia_gid; ++ if ((attr->ia_valid & ATTR_TAG) && ++ IS_TAGGED(inode)) ++ inode->i_tag = attr->ia_tag; + mark_inode_dirty(inode); + error = journal_end(&th, inode->i_sb, jbegin_count); + if (error) +diff -NurpP --minimal linux-3.0.9/fs/reiserfs/ioctl.c linux-3.0.9-vs2.3.2.1/fs/reiserfs/ioctl.c +--- linux-3.0.9/fs/reiserfs/ioctl.c 2011-05-22 16:17:53.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/fs/reiserfs/ioctl.c 2011-06-10 22:11:24.000000000 +0200 +@@ -11,6 +11,21 @@ + #include <linux/pagemap.h> + #include <linux/compat.h> + ++ ++int reiserfs_sync_flags(struct inode *inode, int flags, int vflags) ++{ ++ __u16 sd_attrs = 0; ++ ++ inode->i_flags = flags; ++ inode->i_vflags = vflags; ++ ++ i_attrs_to_sd_attrs(inode, &sd_attrs); ++ REISERFS_I(inode)->i_attrs = sd_attrs; ++ inode->i_ctime = CURRENT_TIME_SEC; ++ mark_inode_dirty(inode); ++ return 0; ++} ++ + /* + * reiserfs_ioctl - handler for ioctl for inode + * supported commands: +@@ -22,7 +37,7 @@ + long reiserfs_ioctl(struct file *filp, unsigned int cmd, unsigned long arg) + { + struct inode *inode = filp->f_path.dentry->d_inode; +- unsigned int flags; ++ unsigned int flags, oldflags; + int err = 0; + + reiserfs_write_lock(inode->i_sb); +@@ -47,6 +62,7 @@ long reiserfs_ioctl(struct file *filp, u + + flags = REISERFS_I(inode)->i_attrs; + i_attrs_to_sd_attrs(inode, (__u16 *) & flags); ++ flags &= REISERFS_FL_USER_VISIBLE; + err = put_user(flags, (int __user *)arg); + break; + case REISERFS_IOC_SETFLAGS:{ +@@ -67,6 +83,10 @@ long reiserfs_ioctl(struct file *filp, u + err = -EFAULT; + goto setflags_out; + } ++ if (IS_BARRIER(inode)) { ++ vxwprintk_task(1, "messing with the barrier."); ++ return -EACCES; ++ } + /* + * Is it quota file? Do not allow user to mess with it + */ +@@ -91,6 +111,10 @@ long reiserfs_ioctl(struct file *filp, u + goto setflags_out; + } + } ++ ++ oldflags = REISERFS_I(inode)->i_attrs; ++ flags &= REISERFS_FL_USER_MODIFIABLE; ++ flags |= oldflags & ~REISERFS_FL_USER_MODIFIABLE; + sd_attrs_to_i_attrs(flags, inode); + REISERFS_I(inode)->i_attrs = flags; + inode->i_ctime = CURRENT_TIME_SEC; +diff -NurpP --minimal linux-3.0.9/fs/reiserfs/namei.c linux-3.0.9-vs2.3.2.1/fs/reiserfs/namei.c +--- linux-3.0.9/fs/reiserfs/namei.c 2011-05-22 16:17:53.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/fs/reiserfs/namei.c 2011-06-10 22:11:24.000000000 +0200 +@@ -18,6 +18,7 @@ + #include <linux/reiserfs_acl.h> + #include <linux/reiserfs_xattr.h> + #include <linux/quotaops.h> ++#include <linux/vs_tag.h> + + #define INC_DIR_INODE_NLINK(i) if (i->i_nlink != 1) { inc_nlink(i); if (i->i_nlink >= REISERFS_LINK_MAX) i->i_nlink=1; } + #define DEC_DIR_INODE_NLINK(i) if (i->i_nlink != 1) drop_nlink(i); +@@ -362,6 +363,7 @@ static struct dentry *reiserfs_lookup(st + if (retval == IO_ERROR) { + return ERR_PTR(-EIO); + } ++ dx_propagate_tag(nd, inode); + + return d_splice_alias(inode, dentry); + } +@@ -1529,6 +1531,7 @@ const struct inode_operations reiserfs_d + .listxattr = reiserfs_listxattr, + .removexattr = reiserfs_removexattr, + .permission = reiserfs_permission, ++ .sync_flags = reiserfs_sync_flags, + }; + + /* +diff -NurpP --minimal linux-3.0.9/fs/reiserfs/super.c linux-3.0.9-vs2.3.2.1/fs/reiserfs/super.c +--- linux-3.0.9/fs/reiserfs/super.c 2011-07-22 11:18:06.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/fs/reiserfs/super.c 2011-06-10 22:11:24.000000000 +0200 +@@ -899,6 +899,14 @@ static int reiserfs_parse_options(struct + {"user_xattr",.setmask = 1 << REISERFS_UNSUPPORTED_OPT}, + {"nouser_xattr",.clrmask = 1 << REISERFS_UNSUPPORTED_OPT}, + #endif ++#ifndef CONFIG_TAGGING_NONE ++ {"tagxid",.setmask = 1 << REISERFS_TAGGED}, ++ {"tag",.setmask = 1 << REISERFS_TAGGED}, ++ {"notag",.clrmask = 1 << REISERFS_TAGGED}, ++#endif ++#ifdef CONFIG_PROPAGATE ++ {"tag",.arg_required = 'T',.values = NULL}, ++#endif + #ifdef CONFIG_REISERFS_FS_POSIX_ACL + {"acl",.setmask = 1 << REISERFS_POSIXACL}, + {"noacl",.clrmask = 1 << REISERFS_POSIXACL}, +@@ -1208,6 +1216,14 @@ static int reiserfs_remount(struct super + handle_quota_files(s, qf_names, &qfmt); + #endif + ++ if ((mount_options & (1 << REISERFS_TAGGED)) && ++ !(s->s_flags & MS_TAGGED)) { ++ reiserfs_warning(s, "super-vs01", ++ "reiserfs: tagging not permitted on remount."); ++ err = -EINVAL; ++ goto out_err; ++ } ++ + handle_attrs(s); + + /* Add options that are safe here */ +@@ -1690,6 +1706,10 @@ static int reiserfs_fill_super(struct su + goto error; + } + ++ /* map mount option tagxid */ ++ if (REISERFS_SB(s)->s_mount_opt & (1 << REISERFS_TAGGED)) ++ s->s_flags |= MS_TAGGED; ++ + rs = SB_DISK_SUPER_BLOCK(s); + /* Let's do basic sanity check to verify that underlying device is not + smaller than the filesystem. If the check fails then abort and scream, +diff -NurpP --minimal linux-3.0.9/fs/reiserfs/xattr.c linux-3.0.9-vs2.3.2.1/fs/reiserfs/xattr.c +--- linux-3.0.9/fs/reiserfs/xattr.c 2011-07-22 11:18:06.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/fs/reiserfs/xattr.c 2011-06-22 12:39:15.000000000 +0200 +@@ -40,6 +40,7 @@ + #include <linux/errno.h> + #include <linux/gfp.h> + #include <linux/fs.h> ++#include <linux/mount.h> + #include <linux/file.h> + #include <linux/pagemap.h> + #include <linux/xattr.h> +diff -NurpP --minimal linux-3.0.9/fs/stat.c linux-3.0.9-vs2.3.2.1/fs/stat.c +--- linux-3.0.9/fs/stat.c 2011-11-15 16:40:47.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/fs/stat.c 2011-11-15 17:37:07.000000000 +0100 +@@ -26,6 +26,7 @@ void generic_fillattr(struct inode *inod + stat->nlink = inode->i_nlink; + stat->uid = inode->i_uid; + stat->gid = inode->i_gid; ++ stat->tag = inode->i_tag; + stat->rdev = inode->i_rdev; + stat->atime = inode->i_atime; + stat->mtime = inode->i_mtime; +diff -NurpP --minimal linux-3.0.9/fs/statfs.c linux-3.0.9-vs2.3.2.1/fs/statfs.c +--- linux-3.0.9/fs/statfs.c 2011-11-15 16:40:47.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/fs/statfs.c 2011-11-15 17:37:07.000000000 +0100 +@@ -7,6 +7,8 @@ + #include <linux/statfs.h> + #include <linux/security.h> + #include <linux/uaccess.h> ++#include <linux/vs_base.h> ++#include <linux/vs_dlimit.h> + + static int flags_by_mnt(int mnt_flags) + { +@@ -59,6 +61,8 @@ int statfs_by_dentry(struct dentry *dent + retval = dentry->d_sb->s_op->statfs(dentry, buf); + if (retval == 0 && buf->f_frsize == 0) + buf->f_frsize = buf->f_bsize; ++ if (!vx_check(0, VS_ADMIN|VS_WATCH)) ++ vx_vsi_statfs(dentry->d_sb, buf); + return retval; + } + +diff -NurpP --minimal linux-3.0.9/fs/super.c linux-3.0.9-vs2.3.2.1/fs/super.c +--- linux-3.0.9/fs/super.c 2011-07-22 11:18:06.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/fs/super.c 2011-07-23 16:49:09.000000000 +0200 +@@ -32,6 +32,9 @@ + #include <linux/backing-dev.h> + #include <linux/rculist_bl.h> + #include <linux/cleancache.h> ++#include <linux/devpts_fs.h> ++#include <linux/proc_fs.h> ++#include <linux/vs_context.h> + #include "internal.h" + + +@@ -943,6 +946,13 @@ mount_fs(struct file_system_type *type, + WARN_ON(sb->s_bdi == &default_backing_dev_info); + sb->s_flags |= MS_BORN; + ++ error = -EPERM; ++ if (!vx_capable(CAP_SYS_ADMIN, VXC_BINARY_MOUNT) && ++ !sb->s_bdev && ++ (sb->s_magic != PROC_SUPER_MAGIC) && ++ (sb->s_magic != DEVPTS_SUPER_MAGIC)) ++ goto out_sb; ++ + error = security_sb_kern_mount(sb, flags, secdata); + if (error) + goto out_sb; +diff -NurpP --minimal linux-3.0.9/fs/sysfs/mount.c linux-3.0.9-vs2.3.2.1/fs/sysfs/mount.c +--- linux-3.0.9/fs/sysfs/mount.c 2011-07-22 11:18:06.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/fs/sysfs/mount.c 2011-06-22 12:39:15.000000000 +0200 +@@ -47,7 +47,7 @@ static int sysfs_fill_super(struct super + + sb->s_blocksize = PAGE_CACHE_SIZE; + sb->s_blocksize_bits = PAGE_CACHE_SHIFT; +- sb->s_magic = SYSFS_MAGIC; ++ sb->s_magic = SYSFS_SUPER_MAGIC; + sb->s_op = &sysfs_ops; + sb->s_time_gran = 1; + +diff -NurpP --minimal linux-3.0.9/fs/utimes.c linux-3.0.9-vs2.3.2.1/fs/utimes.c +--- linux-3.0.9/fs/utimes.c 2011-05-22 16:17:54.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/fs/utimes.c 2011-06-10 22:11:24.000000000 +0200 +@@ -8,6 +8,8 @@ + #include <linux/stat.h> + #include <linux/utime.h> + #include <linux/syscalls.h> ++#include <linux/mount.h> ++#include <linux/vs_cowbl.h> + #include <asm/uaccess.h> + #include <asm/unistd.h> + +@@ -52,12 +54,18 @@ static int utimes_common(struct path *pa + { + int error; + struct iattr newattrs; +- struct inode *inode = path->dentry->d_inode; ++ struct inode *inode; + + error = mnt_want_write(path->mnt); + if (error) + goto out; + ++ error = cow_check_and_break(path); ++ if (error) ++ goto mnt_drop_write_and_out; ++ ++ inode = path->dentry->d_inode; ++ + if (times && times[0].tv_nsec == UTIME_NOW && + times[1].tv_nsec == UTIME_NOW) + times = NULL; +diff -NurpP --minimal linux-3.0.9/fs/xattr.c linux-3.0.9-vs2.3.2.1/fs/xattr.c +--- linux-3.0.9/fs/xattr.c 2011-07-22 11:18:09.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/fs/xattr.c 2011-06-10 23:10:19.000000000 +0200 +@@ -18,6 +18,7 @@ + #include <linux/module.h> + #include <linux/fsnotify.h> + #include <linux/audit.h> ++#include <linux/mount.h> + #include <asm/uaccess.h> + + +@@ -49,7 +50,7 @@ xattr_permission(struct inode *inode, co + * The trusted.* namespace can only be accessed by privileged users. + */ + if (!strncmp(name, XATTR_TRUSTED_PREFIX, XATTR_TRUSTED_PREFIX_LEN)) { +- if (!capable(CAP_SYS_ADMIN)) ++ if (!vx_capable(CAP_SYS_ADMIN, VXC_FS_TRUSTED)) + return (mask & MAY_WRITE) ? -EPERM : -ENODATA; + return 0; + } +diff -NurpP --minimal linux-3.0.9/fs/xfs/linux-2.6/xfs_ioctl.c linux-3.0.9-vs2.3.2.1/fs/xfs/linux-2.6/xfs_ioctl.c +--- linux-3.0.9/fs/xfs/linux-2.6/xfs_ioctl.c 2011-05-22 16:17:54.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/fs/xfs/linux-2.6/xfs_ioctl.c 2011-06-10 22:11:24.000000000 +0200 +@@ -28,7 +28,7 @@ + #include "xfs_bmap_btree.h" + #include "xfs_dinode.h" + #include "xfs_inode.h" +-#include "xfs_ioctl.h" ++// #include "xfs_ioctl.h" + #include "xfs_rtalloc.h" + #include "xfs_itable.h" + #include "xfs_error.h" +@@ -748,6 +748,10 @@ xfs_merge_ioc_xflags( + xflags |= XFS_XFLAG_IMMUTABLE; + else + xflags &= ~XFS_XFLAG_IMMUTABLE; ++ if (flags & FS_IXUNLINK_FL) ++ xflags |= XFS_XFLAG_IXUNLINK; ++ else ++ xflags &= ~XFS_XFLAG_IXUNLINK; + if (flags & FS_APPEND_FL) + xflags |= XFS_XFLAG_APPEND; + else +@@ -776,6 +780,8 @@ xfs_di2lxflags( + + if (di_flags & XFS_DIFLAG_IMMUTABLE) + flags |= FS_IMMUTABLE_FL; ++ if (di_flags & XFS_DIFLAG_IXUNLINK) ++ flags |= FS_IXUNLINK_FL; + if (di_flags & XFS_DIFLAG_APPEND) + flags |= FS_APPEND_FL; + if (di_flags & XFS_DIFLAG_SYNC) +@@ -836,6 +842,8 @@ xfs_set_diflags( + di_flags = (ip->i_d.di_flags & XFS_DIFLAG_PREALLOC); + if (xflags & XFS_XFLAG_IMMUTABLE) + di_flags |= XFS_DIFLAG_IMMUTABLE; ++ if (xflags & XFS_XFLAG_IXUNLINK) ++ di_flags |= XFS_DIFLAG_IXUNLINK; + if (xflags & XFS_XFLAG_APPEND) + di_flags |= XFS_DIFLAG_APPEND; + if (xflags & XFS_XFLAG_SYNC) +@@ -878,6 +886,10 @@ xfs_diflags_to_linux( + inode->i_flags |= S_IMMUTABLE; + else + inode->i_flags &= ~S_IMMUTABLE; ++ if (xflags & XFS_XFLAG_IXUNLINK) ++ inode->i_flags |= S_IXUNLINK; ++ else ++ inode->i_flags &= ~S_IXUNLINK; + if (xflags & XFS_XFLAG_APPEND) + inode->i_flags |= S_APPEND; + else +@@ -1370,10 +1382,18 @@ xfs_file_ioctl( + case XFS_IOC_FSGETXATTRA: + return xfs_ioc_fsgetxattr(ip, 1, arg); + case XFS_IOC_FSSETXATTR: ++ if (IS_BARRIER(inode)) { ++ vxwprintk_task(1, "messing with the barrier."); ++ return -XFS_ERROR(EACCES); ++ } + return xfs_ioc_fssetxattr(ip, filp, arg); + case XFS_IOC_GETXFLAGS: + return xfs_ioc_getxflags(ip, arg); + case XFS_IOC_SETXFLAGS: ++ if (IS_BARRIER(inode)) { ++ vxwprintk_task(1, "messing with the barrier."); ++ return -XFS_ERROR(EACCES); ++ } + return xfs_ioc_setxflags(ip, filp, arg); + + case XFS_IOC_FSSETDM: { +diff -NurpP --minimal linux-3.0.9/fs/xfs/linux-2.6/xfs_ioctl.h linux-3.0.9-vs2.3.2.1/fs/xfs/linux-2.6/xfs_ioctl.h +--- linux-3.0.9/fs/xfs/linux-2.6/xfs_ioctl.h 2010-07-07 18:31:54.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/fs/xfs/linux-2.6/xfs_ioctl.h 2011-06-10 22:11:24.000000000 +0200 +@@ -70,6 +70,12 @@ xfs_handle_to_dentry( + void __user *uhandle, + u32 hlen); + ++extern int ++xfs_sync_flags( ++ struct inode *inode, ++ int flags, ++ int vflags); ++ + extern long + xfs_file_ioctl( + struct file *filp, +diff -NurpP --minimal linux-3.0.9/fs/xfs/linux-2.6/xfs_iops.c linux-3.0.9-vs2.3.2.1/fs/xfs/linux-2.6/xfs_iops.c +--- linux-3.0.9/fs/xfs/linux-2.6/xfs_iops.c 2011-07-22 11:18:09.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/fs/xfs/linux-2.6/xfs_iops.c 2011-06-22 12:39:15.000000000 +0200 +@@ -30,6 +30,7 @@ + #include "xfs_bmap_btree.h" + #include "xfs_dinode.h" + #include "xfs_inode.h" ++#include "xfs_ioctl.h" + #include "xfs_bmap.h" + #include "xfs_rtalloc.h" + #include "xfs_error.h" +@@ -48,6 +49,7 @@ + #include <linux/security.h> + #include <linux/fiemap.h> + #include <linux/slab.h> ++#include <linux/vs_tag.h> + + /* + * Bring the timestamps in the XFS inode uptodate. +@@ -464,6 +466,7 @@ xfs_vn_getattr( + stat->nlink = ip->i_d.di_nlink; + stat->uid = ip->i_d.di_uid; + stat->gid = ip->i_d.di_gid; ++ stat->tag = ip->i_d.di_tag; + stat->ino = ip->i_ino; + stat->atime = inode->i_atime; + stat->mtime = inode->i_mtime; +@@ -599,6 +602,7 @@ static const struct inode_operations xfs + .removexattr = generic_removexattr, + .listxattr = xfs_vn_listxattr, + .fiemap = xfs_vn_fiemap, ++ .sync_flags = xfs_sync_flags, + }; + + static const struct inode_operations xfs_dir_inode_operations = { +@@ -624,6 +628,7 @@ static const struct inode_operations xfs + .getxattr = generic_getxattr, + .removexattr = generic_removexattr, + .listxattr = xfs_vn_listxattr, ++ .sync_flags = xfs_sync_flags, + }; + + static const struct inode_operations xfs_dir_ci_inode_operations = { +@@ -673,6 +678,10 @@ xfs_diflags_to_iflags( + inode->i_flags |= S_IMMUTABLE; + else + inode->i_flags &= ~S_IMMUTABLE; ++ if (ip->i_d.di_flags & XFS_DIFLAG_IXUNLINK) ++ inode->i_flags |= S_IXUNLINK; ++ else ++ inode->i_flags &= ~S_IXUNLINK; + if (ip->i_d.di_flags & XFS_DIFLAG_APPEND) + inode->i_flags |= S_APPEND; + else +@@ -685,6 +694,15 @@ xfs_diflags_to_iflags( + inode->i_flags |= S_NOATIME; + else + inode->i_flags &= ~S_NOATIME; ++ ++ if (ip->i_d.di_vflags & XFS_DIVFLAG_BARRIER) ++ inode->i_vflags |= V_BARRIER; ++ else ++ inode->i_vflags &= ~V_BARRIER; ++ if (ip->i_d.di_vflags & XFS_DIVFLAG_COW) ++ inode->i_vflags |= V_COW; ++ else ++ inode->i_vflags &= ~V_COW; + } + + /* +@@ -716,6 +734,7 @@ xfs_setup_inode( + inode->i_nlink = ip->i_d.di_nlink; + inode->i_uid = ip->i_d.di_uid; + inode->i_gid = ip->i_d.di_gid; ++ inode->i_tag = ip->i_d.di_tag; + + switch (inode->i_mode & S_IFMT) { + case S_IFBLK: +diff -NurpP --minimal linux-3.0.9/fs/xfs/linux-2.6/xfs_linux.h linux-3.0.9-vs2.3.2.1/fs/xfs/linux-2.6/xfs_linux.h +--- linux-3.0.9/fs/xfs/linux-2.6/xfs_linux.h 2011-11-15 16:40:47.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/fs/xfs/linux-2.6/xfs_linux.h 2011-11-15 17:37:07.000000000 +0100 +@@ -117,6 +117,7 @@ + + #define current_cpu() (raw_smp_processor_id()) + #define current_pid() (current->pid) ++#define current_fstag(vp) (dx_current_fstag((vp)->i_sb)) + #define current_test_flags(f) (current->flags & (f)) + #define current_set_flags_nested(sp, f) \ + (*(sp) = current->flags, current->flags |= (f)) +diff -NurpP --minimal linux-3.0.9/fs/xfs/linux-2.6/xfs_super.c linux-3.0.9-vs2.3.2.1/fs/xfs/linux-2.6/xfs_super.c +--- linux-3.0.9/fs/xfs/linux-2.6/xfs_super.c 2011-11-15 16:40:47.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/fs/xfs/linux-2.6/xfs_super.c 2011-11-15 17:37:07.000000000 +0100 +@@ -114,6 +114,9 @@ mempool_t *xfs_ioend_pool; + #define MNTOPT_NODELAYLOG "nodelaylog" /* Delayed logging disabled */ + #define MNTOPT_DISCARD "discard" /* Discard unused blocks */ + #define MNTOPT_NODISCARD "nodiscard" /* Do not discard unused blocks */ ++#define MNTOPT_TAGXID "tagxid" /* context tagging for inodes */ ++#define MNTOPT_TAGGED "tag" /* context tagging for inodes */ ++#define MNTOPT_NOTAGTAG "notag" /* do not use context tagging */ + + /* + * Table driven mount option parser. +@@ -122,10 +125,14 @@ mempool_t *xfs_ioend_pool; + * in the future, too. + */ + enum { ++ Opt_tag, Opt_notag, + Opt_barrier, Opt_nobarrier, Opt_err + }; + + static const match_table_t tokens = { ++ {Opt_tag, "tagxid"}, ++ {Opt_tag, "tag"}, ++ {Opt_notag, "notag"}, + {Opt_barrier, "barrier"}, + {Opt_nobarrier, "nobarrier"}, + {Opt_err, NULL} +@@ -373,6 +380,19 @@ xfs_parseargs( + } else if (!strcmp(this_char, "irixsgid")) { + xfs_warn(mp, + "irixsgid is now a sysctl(2) variable, option is deprecated."); ++#ifndef CONFIG_TAGGING_NONE ++ } else if (!strcmp(this_char, MNTOPT_TAGGED)) { ++ mp->m_flags |= XFS_MOUNT_TAGGED; ++ } else if (!strcmp(this_char, MNTOPT_NOTAGTAG)) { ++ mp->m_flags &= ~XFS_MOUNT_TAGGED; ++ } else if (!strcmp(this_char, MNTOPT_TAGXID)) { ++ mp->m_flags |= XFS_MOUNT_TAGGED; ++#endif ++#ifdef CONFIG_PROPAGATE ++ } else if (!strcmp(this_char, MNTOPT_TAGGED)) { ++ /* use value */ ++ mp->m_flags |= XFS_MOUNT_TAGGED; ++#endif + } else { + xfs_warn(mp, "unknown mount option [%s].", this_char); + return EINVAL; +@@ -1182,6 +1202,16 @@ xfs_fs_remount( + case Opt_nobarrier: + mp->m_flags &= ~XFS_MOUNT_BARRIER; + break; ++ case Opt_tag: ++ if (!(sb->s_flags & MS_TAGGED)) { ++ printk(KERN_INFO ++ "XFS: %s: tagging not permitted on remount.\n", ++ sb->s_id); ++ return -EINVAL; ++ } ++ break; ++ case Opt_notag: ++ break; + default: + /* + * Logically we would return an error here to prevent +@@ -1397,6 +1427,9 @@ xfs_fs_fill_super( + if (error) + goto out_free_sb; + ++ if (mp->m_flags & XFS_MOUNT_TAGGED) ++ sb->s_flags |= MS_TAGGED; ++ + /* + * we must configure the block size in the superblock before we run the + * full mount process as the mount process can lookup and cache inodes. +diff -NurpP --minimal linux-3.0.9/fs/xfs/xfs_dinode.h linux-3.0.9-vs2.3.2.1/fs/xfs/xfs_dinode.h +--- linux-3.0.9/fs/xfs/xfs_dinode.h 2011-01-05 21:50:28.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/fs/xfs/xfs_dinode.h 2011-06-10 22:11:24.000000000 +0200 +@@ -51,7 +51,9 @@ typedef struct xfs_dinode { + __be32 di_nlink; /* number of links to file */ + __be16 di_projid_lo; /* lower part of owner's project id */ + __be16 di_projid_hi; /* higher part owner's project id */ +- __u8 di_pad[6]; /* unused, zeroed space */ ++ __u8 di_pad[2]; /* unused, zeroed space */ ++ __be16 di_tag; /* context tagging */ ++ __be16 di_vflags; /* vserver specific flags */ + __be16 di_flushiter; /* incremented on flush */ + xfs_timestamp_t di_atime; /* time last accessed */ + xfs_timestamp_t di_mtime; /* time last modified */ +@@ -184,6 +186,8 @@ static inline void xfs_dinode_put_rdev(s + #define XFS_DIFLAG_EXTSZINHERIT_BIT 12 /* inherit inode extent size */ + #define XFS_DIFLAG_NODEFRAG_BIT 13 /* do not reorganize/defragment */ + #define XFS_DIFLAG_FILESTREAM_BIT 14 /* use filestream allocator */ ++#define XFS_DIFLAG_IXUNLINK_BIT 15 /* Immutable inver on unlink */ ++ + #define XFS_DIFLAG_REALTIME (1 << XFS_DIFLAG_REALTIME_BIT) + #define XFS_DIFLAG_PREALLOC (1 << XFS_DIFLAG_PREALLOC_BIT) + #define XFS_DIFLAG_NEWRTBM (1 << XFS_DIFLAG_NEWRTBM_BIT) +@@ -199,6 +203,7 @@ static inline void xfs_dinode_put_rdev(s + #define XFS_DIFLAG_EXTSZINHERIT (1 << XFS_DIFLAG_EXTSZINHERIT_BIT) + #define XFS_DIFLAG_NODEFRAG (1 << XFS_DIFLAG_NODEFRAG_BIT) + #define XFS_DIFLAG_FILESTREAM (1 << XFS_DIFLAG_FILESTREAM_BIT) ++#define XFS_DIFLAG_IXUNLINK (1 << XFS_DIFLAG_IXUNLINK_BIT) + + #ifdef CONFIG_XFS_RT + #define XFS_IS_REALTIME_INODE(ip) ((ip)->i_d.di_flags & XFS_DIFLAG_REALTIME) +@@ -211,6 +216,10 @@ static inline void xfs_dinode_put_rdev(s + XFS_DIFLAG_IMMUTABLE | XFS_DIFLAG_APPEND | XFS_DIFLAG_SYNC | \ + XFS_DIFLAG_NOATIME | XFS_DIFLAG_NODUMP | XFS_DIFLAG_RTINHERIT | \ + XFS_DIFLAG_PROJINHERIT | XFS_DIFLAG_NOSYMLINKS | XFS_DIFLAG_EXTSIZE | \ +- XFS_DIFLAG_EXTSZINHERIT | XFS_DIFLAG_NODEFRAG | XFS_DIFLAG_FILESTREAM) ++ XFS_DIFLAG_EXTSZINHERIT | XFS_DIFLAG_NODEFRAG | XFS_DIFLAG_FILESTREAM | \ ++ XFS_DIFLAG_IXUNLINK) ++ ++#define XFS_DIVFLAG_BARRIER 0x01 ++#define XFS_DIVFLAG_COW 0x02 + + #endif /* __XFS_DINODE_H__ */ +diff -NurpP --minimal linux-3.0.9/fs/xfs/xfs_fs.h linux-3.0.9-vs2.3.2.1/fs/xfs/xfs_fs.h +--- linux-3.0.9/fs/xfs/xfs_fs.h 2011-01-05 21:50:28.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/fs/xfs/xfs_fs.h 2011-06-10 22:11:24.000000000 +0200 +@@ -67,6 +67,9 @@ struct fsxattr { + #define XFS_XFLAG_EXTSZINHERIT 0x00001000 /* inherit inode extent size */ + #define XFS_XFLAG_NODEFRAG 0x00002000 /* do not defragment */ + #define XFS_XFLAG_FILESTREAM 0x00004000 /* use filestream allocator */ ++#define XFS_XFLAG_IXUNLINK 0x00008000 /* immutable invert on unlink */ ++#define XFS_XFLAG_BARRIER 0x10000000 /* chroot() barrier */ ++#define XFS_XFLAG_COW 0x20000000 /* copy on write mark */ + #define XFS_XFLAG_HASATTR 0x80000000 /* no DIFLAG for this */ + + /* +@@ -297,7 +300,8 @@ typedef struct xfs_bstat { + #define bs_projid bs_projid_lo /* (previously just bs_projid) */ + __u16 bs_forkoff; /* inode fork offset in bytes */ + __u16 bs_projid_hi; /* higher part of project id */ +- unsigned char bs_pad[10]; /* pad space, unused */ ++ unsigned char bs_pad[8]; /* pad space, unused */ ++ __u16 bs_tag; /* context tagging */ + __u32 bs_dmevmask; /* DMIG event mask */ + __u16 bs_dmstate; /* DMIG state info */ + __u16 bs_aextents; /* attribute number of extents */ +diff -NurpP --minimal linux-3.0.9/fs/xfs/xfs_ialloc.c linux-3.0.9-vs2.3.2.1/fs/xfs/xfs_ialloc.c +--- linux-3.0.9/fs/xfs/xfs_ialloc.c 2011-05-22 16:17:54.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/fs/xfs/xfs_ialloc.c 2011-06-10 22:11:24.000000000 +0200 +@@ -37,7 +37,6 @@ + #include "xfs_error.h" + #include "xfs_bmap.h" + +- + /* + * Allocation group level functions. + */ +diff -NurpP --minimal linux-3.0.9/fs/xfs/xfs_inode.c linux-3.0.9-vs2.3.2.1/fs/xfs/xfs_inode.c +--- linux-3.0.9/fs/xfs/xfs_inode.c 2011-07-22 11:18:10.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/fs/xfs/xfs_inode.c 2011-06-10 22:11:24.000000000 +0200 +@@ -243,6 +243,7 @@ xfs_inotobp( + return 0; + } + ++#include <linux/vs_tag.h> + + /* + * This routine is called to map an inode to the buffer containing +@@ -641,15 +642,25 @@ xfs_iformat_btree( + STATIC void + xfs_dinode_from_disk( + xfs_icdinode_t *to, +- xfs_dinode_t *from) ++ xfs_dinode_t *from, ++ int tagged) + { ++ uint32_t uid, gid, tag; ++ + to->di_magic = be16_to_cpu(from->di_magic); + to->di_mode = be16_to_cpu(from->di_mode); + to->di_version = from ->di_version; + to->di_format = from->di_format; + to->di_onlink = be16_to_cpu(from->di_onlink); +- to->di_uid = be32_to_cpu(from->di_uid); +- to->di_gid = be32_to_cpu(from->di_gid); ++ ++ uid = be32_to_cpu(from->di_uid); ++ gid = be32_to_cpu(from->di_gid); ++ tag = be16_to_cpu(from->di_tag); ++ ++ to->di_uid = INOTAG_UID(tagged, uid, gid); ++ to->di_gid = INOTAG_GID(tagged, uid, gid); ++ to->di_tag = INOTAG_TAG(tagged, uid, gid, tag); ++ + to->di_nlink = be32_to_cpu(from->di_nlink); + to->di_projid_lo = be16_to_cpu(from->di_projid_lo); + to->di_projid_hi = be16_to_cpu(from->di_projid_hi); +@@ -671,21 +682,26 @@ xfs_dinode_from_disk( + to->di_dmevmask = be32_to_cpu(from->di_dmevmask); + to->di_dmstate = be16_to_cpu(from->di_dmstate); + to->di_flags = be16_to_cpu(from->di_flags); ++ to->di_vflags = be16_to_cpu(from->di_vflags); + to->di_gen = be32_to_cpu(from->di_gen); + } + + void + xfs_dinode_to_disk( + xfs_dinode_t *to, +- xfs_icdinode_t *from) ++ xfs_icdinode_t *from, ++ int tagged) + { + to->di_magic = cpu_to_be16(from->di_magic); + to->di_mode = cpu_to_be16(from->di_mode); + to->di_version = from ->di_version; + to->di_format = from->di_format; + to->di_onlink = cpu_to_be16(from->di_onlink); +- to->di_uid = cpu_to_be32(from->di_uid); +- to->di_gid = cpu_to_be32(from->di_gid); ++ ++ to->di_uid = cpu_to_be32(TAGINO_UID(tagged, from->di_uid, from->di_tag)); ++ to->di_gid = cpu_to_be32(TAGINO_GID(tagged, from->di_gid, from->di_tag)); ++ to->di_tag = cpu_to_be16(TAGINO_TAG(tagged, from->di_tag)); ++ + to->di_nlink = cpu_to_be32(from->di_nlink); + to->di_projid_lo = cpu_to_be16(from->di_projid_lo); + to->di_projid_hi = cpu_to_be16(from->di_projid_hi); +@@ -707,12 +723,14 @@ xfs_dinode_to_disk( + to->di_dmevmask = cpu_to_be32(from->di_dmevmask); + to->di_dmstate = cpu_to_be16(from->di_dmstate); + to->di_flags = cpu_to_be16(from->di_flags); ++ to->di_vflags = cpu_to_be16(from->di_vflags); + to->di_gen = cpu_to_be32(from->di_gen); + } + + STATIC uint + _xfs_dic2xflags( +- __uint16_t di_flags) ++ __uint16_t di_flags, ++ __uint16_t di_vflags) + { + uint flags = 0; + +@@ -723,6 +741,8 @@ _xfs_dic2xflags( + flags |= XFS_XFLAG_PREALLOC; + if (di_flags & XFS_DIFLAG_IMMUTABLE) + flags |= XFS_XFLAG_IMMUTABLE; ++ if (di_flags & XFS_DIFLAG_IXUNLINK) ++ flags |= XFS_XFLAG_IXUNLINK; + if (di_flags & XFS_DIFLAG_APPEND) + flags |= XFS_XFLAG_APPEND; + if (di_flags & XFS_DIFLAG_SYNC) +@@ -747,6 +767,10 @@ _xfs_dic2xflags( + flags |= XFS_XFLAG_FILESTREAM; + } + ++ if (di_vflags & XFS_DIVFLAG_BARRIER) ++ flags |= FS_BARRIER_FL; ++ if (di_vflags & XFS_DIVFLAG_COW) ++ flags |= FS_COW_FL; + return flags; + } + +@@ -756,7 +780,7 @@ xfs_ip2xflags( + { + xfs_icdinode_t *dic = &ip->i_d; + +- return _xfs_dic2xflags(dic->di_flags) | ++ return _xfs_dic2xflags(dic->di_flags, dic->di_vflags) | + (XFS_IFORK_Q(ip) ? XFS_XFLAG_HASATTR : 0); + } + +@@ -764,7 +788,8 @@ uint + xfs_dic2xflags( + xfs_dinode_t *dip) + { +- return _xfs_dic2xflags(be16_to_cpu(dip->di_flags)) | ++ return _xfs_dic2xflags(be16_to_cpu(dip->di_flags), ++ be16_to_cpu(dip->di_vflags)) | + (XFS_DFORK_Q(dip) ? XFS_XFLAG_HASATTR : 0); + } + +@@ -797,7 +822,6 @@ xfs_iread( + if (error) + return error; + dip = (xfs_dinode_t *)xfs_buf_offset(bp, ip->i_imap.im_boffset); +- + /* + * If we got something that isn't an inode it means someone + * (nfs or dmi) has a stale handle. +@@ -820,7 +844,8 @@ xfs_iread( + * Otherwise, just get the truly permanent information. + */ + if (dip->di_mode) { +- xfs_dinode_from_disk(&ip->i_d, dip); ++ xfs_dinode_from_disk(&ip->i_d, dip, ++ mp->m_flags & XFS_MOUNT_TAGGED); + error = xfs_iformat(ip, dip); + if (error) { + #ifdef DEBUG +@@ -1015,6 +1040,7 @@ xfs_ialloc( + ASSERT(ip->i_d.di_nlink == nlink); + ip->i_d.di_uid = current_fsuid(); + ip->i_d.di_gid = current_fsgid(); ++ ip->i_d.di_tag = current_fstag(&ip->i_vnode); + xfs_set_projid(ip, prid); + memset(&(ip->i_d.di_pad[0]), 0, sizeof(ip->i_d.di_pad)); + +@@ -1075,6 +1101,7 @@ xfs_ialloc( + ip->i_d.di_dmevmask = 0; + ip->i_d.di_dmstate = 0; + ip->i_d.di_flags = 0; ++ ip->i_d.di_vflags = 0; + flags = XFS_ILOG_CORE; + switch (mode & S_IFMT) { + case S_IFIFO: +@@ -2108,6 +2135,7 @@ xfs_ifree( + } + ip->i_d.di_mode = 0; /* mark incore inode as free */ + ip->i_d.di_flags = 0; ++ ip->i_d.di_vflags = 0; + ip->i_d.di_dmevmask = 0; + ip->i_d.di_forkoff = 0; /* mark the attr fork not in use */ + ip->i_df.if_ext_max = +@@ -2987,7 +3015,8 @@ xfs_iflush_int( + * because if the inode is dirty at all the core must + * be. + */ +- xfs_dinode_to_disk(dip, &ip->i_d); ++ xfs_dinode_to_disk(dip, &ip->i_d, ++ mp->m_flags & XFS_MOUNT_TAGGED); + + /* Wrap, we never let the log put out DI_MAX_FLUSH */ + if (ip->i_d.di_flushiter == DI_MAX_FLUSH) +diff -NurpP --minimal linux-3.0.9/fs/xfs/xfs_inode.h linux-3.0.9-vs2.3.2.1/fs/xfs/xfs_inode.h +--- linux-3.0.9/fs/xfs/xfs_inode.h 2011-07-22 11:18:10.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/fs/xfs/xfs_inode.h 2011-07-01 11:35:35.000000000 +0200 +@@ -135,7 +135,9 @@ typedef struct xfs_icdinode { + __uint32_t di_nlink; /* number of links to file */ + __uint16_t di_projid_lo; /* lower part of owner's project id */ + __uint16_t di_projid_hi; /* higher part of owner's project id */ +- __uint8_t di_pad[6]; /* unused, zeroed space */ ++ __uint8_t di_pad[2]; /* unused, zeroed space */ ++ __uint16_t di_tag; /* context tagging */ ++ __uint16_t di_vflags; /* vserver specific flags */ + __uint16_t di_flushiter; /* incremented on flush */ + xfs_ictimestamp_t di_atime; /* time last accessed */ + xfs_ictimestamp_t di_mtime; /* time last modified */ +@@ -546,7 +548,7 @@ int xfs_itobp(struct xfs_mount *, struc + int xfs_iread(struct xfs_mount *, struct xfs_trans *, + struct xfs_inode *, uint); + void xfs_dinode_to_disk(struct xfs_dinode *, +- struct xfs_icdinode *); ++ struct xfs_icdinode *, int); + void xfs_idestroy_fork(struct xfs_inode *, int); + void xfs_idata_realloc(struct xfs_inode *, int, int); + void xfs_iroot_realloc(struct xfs_inode *, int, int); +diff -NurpP --minimal linux-3.0.9/fs/xfs/xfs_itable.c linux-3.0.9-vs2.3.2.1/fs/xfs/xfs_itable.c +--- linux-3.0.9/fs/xfs/xfs_itable.c 2011-05-22 16:17:54.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/fs/xfs/xfs_itable.c 2011-06-10 22:11:24.000000000 +0200 +@@ -98,6 +98,7 @@ xfs_bulkstat_one_int( + buf->bs_mode = dic->di_mode; + buf->bs_uid = dic->di_uid; + buf->bs_gid = dic->di_gid; ++ buf->bs_tag = dic->di_tag; + buf->bs_size = dic->di_size; + + /* +diff -NurpP --minimal linux-3.0.9/fs/xfs/xfs_log_recover.c linux-3.0.9-vs2.3.2.1/fs/xfs/xfs_log_recover.c +--- linux-3.0.9/fs/xfs/xfs_log_recover.c 2011-07-22 11:18:10.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/fs/xfs/xfs_log_recover.c 2011-06-10 22:11:24.000000000 +0200 +@@ -2343,7 +2343,8 @@ xlog_recover_inode_pass2( + } + + /* The core is in in-core format */ +- xfs_dinode_to_disk(dip, item->ri_buf[1].i_addr); ++ xfs_dinode_to_disk(dip, item->ri_buf[1].i_addr, ++ mp->m_flags & XFS_MOUNT_TAGGED); + + /* the rest is in on-disk format */ + if (item->ri_buf[1].i_len > sizeof(struct xfs_icdinode)) { +diff -NurpP --minimal linux-3.0.9/fs/xfs/xfs_mount.h linux-3.0.9-vs2.3.2.1/fs/xfs/xfs_mount.h +--- linux-3.0.9/fs/xfs/xfs_mount.h 2011-07-22 11:18:10.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/fs/xfs/xfs_mount.h 2011-06-10 22:11:24.000000000 +0200 +@@ -249,6 +249,7 @@ typedef struct xfs_mount { + allocator */ + #define XFS_MOUNT_NOATTR2 (1ULL << 25) /* disable use of attr2 format */ + ++#define XFS_MOUNT_TAGGED (1ULL << 31) /* context tagging */ + + /* + * Default minimum read and write sizes. +diff -NurpP --minimal linux-3.0.9/fs/xfs/xfs_vnodeops.c linux-3.0.9-vs2.3.2.1/fs/xfs/xfs_vnodeops.c +--- linux-3.0.9/fs/xfs/xfs_vnodeops.c 2011-07-22 11:18:10.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/fs/xfs/xfs_vnodeops.c 2011-07-01 11:35:35.000000000 +0200 +@@ -50,6 +50,78 @@ + #include "xfs_vnodeops.h" + #include "xfs_trace.h" + ++ ++STATIC void ++xfs_get_inode_flags( ++ xfs_inode_t *ip) ++{ ++ struct inode *inode = VFS_I(ip); ++ unsigned int flags = inode->i_flags; ++ unsigned int vflags = inode->i_vflags; ++ ++ if (flags & S_IMMUTABLE) ++ ip->i_d.di_flags |= XFS_DIFLAG_IMMUTABLE; ++ else ++ ip->i_d.di_flags &= ~XFS_DIFLAG_IMMUTABLE; ++ if (flags & S_IXUNLINK) ++ ip->i_d.di_flags |= XFS_DIFLAG_IXUNLINK; ++ else ++ ip->i_d.di_flags &= ~XFS_DIFLAG_IXUNLINK; ++ ++ if (vflags & V_BARRIER) ++ ip->i_d.di_vflags |= XFS_DIVFLAG_BARRIER; ++ else ++ ip->i_d.di_vflags &= ~XFS_DIVFLAG_BARRIER; ++ if (vflags & V_COW) ++ ip->i_d.di_vflags |= XFS_DIVFLAG_COW; ++ else ++ ip->i_d.di_vflags &= ~XFS_DIVFLAG_COW; ++} ++ ++int ++xfs_sync_flags( ++ struct inode *inode, ++ int flags, ++ int vflags) ++{ ++ struct xfs_inode *ip = XFS_I(inode); ++ struct xfs_mount *mp = ip->i_mount; ++ struct xfs_trans *tp; ++ unsigned int lock_flags = 0; ++ int code; ++ ++ tp = xfs_trans_alloc(mp, XFS_TRANS_SETATTR_NOT_SIZE); ++ code = xfs_trans_reserve(tp, 0, XFS_ICHANGE_LOG_RES(mp), 0, 0, 0); ++ if (code) ++ goto error_out; ++ ++ xfs_ilock(ip, XFS_ILOCK_EXCL); ++ ++ xfs_trans_ijoin(tp, ip); ++ ++ inode->i_flags = flags; ++ inode->i_vflags = vflags; ++ xfs_get_inode_flags(ip); ++ ++ xfs_trans_log_inode(tp, ip, XFS_ILOG_CORE); ++ xfs_trans_ichgtime(tp, ip, XFS_ICHGTIME_CHG); ++ ++ XFS_STATS_INC(xs_ig_attrchg); ++ ++ if (mp->m_flags & XFS_MOUNT_WSYNC) ++ xfs_trans_set_sync(tp); ++ code = xfs_trans_commit(tp, 0); ++ xfs_iunlock(ip, XFS_ILOCK_EXCL); ++ return code; ++ ++error_out: ++ xfs_trans_cancel(tp, 0); ++ if (lock_flags) ++ xfs_iunlock(ip, XFS_ILOCK_EXCL); ++ return code; ++} ++ ++ + int + xfs_setattr( + struct xfs_inode *ip, +@@ -65,6 +137,7 @@ xfs_setattr( + uint commit_flags=0; + uid_t uid=0, iuid=0; + gid_t gid=0, igid=0; ++ tag_t tag=0, itag=0; + struct xfs_dquot *udqp, *gdqp, *olddquot1, *olddquot2; + int need_iolock = 1; + +@@ -147,7 +220,7 @@ xfs_setattr( + /* + * Change file ownership. Must be the owner or privileged. + */ +- if (mask & (ATTR_UID|ATTR_GID)) { ++ if (mask & (ATTR_UID|ATTR_GID|ATTR_TAG)) { + /* + * These IDs could have changed since we last looked at them. + * But, we're assured that if the ownership did change +@@ -156,8 +229,10 @@ xfs_setattr( + */ + iuid = ip->i_d.di_uid; + igid = ip->i_d.di_gid; ++ itag = ip->i_d.di_tag; + gid = (mask & ATTR_GID) ? iattr->ia_gid : igid; + uid = (mask & ATTR_UID) ? iattr->ia_uid : iuid; ++ tag = (mask & ATTR_TAG) ? iattr->ia_tag : itag; + + /* + * Do a quota reservation only if uid/gid is actually +@@ -165,7 +240,8 @@ xfs_setattr( + */ + if (XFS_IS_QUOTA_RUNNING(mp) && + ((XFS_IS_UQUOTA_ON(mp) && iuid != uid) || +- (XFS_IS_GQUOTA_ON(mp) && igid != gid))) { ++ (XFS_IS_GQUOTA_ON(mp) && igid != gid) || ++ (XFS_IS_GQUOTA_ON(mp) && itag != tag))) { + ASSERT(tp); + code = xfs_qm_vop_chown_reserve(tp, ip, udqp, gdqp, + capable(CAP_FOWNER) ? +@@ -329,7 +405,7 @@ xfs_setattr( + /* + * Change file ownership. Must be the owner or privileged. + */ +- if (mask & (ATTR_UID|ATTR_GID)) { ++ if (mask & (ATTR_UID|ATTR_GID|ATTR_TAG)) { + /* + * CAP_FSETID overrides the following restrictions: + * +@@ -345,6 +421,10 @@ xfs_setattr( + * Change the ownerships and register quota modifications + * in the transaction. + */ ++ if (itag != tag) { ++ ip->i_d.di_tag = tag; ++ inode->i_tag = tag; ++ } + if (iuid != uid) { + if (XFS_IS_QUOTA_RUNNING(mp) && XFS_IS_UQUOTA_ON(mp)) { + ASSERT(mask & ATTR_UID); +diff -NurpP --minimal linux-3.0.9/fs/xfs/xfs_vnodeops.h linux-3.0.9-vs2.3.2.1/fs/xfs/xfs_vnodeops.h +--- linux-3.0.9/fs/xfs/xfs_vnodeops.h 2011-05-22 16:17:54.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/fs/xfs/xfs_vnodeops.h 2011-06-10 22:11:24.000000000 +0200 +@@ -13,6 +13,7 @@ struct xfs_inode; + struct xfs_iomap; + + ++int xfs_sync_xflags(struct xfs_inode *ip); + int xfs_setattr(struct xfs_inode *ip, struct iattr *vap, int flags); + #define XFS_ATTR_DMI 0x01 /* invocation from a DMI function */ + #define XFS_ATTR_NONBLOCK 0x02 /* return EAGAIN if operation would block */ +diff -NurpP --minimal linux-3.0.9/include/asm-generic/tlb.h linux-3.0.9-vs2.3.2.1/include/asm-generic/tlb.h +--- linux-3.0.9/include/asm-generic/tlb.h 2011-07-22 11:18:10.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/include/asm-generic/tlb.h 2011-06-10 22:11:24.000000000 +0200 +@@ -16,6 +16,7 @@ + #define _ASM_GENERIC__TLB_H + + #include <linux/swap.h> ++#include <linux/vs_memory.h> + #include <asm/pgalloc.h> + #include <asm/tlbflush.h> + +diff -NurpP --minimal linux-3.0.9/include/linux/Kbuild linux-3.0.9-vs2.3.2.1/include/linux/Kbuild +--- linux-3.0.9/include/linux/Kbuild 2011-07-22 11:18:10.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/include/linux/Kbuild 2011-06-10 22:11:24.000000000 +0200 +@@ -17,6 +17,7 @@ header-y += netfilter_bridge/ + header-y += netfilter_ipv4/ + header-y += netfilter_ipv6/ + header-y += usb/ ++header-y += vserver/ + header-y += wimax/ + + objhdr-y += version.h +diff -NurpP --minimal linux-3.0.9/include/linux/capability.h linux-3.0.9-vs2.3.2.1/include/linux/capability.h +--- linux-3.0.9/include/linux/capability.h 2011-07-22 11:18:10.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/include/linux/capability.h 2011-06-10 22:11:24.000000000 +0200 +@@ -279,6 +279,7 @@ struct cpu_vfs_cap_data { + arbitrary SCSI commands */ + /* Allow setting encryption key on loopback filesystem */ + /* Allow setting zone reclaim policy */ ++/* Allow the selection of a security context */ + + #define CAP_SYS_ADMIN 21 + +@@ -362,7 +363,12 @@ struct cpu_vfs_cap_data { + + #define CAP_LAST_CAP CAP_WAKE_ALARM + +-#define cap_valid(x) ((x) >= 0 && (x) <= CAP_LAST_CAP) ++/* Allow context manipulations */ ++/* Allow changing context info on files */ ++ ++#define CAP_CONTEXT 63 ++ ++#define cap_valid(x) ((x) >= 0 && ((x) <= CAP_LAST_CAP || (x) == CAP_CONTEXT)) + + /* + * Bit location of each capability (used by user-space library and kernel) +diff -NurpP --minimal linux-3.0.9/include/linux/cred.h linux-3.0.9-vs2.3.2.1/include/linux/cred.h +--- linux-3.0.9/include/linux/cred.h 2011-07-22 11:18:10.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/include/linux/cred.h 2011-06-10 22:11:24.000000000 +0200 +@@ -156,6 +156,7 @@ extern void exit_creds(struct task_struc + extern int copy_creds(struct task_struct *, unsigned long); + extern const struct cred *get_task_cred(struct task_struct *); + extern struct cred *cred_alloc_blank(void); ++extern struct cred *__prepare_creds(const struct cred *); + extern struct cred *prepare_creds(void); + extern struct cred *prepare_exec_creds(void); + extern int commit_creds(struct cred *); +@@ -209,6 +210,31 @@ static inline void validate_process_cred + } + #endif + ++static inline void set_cred_subscribers(struct cred *cred, int n) ++{ ++#ifdef CONFIG_DEBUG_CREDENTIALS ++ atomic_set(&cred->subscribers, n); ++#endif ++} ++ ++static inline int read_cred_subscribers(const struct cred *cred) ++{ ++#ifdef CONFIG_DEBUG_CREDENTIALS ++ return atomic_read(&cred->subscribers); ++#else ++ return 0; ++#endif ++} ++ ++static inline void alter_cred_subscribers(const struct cred *_cred, int n) ++{ ++#ifdef CONFIG_DEBUG_CREDENTIALS ++ struct cred *cred = (struct cred *) _cred; ++ ++ atomic_add(n, &cred->subscribers); ++#endif ++} ++ + /** + * get_new_cred - Get a reference on a new set of credentials + * @cred: The new credentials to reference +diff -NurpP --minimal linux-3.0.9/include/linux/devpts_fs.h linux-3.0.9-vs2.3.2.1/include/linux/devpts_fs.h +--- linux-3.0.9/include/linux/devpts_fs.h 2008-12-25 00:26:37.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/include/linux/devpts_fs.h 2011-06-10 22:11:24.000000000 +0200 +@@ -45,5 +45,4 @@ static inline void devpts_pty_kill(struc + + #endif + +- + #endif /* _LINUX_DEVPTS_FS_H */ +diff -NurpP --minimal linux-3.0.9/include/linux/ext2_fs.h linux-3.0.9-vs2.3.2.1/include/linux/ext2_fs.h +--- linux-3.0.9/include/linux/ext2_fs.h 2011-11-15 16:40:47.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/include/linux/ext2_fs.h 2011-11-15 17:37:07.000000000 +0100 +@@ -189,8 +189,12 @@ struct ext2_group_desc + #define EXT2_NOTAIL_FL FS_NOTAIL_FL /* file tail should not be merged */ + #define EXT2_DIRSYNC_FL FS_DIRSYNC_FL /* dirsync behaviour (directories only) */ + #define EXT2_TOPDIR_FL FS_TOPDIR_FL /* Top of directory hierarchies*/ ++#define EXT2_IXUNLINK_FL FS_IXUNLINK_FL /* Immutable invert on unlink */ + #define EXT2_RESERVED_FL FS_RESERVED_FL /* reserved for ext2 lib */ + ++#define EXT2_BARRIER_FL FS_BARRIER_FL /* Barrier for chroot() */ ++#define EXT2_COW_FL FS_COW_FL /* Copy on Write marker */ ++ + #define EXT2_FL_USER_VISIBLE FS_FL_USER_VISIBLE /* User visible flags */ + #define EXT2_FL_USER_MODIFIABLE FS_FL_USER_MODIFIABLE /* User modifiable flags */ + +@@ -274,7 +278,8 @@ struct ext2_inode { + __u16 i_pad1; + __le16 l_i_uid_high; /* these 2 fields */ + __le16 l_i_gid_high; /* were reserved2[0] */ +- __u32 l_i_reserved2; ++ __le16 l_i_tag; /* Context Tag */ ++ __u16 l_i_reserved2; + } linux2; + struct { + __u8 h_i_frag; /* Fragment number */ +@@ -303,6 +308,7 @@ struct ext2_inode { + #define i_gid_low i_gid + #define i_uid_high osd2.linux2.l_i_uid_high + #define i_gid_high osd2.linux2.l_i_gid_high ++#define i_raw_tag osd2.linux2.l_i_tag + #define i_reserved2 osd2.linux2.l_i_reserved2 + #endif + +@@ -347,6 +353,7 @@ struct ext2_inode { + #define EXT2_MOUNT_USRQUOTA 0x020000 /* user quota */ + #define EXT2_MOUNT_GRPQUOTA 0x040000 /* group quota */ + #define EXT2_MOUNT_RESERVATION 0x080000 /* Preallocation */ ++#define EXT2_MOUNT_TAGGED (1<<24) /* Enable Context Tags */ + + + #define clear_opt(o, opt) o &= ~EXT2_MOUNT_##opt +diff -NurpP --minimal linux-3.0.9/include/linux/ext3_fs.h linux-3.0.9-vs2.3.2.1/include/linux/ext3_fs.h +--- linux-3.0.9/include/linux/ext3_fs.h 2011-11-15 16:40:47.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/include/linux/ext3_fs.h 2011-11-15 17:37:07.000000000 +0100 +@@ -173,10 +173,14 @@ struct ext3_group_desc + #define EXT3_NOTAIL_FL 0x00008000 /* file tail should not be merged */ + #define EXT3_DIRSYNC_FL 0x00010000 /* dirsync behaviour (directories only) */ + #define EXT3_TOPDIR_FL 0x00020000 /* Top of directory hierarchies*/ ++#define EXT3_IXUNLINK_FL 0x08000000 /* Immutable invert on unlink */ + #define EXT3_RESERVED_FL 0x80000000 /* reserved for ext3 lib */ + +-#define EXT3_FL_USER_VISIBLE 0x0003DFFF /* User visible flags */ +-#define EXT3_FL_USER_MODIFIABLE 0x000380FF /* User modifiable flags */ ++#define EXT3_BARRIER_FL 0x04000000 /* Barrier for chroot() */ ++#define EXT3_COW_FL 0x20000000 /* Copy on Write marker */ ++ ++#define EXT3_FL_USER_VISIBLE 0x0103DFFF /* User visible flags */ ++#define EXT3_FL_USER_MODIFIABLE 0x010380FF /* User modifiable flags */ + + /* Flags that should be inherited by new inodes from their parent. */ + #define EXT3_FL_INHERITED (EXT3_SECRM_FL | EXT3_UNRM_FL | EXT3_COMPR_FL |\ +@@ -312,7 +316,8 @@ struct ext3_inode { + __u16 i_pad1; + __le16 l_i_uid_high; /* these 2 fields */ + __le16 l_i_gid_high; /* were reserved2[0] */ +- __u32 l_i_reserved2; ++ __le16 l_i_tag; /* Context Tag */ ++ __u16 l_i_reserved2; + } linux2; + struct { + __u8 h_i_frag; /* Fragment number */ +@@ -343,6 +348,7 @@ struct ext3_inode { + #define i_gid_low i_gid + #define i_uid_high osd2.linux2.l_i_uid_high + #define i_gid_high osd2.linux2.l_i_gid_high ++#define i_raw_tag osd2.linux2.l_i_tag + #define i_reserved2 osd2.linux2.l_i_reserved2 + + #elif defined(__GNU__) +@@ -405,6 +411,7 @@ struct ext3_inode { + #define EXT3_MOUNT_GRPQUOTA 0x200000 /* "old" group quota */ + #define EXT3_MOUNT_DATA_ERR_ABORT 0x400000 /* Abort on file data write + * error in ordered mode */ ++#define EXT3_MOUNT_TAGGED (1<<24) /* Enable Context Tags */ + + /* Compatibility, for having both ext2_fs.h and ext3_fs.h included at once */ + #ifndef _LINUX_EXT2_FS_H +@@ -919,6 +926,7 @@ extern void ext3_get_inode_flags(struct + extern void ext3_set_aops(struct inode *inode); + extern int ext3_fiemap(struct inode *inode, struct fiemap_extent_info *fieinfo, + u64 start, u64 len); ++extern int ext3_sync_flags(struct inode *, int, int); + + /* ioctl.c */ + extern long ext3_ioctl(struct file *, unsigned int, unsigned long); +diff -NurpP --minimal linux-3.0.9/include/linux/fs.h linux-3.0.9-vs2.3.2.1/include/linux/fs.h +--- linux-3.0.9/include/linux/fs.h 2011-07-22 11:18:10.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/include/linux/fs.h 2011-07-01 11:35:35.000000000 +0200 +@@ -208,6 +208,9 @@ struct inodes_stat_t { + #define MS_KERNMOUNT (1<<22) /* this is a kern_mount call */ + #define MS_I_VERSION (1<<23) /* Update inode I_version field */ + #define MS_STRICTATIME (1<<24) /* Always perform atime updates */ ++#define MS_TAGGED (1<<25) /* use generic inode tagging */ ++#define MS_TAGID (1<<26) /* use specific tag for this mount */ ++#define MS_NOTAGCHECK (1<<27) /* don't check tags */ + #define MS_NOSEC (1<<28) + #define MS_BORN (1<<29) + #define MS_ACTIVE (1<<30) +@@ -239,6 +242,14 @@ struct inodes_stat_t { + #define S_IMA 1024 /* Inode has an associated IMA struct */ + #define S_AUTOMOUNT 2048 /* Automount/referral quasi-directory */ + #define S_NOSEC 4096 /* no suid or xattr security attributes */ ++#define S_IXUNLINK 8192 /* Immutable Invert on unlink */ ++ ++/* Linux-VServer related Inode flags */ ++ ++#define V_VALID 1 ++#define V_XATTR 2 ++#define V_BARRIER 4 /* Barrier for chroot() */ ++#define V_COW 8 /* Copy on Write */ + + /* + * Note that nosuid etc flags are inode-specific: setting some file-system +@@ -261,12 +272,15 @@ struct inodes_stat_t { + #define IS_DIRSYNC(inode) (__IS_FLG(inode, MS_SYNCHRONOUS|MS_DIRSYNC) || \ + ((inode)->i_flags & (S_SYNC|S_DIRSYNC))) + #define IS_MANDLOCK(inode) __IS_FLG(inode, MS_MANDLOCK) +-#define IS_NOATIME(inode) __IS_FLG(inode, MS_RDONLY|MS_NOATIME) +-#define IS_I_VERSION(inode) __IS_FLG(inode, MS_I_VERSION) ++#define IS_NOATIME(inode) __IS_FLG(inode, MS_RDONLY|MS_NOATIME) ++#define IS_I_VERSION(inode) __IS_FLG(inode, MS_I_VERSION) ++#define IS_TAGGED(inode) __IS_FLG(inode, MS_TAGGED) + + #define IS_NOQUOTA(inode) ((inode)->i_flags & S_NOQUOTA) + #define IS_APPEND(inode) ((inode)->i_flags & S_APPEND) + #define IS_IMMUTABLE(inode) ((inode)->i_flags & S_IMMUTABLE) ++#define IS_IXUNLINK(inode) ((inode)->i_flags & S_IXUNLINK) ++#define IS_IXORUNLINK(inode) ((IS_IXUNLINK(inode) ? S_IMMUTABLE : 0) ^ IS_IMMUTABLE(inode)) + #define IS_POSIXACL(inode) __IS_FLG(inode, MS_POSIXACL) + + #define IS_DEADDIR(inode) ((inode)->i_flags & S_DEAD) +@@ -277,6 +291,16 @@ struct inodes_stat_t { + #define IS_AUTOMOUNT(inode) ((inode)->i_flags & S_AUTOMOUNT) + #define IS_NOSEC(inode) ((inode)->i_flags & S_NOSEC) + ++#define IS_BARRIER(inode) (S_ISDIR((inode)->i_mode) && ((inode)->i_vflags & V_BARRIER)) ++ ++#ifdef CONFIG_VSERVER_COWBL ++# define IS_COW(inode) (IS_IXUNLINK(inode) && IS_IMMUTABLE(inode)) ++# define IS_COW_LINK(inode) (S_ISREG((inode)->i_mode) && ((inode)->i_nlink > 1)) ++#else ++# define IS_COW(inode) (0) ++# define IS_COW_LINK(inode) (0) ++#endif ++ + /* the read-only stuff doesn't really belong here, but any other place is + probably as bad and I don't want to create yet another include file. */ + +@@ -362,11 +386,14 @@ struct inodes_stat_t { + #define FS_EXTENT_FL 0x00080000 /* Extents */ + #define FS_DIRECTIO_FL 0x00100000 /* Use direct i/o */ + #define FS_NOCOW_FL 0x00800000 /* Do not cow file */ ++#define FS_IXUNLINK_FL 0x08000000 /* Immutable invert on unlink */ + #define FS_RESERVED_FL 0x80000000 /* reserved for ext2 lib */ + +-#define FS_FL_USER_VISIBLE 0x0003DFFF /* User visible flags */ +-#define FS_FL_USER_MODIFIABLE 0x000380FF /* User modifiable flags */ +- ++#define FS_BARRIER_FL 0x04000000 /* Barrier for chroot() */ ++#define FS_COW_FL 0x20000000 /* Copy on Write marker */ ++ ++#define FS_FL_USER_VISIBLE 0x0103DFFF /* User visible flags */ ++#define FS_FL_USER_MODIFIABLE 0x010380FF /* User modifiable flags */ + + #define SYNC_FILE_RANGE_WAIT_BEFORE 1 + #define SYNC_FILE_RANGE_WRITE 2 +@@ -447,6 +474,7 @@ typedef void (dio_iodone_t)(struct kiocb + #define ATTR_KILL_PRIV (1 << 14) + #define ATTR_OPEN (1 << 15) /* Truncating from open(O_TRUNC) */ + #define ATTR_TIMES_SET (1 << 16) ++#define ATTR_TAG (1 << 17) + + /* + * This is the Inode Attributes structure, used for notify_change(). It +@@ -462,6 +490,7 @@ struct iattr { + umode_t ia_mode; + uid_t ia_uid; + gid_t ia_gid; ++ tag_t ia_tag; + loff_t ia_size; + struct timespec ia_atime; + struct timespec ia_mtime; +@@ -475,6 +504,9 @@ struct iattr { + struct file *ia_file; + }; + ++#define ATTR_FLAG_BARRIER 512 /* Barrier for chroot() */ ++#define ATTR_FLAG_IXUNLINK 1024 /* Immutable invert on unlink */ ++ + /* + * Includes for diskquotas. + */ +@@ -740,11 +772,13 @@ struct inode { + umode_t i_mode; + uid_t i_uid; + gid_t i_gid; ++ tag_t i_tag; + const struct inode_operations *i_op; + struct super_block *i_sb; + + spinlock_t i_lock; /* i_blocks, i_bytes, maybe i_size */ +- unsigned int i_flags; ++ unsigned short i_flags; ++ unsigned short i_vflags; + unsigned long i_state; + #ifdef CONFIG_SECURITY + void *i_security; +@@ -766,6 +800,7 @@ struct inode { + atomic_t i_count; + unsigned int i_nlink; + dev_t i_rdev; ++ dev_t i_mdev; + unsigned int i_blkbits; + u64 i_version; + loff_t i_size; +@@ -890,12 +925,12 @@ static inline void i_size_write(struct i + + static inline unsigned iminor(const struct inode *inode) + { +- return MINOR(inode->i_rdev); ++ return MINOR(inode->i_mdev); + } + + static inline unsigned imajor(const struct inode *inode) + { +- return MAJOR(inode->i_rdev); ++ return MAJOR(inode->i_mdev); + } + + extern struct block_device *I_BDEV(struct inode *inode); +@@ -957,6 +992,7 @@ struct file { + loff_t f_pos; + struct fown_struct f_owner; + const struct cred *f_cred; ++ xid_t f_xid; + struct file_ra_state f_ra; + + u64 f_version; +@@ -1101,6 +1137,7 @@ struct file_lock { + struct file *fl_file; + loff_t fl_start; + loff_t fl_end; ++ xid_t fl_xid; + + struct fasync_struct * fl_fasync; /* for lease break notifications */ + unsigned long fl_break_time; /* for nonblocking lease breaks */ +@@ -1600,6 +1637,7 @@ struct inode_operations { + ssize_t (*getxattr) (struct dentry *, const char *, void *, size_t); + ssize_t (*listxattr) (struct dentry *, char *, size_t); + int (*removexattr) (struct dentry *, const char *); ++ int (*sync_flags) (struct inode *, int, int); + void (*truncate_range)(struct inode *, loff_t, loff_t); + int (*fiemap)(struct inode *, struct fiemap_extent_info *, u64 start, + u64 len); +@@ -1618,6 +1656,7 @@ extern ssize_t vfs_readv(struct file *, + unsigned long, loff_t *); + extern ssize_t vfs_writev(struct file *, const struct iovec __user *, + unsigned long, loff_t *); ++ssize_t vfs_sendfile(struct file *, struct file *, loff_t *, size_t, loff_t); + + struct super_operations { + struct inode *(*alloc_inode)(struct super_block *sb); +@@ -2437,6 +2476,7 @@ extern int dcache_dir_open(struct inode + extern int dcache_dir_close(struct inode *, struct file *); + extern loff_t dcache_dir_lseek(struct file *, loff_t, int); + extern int dcache_readdir(struct file *, void *, filldir_t); ++extern int dcache_readdir_filter(struct file *, void *, filldir_t, int (*)(struct dentry *)); + extern int simple_setattr(struct dentry *, struct iattr *); + extern int simple_getattr(struct vfsmount *, struct dentry *, struct kstat *); + extern int simple_statfs(struct dentry *, struct kstatfs *); +diff -NurpP --minimal linux-3.0.9/include/linux/gfs2_ondisk.h linux-3.0.9-vs2.3.2.1/include/linux/gfs2_ondisk.h +--- linux-3.0.9/include/linux/gfs2_ondisk.h 2010-07-07 18:31:55.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/include/linux/gfs2_ondisk.h 2011-06-10 22:11:24.000000000 +0200 +@@ -211,6 +211,9 @@ enum { + gfs2fl_NoAtime = 7, + gfs2fl_Sync = 8, + gfs2fl_System = 9, ++ gfs2fl_IXUnlink = 16, ++ gfs2fl_Barrier = 17, ++ gfs2fl_Cow = 18, + gfs2fl_TruncInProg = 29, + gfs2fl_InheritDirectio = 30, + gfs2fl_InheritJdata = 31, +@@ -227,6 +230,9 @@ enum { + #define GFS2_DIF_NOATIME 0x00000080 + #define GFS2_DIF_SYNC 0x00000100 + #define GFS2_DIF_SYSTEM 0x00000200 /* New in gfs2 */ ++#define GFS2_DIF_IXUNLINK 0x00010000 ++#define GFS2_DIF_BARRIER 0x00020000 ++#define GFS2_DIF_COW 0x00040000 + #define GFS2_DIF_TRUNC_IN_PROG 0x20000000 /* New in gfs2 */ + #define GFS2_DIF_INHERIT_DIRECTIO 0x40000000 + #define GFS2_DIF_INHERIT_JDATA 0x80000000 +diff -NurpP --minimal linux-3.0.9/include/linux/if_tun.h linux-3.0.9-vs2.3.2.1/include/linux/if_tun.h +--- linux-3.0.9/include/linux/if_tun.h 2010-08-02 16:52:54.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/include/linux/if_tun.h 2011-06-10 22:11:24.000000000 +0200 +@@ -53,6 +53,7 @@ + #define TUNDETACHFILTER _IOW('T', 214, struct sock_fprog) + #define TUNGETVNETHDRSZ _IOR('T', 215, int) + #define TUNSETVNETHDRSZ _IOW('T', 216, int) ++#define TUNSETNID _IOW('T', 217, int) + + /* TUNSETIFF ifr flags */ + #define IFF_TUN 0x0001 +diff -NurpP --minimal linux-3.0.9/include/linux/init_task.h linux-3.0.9-vs2.3.2.1/include/linux/init_task.h +--- linux-3.0.9/include/linux/init_task.h 2011-07-22 11:18:10.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/include/linux/init_task.h 2011-06-10 22:11:24.000000000 +0200 +@@ -193,6 +193,10 @@ extern struct cred init_cred; + INIT_FTRACE_GRAPH \ + INIT_TRACE_RECURSION \ + INIT_TASK_RCU_PREEMPT(tsk) \ ++ .xid = 0, \ ++ .vx_info = NULL, \ ++ .nid = 0, \ ++ .nx_info = NULL, \ + } + + +diff -NurpP --minimal linux-3.0.9/include/linux/ipc.h linux-3.0.9-vs2.3.2.1/include/linux/ipc.h +--- linux-3.0.9/include/linux/ipc.h 2009-12-03 20:02:55.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/include/linux/ipc.h 2011-06-10 22:11:24.000000000 +0200 +@@ -91,6 +91,7 @@ struct kern_ipc_perm + key_t key; + uid_t uid; + gid_t gid; ++ xid_t xid; + uid_t cuid; + gid_t cgid; + mode_t mode; +diff -NurpP --minimal linux-3.0.9/include/linux/ipc_namespace.h linux-3.0.9-vs2.3.2.1/include/linux/ipc_namespace.h +--- linux-3.0.9/include/linux/ipc_namespace.h 2011-05-22 16:17:55.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/include/linux/ipc_namespace.h 2011-06-13 14:09:44.000000000 +0200 +@@ -94,7 +94,8 @@ static inline int mq_init_ns(struct ipc_ + + #if defined(CONFIG_IPC_NS) + extern struct ipc_namespace *copy_ipcs(unsigned long flags, +- struct task_struct *tsk); ++ struct ipc_namespace *old_ns, ++ struct user_namespace *user_ns); + static inline struct ipc_namespace *get_ipc_ns(struct ipc_namespace *ns) + { + if (ns) +@@ -105,12 +106,13 @@ static inline struct ipc_namespace *get_ + extern void put_ipc_ns(struct ipc_namespace *ns); + #else + static inline struct ipc_namespace *copy_ipcs(unsigned long flags, +- struct task_struct *tsk) ++ struct ipc_namespace *old_ns, ++ struct user_namespace *user_ns) + { + if (flags & CLONE_NEWIPC) + return ERR_PTR(-EINVAL); + +- return tsk->nsproxy->ipc_ns; ++ return old_ns; + } + + static inline struct ipc_namespace *get_ipc_ns(struct ipc_namespace *ns) +diff -NurpP --minimal linux-3.0.9/include/linux/loop.h linux-3.0.9-vs2.3.2.1/include/linux/loop.h +--- linux-3.0.9/include/linux/loop.h 2009-09-10 15:26:25.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/include/linux/loop.h 2011-06-10 22:11:24.000000000 +0200 +@@ -45,6 +45,7 @@ struct loop_device { + struct loop_func_table *lo_encryption; + __u32 lo_init[2]; + uid_t lo_key_owner; /* Who set the key */ ++ xid_t lo_xid; + int (*ioctl)(struct loop_device *, int cmd, + unsigned long arg); + +diff -NurpP --minimal linux-3.0.9/include/linux/magic.h linux-3.0.9-vs2.3.2.1/include/linux/magic.h +--- linux-3.0.9/include/linux/magic.h 2011-05-22 16:17:55.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/include/linux/magic.h 2011-06-10 22:11:24.000000000 +0200 +@@ -3,7 +3,7 @@ + + #define ADFS_SUPER_MAGIC 0xadf5 + #define AFFS_SUPER_MAGIC 0xadff +-#define AFS_SUPER_MAGIC 0x5346414F ++#define AFS_SUPER_MAGIC 0x5346414F + #define AUTOFS_SUPER_MAGIC 0x0187 + #define CODA_SUPER_MAGIC 0x73757245 + #define CRAMFS_MAGIC 0x28cd3d45 /* some random number */ +@@ -41,6 +41,7 @@ + #define NFS_SUPER_MAGIC 0x6969 + #define OPENPROM_SUPER_MAGIC 0x9fa1 + #define PROC_SUPER_MAGIC 0x9fa0 ++#define DEVPTS_SUPER_MAGIC 0x1cd1 + #define QNX4_SUPER_MAGIC 0x002f /* qnx4 fs detection */ + + #define REISERFS_SUPER_MAGIC 0x52654973 /* used by gcc */ +diff -NurpP --minimal linux-3.0.9/include/linux/major.h linux-3.0.9-vs2.3.2.1/include/linux/major.h +--- linux-3.0.9/include/linux/major.h 2009-09-10 15:26:25.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/include/linux/major.h 2011-06-10 22:11:24.000000000 +0200 +@@ -15,6 +15,7 @@ + #define HD_MAJOR IDE0_MAJOR + #define PTY_SLAVE_MAJOR 3 + #define TTY_MAJOR 4 ++#define VROOT_MAJOR 4 + #define TTYAUX_MAJOR 5 + #define LP_MAJOR 6 + #define VCS_MAJOR 7 +diff -NurpP --minimal linux-3.0.9/include/linux/memcontrol.h linux-3.0.9-vs2.3.2.1/include/linux/memcontrol.h +--- linux-3.0.9/include/linux/memcontrol.h 2011-07-22 11:18:10.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/include/linux/memcontrol.h 2011-06-22 12:39:15.000000000 +0200 +@@ -86,6 +86,13 @@ extern struct mem_cgroup *try_get_mem_cg + extern struct mem_cgroup *mem_cgroup_from_task(struct task_struct *p); + extern struct mem_cgroup *try_get_mem_cgroup_from_mm(struct mm_struct *mm); + ++extern u64 mem_cgroup_res_read_u64(struct mem_cgroup *mem, int member); ++extern u64 mem_cgroup_memsw_read_u64(struct mem_cgroup *mem, int member); ++ ++extern s64 mem_cgroup_stat_read_cache(struct mem_cgroup *mem); ++extern s64 mem_cgroup_stat_read_anon(struct mem_cgroup *mem); ++extern s64 mem_cgroup_stat_read_mapped(struct mem_cgroup *mem); ++ + static inline + int mm_match_cgroup(const struct mm_struct *mm, const struct mem_cgroup *cgroup) + { +diff -NurpP --minimal linux-3.0.9/include/linux/mm_types.h linux-3.0.9-vs2.3.2.1/include/linux/mm_types.h +--- linux-3.0.9/include/linux/mm_types.h 2011-11-15 16:40:47.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/include/linux/mm_types.h 2011-11-15 17:37:07.000000000 +0100 +@@ -282,6 +282,7 @@ struct mm_struct { + + /* Architecture-specific MM context */ + mm_context_t context; ++ struct vx_info *mm_vx_info; + + /* Swap token stuff */ + /* +diff -NurpP --minimal linux-3.0.9/include/linux/mmzone.h linux-3.0.9-vs2.3.2.1/include/linux/mmzone.h +--- linux-3.0.9/include/linux/mmzone.h 2011-07-22 11:18:11.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/include/linux/mmzone.h 2011-07-01 11:35:35.000000000 +0200 +@@ -654,6 +654,13 @@ typedef struct pglist_data { + __pgdat->node_start_pfn + __pgdat->node_spanned_pages;\ + }) + ++#define node_start_pfn(nid) (NODE_DATA(nid)->node_start_pfn) ++ ++#define node_end_pfn(nid) ({\ ++ pg_data_t *__pgdat = NODE_DATA(nid);\ ++ __pgdat->node_start_pfn + __pgdat->node_spanned_pages;\ ++}) ++ + #include <linux/memory_hotplug.h> + + extern struct mutex zonelists_mutex; +diff -NurpP --minimal linux-3.0.9/include/linux/mount.h linux-3.0.9-vs2.3.2.1/include/linux/mount.h +--- linux-3.0.9/include/linux/mount.h 2011-03-15 18:07:39.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/include/linux/mount.h 2011-06-10 22:11:24.000000000 +0200 +@@ -52,6 +52,9 @@ struct mnt_pcp { + int mnt_writers; + }; + ++#define MNT_TAGID 0x10000 ++#define MNT_NOTAG 0x20000 ++ + struct vfsmount { + struct list_head mnt_hash; + struct vfsmount *mnt_parent; /* fs we are mounted on */ +@@ -86,6 +89,7 @@ struct vfsmount { + int mnt_expiry_mark; /* true if marked for expiry */ + int mnt_pinned; + int mnt_ghosts; ++ tag_t mnt_tag; /* tagging used for vfsmount */ + }; + + struct file; /* forward dec */ +diff -NurpP --minimal linux-3.0.9/include/linux/net.h linux-3.0.9-vs2.3.2.1/include/linux/net.h +--- linux-3.0.9/include/linux/net.h 2011-07-22 11:18:11.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/include/linux/net.h 2011-06-10 22:11:24.000000000 +0200 +@@ -72,6 +72,7 @@ struct net; + #define SOCK_NOSPACE 2 + #define SOCK_PASSCRED 3 + #define SOCK_PASSSEC 4 ++#define SOCK_USER_SOCKET 5 + + #ifndef ARCH_HAS_SOCKET_TYPES + /** +diff -NurpP --minimal linux-3.0.9/include/linux/netdevice.h linux-3.0.9-vs2.3.2.1/include/linux/netdevice.h +--- linux-3.0.9/include/linux/netdevice.h 2011-11-15 16:40:47.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/include/linux/netdevice.h 2011-09-17 19:22:49.000000000 +0200 +@@ -1650,6 +1650,7 @@ extern void netdev_resync_ops(struct ne + extern int call_netdevice_notifiers(unsigned long val, struct net_device *dev); + extern struct net_device *dev_get_by_index(struct net *net, int ifindex); + extern struct net_device *__dev_get_by_index(struct net *net, int ifindex); ++extern struct net_device *dev_get_by_index_real_rcu(struct net *net, int ifindex); + extern struct net_device *dev_get_by_index_rcu(struct net *net, int ifindex); + extern int dev_restart(struct net_device *dev); + #ifdef CONFIG_NETPOLL_TRAP +diff -NurpP --minimal linux-3.0.9/include/linux/nfs_mount.h linux-3.0.9-vs2.3.2.1/include/linux/nfs_mount.h +--- linux-3.0.9/include/linux/nfs_mount.h 2011-01-05 21:50:31.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/include/linux/nfs_mount.h 2011-06-10 22:11:24.000000000 +0200 +@@ -63,7 +63,8 @@ struct nfs_mount_data { + #define NFS_MOUNT_SECFLAVOUR 0x2000 /* 5 */ + #define NFS_MOUNT_NORDIRPLUS 0x4000 /* 5 */ + #define NFS_MOUNT_UNSHARED 0x8000 /* 5 */ +-#define NFS_MOUNT_FLAGMASK 0xFFFF ++#define NFS_MOUNT_TAGGED 0x10000 /* context tagging */ ++#define NFS_MOUNT_FLAGMASK 0x1FFFF + + /* The following are for internal use only */ + #define NFS_MOUNT_LOOKUP_CACHE_NONEG 0x10000 +diff -NurpP --minimal linux-3.0.9/include/linux/nsproxy.h linux-3.0.9-vs2.3.2.1/include/linux/nsproxy.h +--- linux-3.0.9/include/linux/nsproxy.h 2011-07-22 11:18:11.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/include/linux/nsproxy.h 2011-06-10 22:11:24.000000000 +0200 +@@ -3,6 +3,7 @@ + + #include <linux/spinlock.h> + #include <linux/sched.h> ++#include <linux/vserver/debug.h> + + struct mnt_namespace; + struct uts_namespace; +@@ -63,22 +64,33 @@ static inline struct nsproxy *task_nspro + } + + int copy_namespaces(unsigned long flags, struct task_struct *tsk); ++struct nsproxy *copy_nsproxy(struct nsproxy *orig); + void exit_task_namespaces(struct task_struct *tsk); + void switch_task_namespaces(struct task_struct *tsk, struct nsproxy *new); + void free_nsproxy(struct nsproxy *ns); + int unshare_nsproxy_namespaces(unsigned long, struct nsproxy **, + struct fs_struct *); + +-static inline void put_nsproxy(struct nsproxy *ns) ++#define get_nsproxy(n) __get_nsproxy(n, __FILE__, __LINE__) ++ ++static inline void __get_nsproxy(struct nsproxy *ns, ++ const char *_file, int _line) + { +- if (atomic_dec_and_test(&ns->count)) { +- free_nsproxy(ns); +- } ++ vxlprintk(VXD_CBIT(space, 0), "get_nsproxy(%p[%u])", ++ ns, atomic_read(&ns->count), _file, _line); ++ atomic_inc(&ns->count); + } + +-static inline void get_nsproxy(struct nsproxy *ns) ++#define put_nsproxy(n) __put_nsproxy(n, __FILE__, __LINE__) ++ ++static inline void __put_nsproxy(struct nsproxy *ns, ++ const char *_file, int _line) + { +- atomic_inc(&ns->count); ++ vxlprintk(VXD_CBIT(space, 0), "put_nsproxy(%p[%u])", ++ ns, atomic_read(&ns->count), _file, _line); ++ if (atomic_dec_and_test(&ns->count)) { ++ free_nsproxy(ns); ++ } + } + + #endif +diff -NurpP --minimal linux-3.0.9/include/linux/pid.h linux-3.0.9-vs2.3.2.1/include/linux/pid.h +--- linux-3.0.9/include/linux/pid.h 2011-07-22 11:18:11.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/include/linux/pid.h 2011-06-10 22:11:24.000000000 +0200 +@@ -8,7 +8,8 @@ enum pid_type + PIDTYPE_PID, + PIDTYPE_PGID, + PIDTYPE_SID, +- PIDTYPE_MAX ++ PIDTYPE_MAX, ++ PIDTYPE_REALPID + }; + + /* +@@ -171,6 +172,7 @@ static inline pid_t pid_nr(struct pid *p + } + + pid_t pid_nr_ns(struct pid *pid, struct pid_namespace *ns); ++pid_t pid_unmapped_nr_ns(struct pid *pid, struct pid_namespace *ns); + pid_t pid_vnr(struct pid *pid); + + #define do_each_pid_task(pid, type, task) \ +diff -NurpP --minimal linux-3.0.9/include/linux/proc_fs.h linux-3.0.9-vs2.3.2.1/include/linux/proc_fs.h +--- linux-3.0.9/include/linux/proc_fs.h 2011-07-22 11:18:11.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/include/linux/proc_fs.h 2011-06-10 22:11:24.000000000 +0200 +@@ -56,6 +56,7 @@ struct proc_dir_entry { + nlink_t nlink; + uid_t uid; + gid_t gid; ++ int vx_flags; + loff_t size; + const struct inode_operations *proc_iops; + /* +@@ -252,12 +253,18 @@ extern const struct proc_ns_operations n + extern const struct proc_ns_operations utsns_operations; + extern const struct proc_ns_operations ipcns_operations; + ++struct vx_info; ++struct nx_info; ++ + union proc_op { + int (*proc_get_link)(struct inode *, struct path *); + int (*proc_read)(struct task_struct *task, char *page); + int (*proc_show)(struct seq_file *m, + struct pid_namespace *ns, struct pid *pid, + struct task_struct *task); ++ int (*proc_vs_read)(char *page); ++ int (*proc_vxi_read)(struct vx_info *vxi, char *page); ++ int (*proc_nxi_read)(struct nx_info *nxi, char *page); + }; + + struct ctl_table_header; +@@ -265,6 +272,7 @@ struct ctl_table; + + struct proc_inode { + struct pid *pid; ++ int vx_flags; + int fd; + union proc_op op; + struct proc_dir_entry *pde; +diff -NurpP --minimal linux-3.0.9/include/linux/quotaops.h linux-3.0.9-vs2.3.2.1/include/linux/quotaops.h +--- linux-3.0.9/include/linux/quotaops.h 2011-05-22 16:17:57.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/include/linux/quotaops.h 2011-06-10 22:11:24.000000000 +0200 +@@ -8,6 +8,7 @@ + #define _LINUX_QUOTAOPS_ + + #include <linux/fs.h> ++#include <linux/vs_dlimit.h> + + #define DQUOT_SPACE_WARN 0x1 + #define DQUOT_SPACE_RESERVE 0x2 +@@ -204,11 +205,12 @@ static inline void dquot_drop(struct ino + + static inline int dquot_alloc_inode(const struct inode *inode) + { +- return 0; ++ return dl_alloc_inode(inode); + } + + static inline void dquot_free_inode(const struct inode *inode) + { ++ dl_free_inode(inode); + } + + static inline int dquot_transfer(struct inode *inode, struct iattr *iattr) +@@ -219,6 +221,10 @@ static inline int dquot_transfer(struct + static inline int __dquot_alloc_space(struct inode *inode, qsize_t number, + int flags) + { ++ int ret = 0; ++ ++ if ((ret = dl_alloc_space(inode, number))) ++ return ret; + if (!(flags & DQUOT_SPACE_RESERVE)) + inode_add_bytes(inode, number); + return 0; +@@ -229,6 +235,7 @@ static inline void __dquot_free_space(st + { + if (!(flags & DQUOT_SPACE_RESERVE)) + inode_sub_bytes(inode, number); ++ dl_free_space(inode, number); + } + + static inline int dquot_claim_space_nodirty(struct inode *inode, qsize_t number) +diff -NurpP --minimal linux-3.0.9/include/linux/reboot.h linux-3.0.9-vs2.3.2.1/include/linux/reboot.h +--- linux-3.0.9/include/linux/reboot.h 2010-07-07 18:31:56.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/include/linux/reboot.h 2011-06-10 22:11:24.000000000 +0200 +@@ -33,6 +33,7 @@ + #define LINUX_REBOOT_CMD_RESTART2 0xA1B2C3D4 + #define LINUX_REBOOT_CMD_SW_SUSPEND 0xD000FCE2 + #define LINUX_REBOOT_CMD_KEXEC 0x45584543 ++#define LINUX_REBOOT_CMD_OOM 0xDEADBEEF + + + #ifdef __KERNEL__ +diff -NurpP --minimal linux-3.0.9/include/linux/reiserfs_fs.h linux-3.0.9-vs2.3.2.1/include/linux/reiserfs_fs.h +--- linux-3.0.9/include/linux/reiserfs_fs.h 2011-05-22 16:17:58.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/include/linux/reiserfs_fs.h 2011-06-10 22:11:24.000000000 +0200 +@@ -976,6 +976,11 @@ struct stat_data_v1 { + #define REISERFS_COMPR_FL FS_COMPR_FL + #define REISERFS_NOTAIL_FL FS_NOTAIL_FL + ++/* unfortunately reiserfs sdattr is only 16 bit */ ++#define REISERFS_IXUNLINK_FL (FS_IXUNLINK_FL >> 16) ++#define REISERFS_BARRIER_FL (FS_BARRIER_FL >> 16) ++#define REISERFS_COW_FL (FS_COW_FL >> 16) ++ + /* persistent flags that file inherits from the parent directory */ + #define REISERFS_INHERIT_MASK ( REISERFS_IMMUTABLE_FL | \ + REISERFS_SYNC_FL | \ +@@ -985,6 +990,9 @@ struct stat_data_v1 { + REISERFS_COMPR_FL | \ + REISERFS_NOTAIL_FL ) + ++#define REISERFS_FL_USER_VISIBLE 0x80FF ++#define REISERFS_FL_USER_MODIFIABLE 0x80FF ++ + /* Stat Data on disk (reiserfs version of UFS disk inode minus the + address blocks) */ + struct stat_data { +@@ -2073,6 +2081,7 @@ static inline void reiserfs_update_sd(st + void sd_attrs_to_i_attrs(__u16 sd_attrs, struct inode *inode); + void i_attrs_to_sd_attrs(struct inode *inode, __u16 * sd_attrs); + int reiserfs_setattr(struct dentry *dentry, struct iattr *attr); ++int reiserfs_sync_flags(struct inode *inode, int, int); + + int __reiserfs_write_begin(struct page *page, unsigned from, unsigned len); + +diff -NurpP --minimal linux-3.0.9/include/linux/reiserfs_fs_sb.h linux-3.0.9-vs2.3.2.1/include/linux/reiserfs_fs_sb.h +--- linux-3.0.9/include/linux/reiserfs_fs_sb.h 2010-02-25 11:52:07.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/include/linux/reiserfs_fs_sb.h 2011-06-10 22:11:24.000000000 +0200 +@@ -476,6 +476,7 @@ enum reiserfs_mount_options { + REISERFS_EXPOSE_PRIVROOT, + REISERFS_BARRIER_NONE, + REISERFS_BARRIER_FLUSH, ++ REISERFS_TAGGED, + + /* Actions on error */ + REISERFS_ERROR_PANIC, +diff -NurpP --minimal linux-3.0.9/include/linux/sched.h linux-3.0.9-vs2.3.2.1/include/linux/sched.h +--- linux-3.0.9/include/linux/sched.h 2011-11-15 16:40:47.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/include/linux/sched.h 2011-10-18 13:51:13.000000000 +0200 +@@ -1406,6 +1406,14 @@ struct task_struct { + #endif + seccomp_t seccomp; + ++/* vserver context data */ ++ struct vx_info *vx_info; ++ struct nx_info *nx_info; ++ ++ xid_t xid; ++ nid_t nid; ++ tag_t tag; ++ + /* Thread group tracking */ + u32 parent_exec_id; + u32 self_exec_id; +@@ -1649,6 +1657,11 @@ struct pid_namespace; + pid_t __task_pid_nr_ns(struct task_struct *task, enum pid_type type, + struct pid_namespace *ns); + ++#include <linux/vserver/base.h> ++#include <linux/vserver/context.h> ++#include <linux/vserver/debug.h> ++#include <linux/vserver/pid.h> ++ + static inline pid_t task_pid_nr(struct task_struct *tsk) + { + return tsk->pid; +@@ -1662,7 +1675,8 @@ static inline pid_t task_pid_nr_ns(struc + + static inline pid_t task_pid_vnr(struct task_struct *tsk) + { +- return __task_pid_nr_ns(tsk, PIDTYPE_PID, NULL); ++ // return __task_pid_nr_ns(tsk, PIDTYPE_PID, NULL); ++ return vx_map_pid(__task_pid_nr_ns(tsk, PIDTYPE_PID, NULL)); + } + + +@@ -1675,7 +1689,7 @@ pid_t task_tgid_nr_ns(struct task_struct + + static inline pid_t task_tgid_vnr(struct task_struct *tsk) + { +- return pid_vnr(task_tgid(tsk)); ++ return vx_map_tgid(pid_vnr(task_tgid(tsk))); + } + + +diff -NurpP --minimal linux-3.0.9/include/linux/shmem_fs.h linux-3.0.9-vs2.3.2.1/include/linux/shmem_fs.h +--- linux-3.0.9/include/linux/shmem_fs.h 2011-07-22 11:18:11.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/include/linux/shmem_fs.h 2011-07-01 11:35:35.000000000 +0200 +@@ -12,6 +12,9 @@ + + #define SHMEM_SYMLINK_INLINE_LEN (SHMEM_NR_DIRECT * sizeof(swp_entry_t)) + ++#define TMPFS_SUPER_MAGIC 0x01021994 ++ ++ + struct shmem_inode_info { + spinlock_t lock; + unsigned long flags; +diff -NurpP --minimal linux-3.0.9/include/linux/stat.h linux-3.0.9-vs2.3.2.1/include/linux/stat.h +--- linux-3.0.9/include/linux/stat.h 2008-12-25 00:26:37.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/include/linux/stat.h 2011-06-10 22:11:24.000000000 +0200 +@@ -66,6 +66,7 @@ struct kstat { + unsigned int nlink; + uid_t uid; + gid_t gid; ++ tag_t tag; + dev_t rdev; + loff_t size; + struct timespec atime; +diff -NurpP --minimal linux-3.0.9/include/linux/sunrpc/auth.h linux-3.0.9-vs2.3.2.1/include/linux/sunrpc/auth.h +--- linux-3.0.9/include/linux/sunrpc/auth.h 2011-03-15 18:07:39.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/include/linux/sunrpc/auth.h 2011-06-10 22:11:24.000000000 +0200 +@@ -25,6 +25,7 @@ + struct auth_cred { + uid_t uid; + gid_t gid; ++ tag_t tag; + struct group_info *group_info; + unsigned char machine_cred : 1; + }; +diff -NurpP --minimal linux-3.0.9/include/linux/sunrpc/clnt.h linux-3.0.9-vs2.3.2.1/include/linux/sunrpc/clnt.h +--- linux-3.0.9/include/linux/sunrpc/clnt.h 2011-05-22 16:17:58.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/include/linux/sunrpc/clnt.h 2011-06-10 22:11:24.000000000 +0200 +@@ -49,7 +49,8 @@ struct rpc_clnt { + unsigned int cl_softrtry : 1,/* soft timeouts */ + cl_discrtry : 1,/* disconnect before retry */ + cl_autobind : 1,/* use getport() */ +- cl_chatty : 1;/* be verbose */ ++ cl_chatty : 1,/* be verbose */ ++ cl_tag : 1;/* context tagging */ + + struct rpc_rtt * cl_rtt; /* RTO estimator data */ + const struct rpc_timeout *cl_timeout; /* Timeout strategy */ +diff -NurpP --minimal linux-3.0.9/include/linux/syscalls.h linux-3.0.9-vs2.3.2.1/include/linux/syscalls.h +--- linux-3.0.9/include/linux/syscalls.h 2011-07-22 11:18:11.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/include/linux/syscalls.h 2011-06-10 22:11:24.000000000 +0200 +@@ -483,6 +483,8 @@ asmlinkage long sys_symlink(const char _ + asmlinkage long sys_unlink(const char __user *pathname); + asmlinkage long sys_rename(const char __user *oldname, + const char __user *newname); ++asmlinkage long sys_copyfile(const char __user *from, const char __user *to, ++ umode_t mode); + asmlinkage long sys_chmod(const char __user *filename, mode_t mode); + asmlinkage long sys_fchmod(unsigned int fd, mode_t mode); + +diff -NurpP --minimal linux-3.0.9/include/linux/sysctl.h linux-3.0.9-vs2.3.2.1/include/linux/sysctl.h +--- linux-3.0.9/include/linux/sysctl.h 2011-03-15 18:07:40.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/include/linux/sysctl.h 2011-06-10 22:11:24.000000000 +0200 +@@ -60,6 +60,7 @@ enum + CTL_ABI=9, /* Binary emulation */ + CTL_CPU=10, /* CPU stuff (speed scaling, etc) */ + CTL_ARLAN=254, /* arlan wireless driver */ ++ CTL_VSERVER=4242, /* Linux-VServer debug */ + CTL_S390DBF=5677, /* s390 debug */ + CTL_SUNRPC=7249, /* sunrpc debug */ + CTL_PM=9899, /* frv power management */ +@@ -94,6 +95,7 @@ enum + + KERN_PANIC=15, /* int: panic timeout */ + KERN_REALROOTDEV=16, /* real root device to mount after initrd */ ++ KERN_VSHELPER=17, /* string: path to vshelper policy agent */ + + KERN_SPARC_REBOOT=21, /* reboot command on Sparc */ + KERN_CTLALTDEL=22, /* int: allow ctl-alt-del to reboot */ +diff -NurpP --minimal linux-3.0.9/include/linux/sysfs.h linux-3.0.9-vs2.3.2.1/include/linux/sysfs.h +--- linux-3.0.9/include/linux/sysfs.h 2011-07-22 11:18:11.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/include/linux/sysfs.h 2011-06-22 12:39:15.000000000 +0200 +@@ -19,6 +19,8 @@ + #include <linux/kobject_ns.h> + #include <asm/atomic.h> + ++#define SYSFS_SUPER_MAGIC 0x62656572 ++ + struct kobject; + struct module; + enum kobj_ns_type; +diff -NurpP --minimal linux-3.0.9/include/linux/time.h linux-3.0.9-vs2.3.2.1/include/linux/time.h +--- linux-3.0.9/include/linux/time.h 2011-07-22 11:18:11.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/include/linux/time.h 2011-06-10 22:11:24.000000000 +0200 +@@ -256,6 +256,9 @@ static __always_inline void timespec_add + a->tv_sec += __iter_div_u64_rem(a->tv_nsec + ns, NSEC_PER_SEC, &ns); + a->tv_nsec = ns; + } ++ ++#include <linux/vs_time.h> ++ + #endif /* __KERNEL__ */ + + #define NFDBITS __NFDBITS +diff -NurpP --minimal linux-3.0.9/include/linux/types.h linux-3.0.9-vs2.3.2.1/include/linux/types.h +--- linux-3.0.9/include/linux/types.h 2011-05-22 16:17:58.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/include/linux/types.h 2011-06-10 22:11:24.000000000 +0200 +@@ -40,6 +40,9 @@ typedef __kernel_uid32_t uid_t; + typedef __kernel_gid32_t gid_t; + typedef __kernel_uid16_t uid16_t; + typedef __kernel_gid16_t gid16_t; ++typedef unsigned int xid_t; ++typedef unsigned int nid_t; ++typedef unsigned int tag_t; + + typedef unsigned long uintptr_t; + +diff -NurpP --minimal linux-3.0.9/include/linux/utsname.h linux-3.0.9-vs2.3.2.1/include/linux/utsname.h +--- linux-3.0.9/include/linux/utsname.h 2011-05-22 16:17:58.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/include/linux/utsname.h 2011-06-13 14:36:48.000000000 +0200 +@@ -54,7 +54,8 @@ static inline void get_uts_ns(struct uts + } + + extern struct uts_namespace *copy_utsname(unsigned long flags, +- struct task_struct *tsk); ++ struct uts_namespace *old_ns, ++ struct user_namespace *user_ns); + extern void free_uts_ns(struct kref *kref); + + static inline void put_uts_ns(struct uts_namespace *ns) +@@ -71,12 +72,13 @@ static inline void put_uts_ns(struct uts + } + + static inline struct uts_namespace *copy_utsname(unsigned long flags, +- struct task_struct *tsk) ++ struct uts_namespace *old_ns, ++ struct user_namespace *user_ns) + { + if (flags & CLONE_NEWUTS) + return ERR_PTR(-EINVAL); + +- return tsk->nsproxy->uts_ns; ++ return old_ns; + } + #endif + +diff -NurpP --minimal linux-3.0.9/include/linux/vroot.h linux-3.0.9-vs2.3.2.1/include/linux/vroot.h +--- linux-3.0.9/include/linux/vroot.h 1970-01-01 01:00:00.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/include/linux/vroot.h 2011-06-10 22:11:24.000000000 +0200 +@@ -0,0 +1,51 @@ ++ ++/* ++ * include/linux/vroot.h ++ * ++ * written by Herbert Pötzl, 9/11/2002 ++ * ported to 2.6 by Herbert Pötzl, 30/12/2004 ++ * ++ * Copyright (C) 2002-2007 by Herbert Pötzl. ++ * Redistribution of this file is permitted under the ++ * GNU General Public License. ++ */ ++ ++#ifndef _LINUX_VROOT_H ++#define _LINUX_VROOT_H ++ ++ ++#ifdef __KERNEL__ ++ ++/* Possible states of device */ ++enum { ++ Vr_unbound, ++ Vr_bound, ++}; ++ ++struct vroot_device { ++ int vr_number; ++ int vr_refcnt; ++ ++ struct semaphore vr_ctl_mutex; ++ struct block_device *vr_device; ++ int vr_state; ++}; ++ ++ ++typedef struct block_device *(vroot_grb_func)(struct block_device *); ++ ++extern int register_vroot_grb(vroot_grb_func *); ++extern int unregister_vroot_grb(vroot_grb_func *); ++ ++#endif /* __KERNEL__ */ ++ ++#define MAX_VROOT_DEFAULT 8 ++ ++/* ++ * IOCTL commands --- we will commandeer 0x56 ('V') ++ */ ++ ++#define VROOT_SET_DEV 0x5600 ++#define VROOT_CLR_DEV 0x5601 ++ ++#endif /* _LINUX_VROOT_H */ +diff -NurpP --minimal linux-3.0.9/include/linux/vs_base.h linux-3.0.9-vs2.3.2.1/include/linux/vs_base.h +--- linux-3.0.9/include/linux/vs_base.h 1970-01-01 01:00:00.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/include/linux/vs_base.h 2011-06-10 22:11:24.000000000 +0200 +@@ -0,0 +1,10 @@ ++#ifndef _VS_BASE_H ++#define _VS_BASE_H ++ ++#include "vserver/base.h" ++#include "vserver/check.h" ++#include "vserver/debug.h" ++ ++#else ++#warning duplicate inclusion ++#endif +diff -NurpP --minimal linux-3.0.9/include/linux/vs_context.h linux-3.0.9-vs2.3.2.1/include/linux/vs_context.h +--- linux-3.0.9/include/linux/vs_context.h 1970-01-01 01:00:00.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/include/linux/vs_context.h 2011-06-10 22:11:24.000000000 +0200 +@@ -0,0 +1,242 @@ ++#ifndef _VS_CONTEXT_H ++#define _VS_CONTEXT_H ++ ++#include "vserver/base.h" ++#include "vserver/check.h" ++#include "vserver/context.h" ++#include "vserver/history.h" ++#include "vserver/debug.h" ++ ++#include <linux/sched.h> ++ ++ ++#define get_vx_info(i) __get_vx_info(i, __FILE__, __LINE__, __HERE__) ++ ++static inline struct vx_info *__get_vx_info(struct vx_info *vxi, ++ const char *_file, int _line, void *_here) ++{ ++ if (!vxi) ++ return NULL; ++ ++ vxlprintk(VXD_CBIT(xid, 2), "get_vx_info(%p[#%d.%d])", ++ vxi, vxi ? vxi->vx_id : 0, ++ vxi ? atomic_read(&vxi->vx_usecnt) : 0, ++ _file, _line); ++ __vxh_get_vx_info(vxi, _here); ++ ++ atomic_inc(&vxi->vx_usecnt); ++ return vxi; ++} ++ ++ ++extern void free_vx_info(struct vx_info *); ++ ++#define put_vx_info(i) __put_vx_info(i, __FILE__, __LINE__, __HERE__) ++ ++static inline void __put_vx_info(struct vx_info *vxi, ++ const char *_file, int _line, void *_here) ++{ ++ if (!vxi) ++ return; ++ ++ vxlprintk(VXD_CBIT(xid, 2), "put_vx_info(%p[#%d.%d])", ++ vxi, vxi ? vxi->vx_id : 0, ++ vxi ? atomic_read(&vxi->vx_usecnt) : 0, ++ _file, _line); ++ __vxh_put_vx_info(vxi, _here); ++ ++ if (atomic_dec_and_test(&vxi->vx_usecnt)) ++ free_vx_info(vxi); ++} ++ ++ ++#define init_vx_info(p, i) \ ++ __init_vx_info(p, i, __FILE__, __LINE__, __HERE__) ++ ++static inline void __init_vx_info(struct vx_info **vxp, struct vx_info *vxi, ++ const char *_file, int _line, void *_here) ++{ ++ if (vxi) { ++ vxlprintk(VXD_CBIT(xid, 3), ++ "init_vx_info(%p[#%d.%d])", ++ vxi, vxi ? vxi->vx_id : 0, ++ vxi ? atomic_read(&vxi->vx_usecnt) : 0, ++ _file, _line); ++ __vxh_init_vx_info(vxi, vxp, _here); ++ ++ atomic_inc(&vxi->vx_usecnt); ++ } ++ *vxp = vxi; ++} ++ ++ ++#define set_vx_info(p, i) \ ++ __set_vx_info(p, i, __FILE__, __LINE__, __HERE__) ++ ++static inline void __set_vx_info(struct vx_info **vxp, struct vx_info *vxi, ++ const char *_file, int _line, void *_here) ++{ ++ struct vx_info *vxo; ++ ++ if (!vxi) ++ return; ++ ++ vxlprintk(VXD_CBIT(xid, 3), "set_vx_info(%p[#%d.%d])", ++ vxi, vxi ? vxi->vx_id : 0, ++ vxi ? atomic_read(&vxi->vx_usecnt) : 0, ++ _file, _line); ++ __vxh_set_vx_info(vxi, vxp, _here); ++ ++ atomic_inc(&vxi->vx_usecnt); ++ vxo = xchg(vxp, vxi); ++ BUG_ON(vxo); ++} ++ ++ ++#define clr_vx_info(p) __clr_vx_info(p, __FILE__, __LINE__, __HERE__) ++ ++static inline void __clr_vx_info(struct vx_info **vxp, ++ const char *_file, int _line, void *_here) ++{ ++ struct vx_info *vxo; ++ ++ vxo = xchg(vxp, NULL); ++ if (!vxo) ++ return; ++ ++ vxlprintk(VXD_CBIT(xid, 3), "clr_vx_info(%p[#%d.%d])", ++ vxo, vxo ? vxo->vx_id : 0, ++ vxo ? atomic_read(&vxo->vx_usecnt) : 0, ++ _file, _line); ++ __vxh_clr_vx_info(vxo, vxp, _here); ++ ++ if (atomic_dec_and_test(&vxo->vx_usecnt)) ++ free_vx_info(vxo); ++} ++ ++ ++#define claim_vx_info(v, p) \ ++ __claim_vx_info(v, p, __FILE__, __LINE__, __HERE__) ++ ++static inline void __claim_vx_info(struct vx_info *vxi, ++ struct task_struct *task, ++ const char *_file, int _line, void *_here) ++{ ++ vxlprintk(VXD_CBIT(xid, 3), "claim_vx_info(%p[#%d.%d.%d]) %p", ++ vxi, vxi ? vxi->vx_id : 0, ++ vxi ? atomic_read(&vxi->vx_usecnt) : 0, ++ vxi ? atomic_read(&vxi->vx_tasks) : 0, ++ task, _file, _line); ++ __vxh_claim_vx_info(vxi, task, _here); ++ ++ atomic_inc(&vxi->vx_tasks); ++} ++ ++ ++extern void unhash_vx_info(struct vx_info *); ++ ++#define release_vx_info(v, p) \ ++ __release_vx_info(v, p, __FILE__, __LINE__, __HERE__) ++ ++static inline void __release_vx_info(struct vx_info *vxi, ++ struct task_struct *task, ++ const char *_file, int _line, void *_here) ++{ ++ vxlprintk(VXD_CBIT(xid, 3), "release_vx_info(%p[#%d.%d.%d]) %p", ++ vxi, vxi ? vxi->vx_id : 0, ++ vxi ? atomic_read(&vxi->vx_usecnt) : 0, ++ vxi ? atomic_read(&vxi->vx_tasks) : 0, ++ task, _file, _line); ++ __vxh_release_vx_info(vxi, task, _here); ++ ++ might_sleep(); ++ ++ if (atomic_dec_and_test(&vxi->vx_tasks)) ++ unhash_vx_info(vxi); ++} ++ ++ ++#define task_get_vx_info(p) \ ++ __task_get_vx_info(p, __FILE__, __LINE__, __HERE__) ++ ++static inline struct vx_info *__task_get_vx_info(struct task_struct *p, ++ const char *_file, int _line, void *_here) ++{ ++ struct vx_info *vxi; ++ ++ task_lock(p); ++ vxlprintk(VXD_CBIT(xid, 5), "task_get_vx_info(%p)", ++ p, _file, _line); ++ vxi = __get_vx_info(p->vx_info, _file, _line, _here); ++ task_unlock(p); ++ return vxi; ++} ++ ++ ++static inline void __wakeup_vx_info(struct vx_info *vxi) ++{ ++ if (waitqueue_active(&vxi->vx_wait)) ++ wake_up_interruptible(&vxi->vx_wait); ++} ++ ++ ++#define enter_vx_info(v, s) __enter_vx_info(v, s, __FILE__, __LINE__) ++ ++static inline void __enter_vx_info(struct vx_info *vxi, ++ struct vx_info_save *vxis, const char *_file, int _line) ++{ ++ vxlprintk(VXD_CBIT(xid, 5), "enter_vx_info(%p[#%d],%p) %p[#%d,%p]", ++ vxi, vxi ? vxi->vx_id : 0, vxis, current, ++ current->xid, current->vx_info, _file, _line); ++ vxis->vxi = xchg(¤t->vx_info, vxi); ++ vxis->xid = current->xid; ++ current->xid = vxi ? vxi->vx_id : 0; ++} ++ ++#define leave_vx_info(s) __leave_vx_info(s, __FILE__, __LINE__) ++ ++static inline void __leave_vx_info(struct vx_info_save *vxis, ++ const char *_file, int _line) ++{ ++ vxlprintk(VXD_CBIT(xid, 5), "leave_vx_info(%p[#%d,%p]) %p[#%d,%p]", ++ vxis, vxis->xid, vxis->vxi, current, ++ current->xid, current->vx_info, _file, _line); ++ (void)xchg(¤t->vx_info, vxis->vxi); ++ current->xid = vxis->xid; ++} ++ ++ ++static inline void __enter_vx_admin(struct vx_info_save *vxis) ++{ ++ vxis->vxi = xchg(¤t->vx_info, NULL); ++ vxis->xid = xchg(¤t->xid, (xid_t)0); ++} ++ ++static inline void __leave_vx_admin(struct vx_info_save *vxis) ++{ ++ (void)xchg(¤t->xid, vxis->xid); ++ (void)xchg(¤t->vx_info, vxis->vxi); ++} ++ ++#define task_is_init(p) \ ++ __task_is_init(p, __FILE__, __LINE__, __HERE__) ++ ++static inline int __task_is_init(struct task_struct *p, ++ const char *_file, int _line, void *_here) ++{ ++ int is_init = is_global_init(p); ++ ++ task_lock(p); ++ if (p->vx_info) ++ is_init = p->vx_info->vx_initpid == p->pid; ++ task_unlock(p); ++ return is_init; ++} ++ ++extern void exit_vx_info(struct task_struct *, int); ++extern void exit_vx_info_early(struct task_struct *, int); ++ ++ ++#else ++#warning duplicate inclusion ++#endif +diff -NurpP --minimal linux-3.0.9/include/linux/vs_cowbl.h linux-3.0.9-vs2.3.2.1/include/linux/vs_cowbl.h +--- linux-3.0.9/include/linux/vs_cowbl.h 1970-01-01 01:00:00.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/include/linux/vs_cowbl.h 2011-06-10 22:11:24.000000000 +0200 +@@ -0,0 +1,48 @@ ++#ifndef _VS_COWBL_H ++#define _VS_COWBL_H ++ ++#include <linux/fs.h> ++#include <linux/dcache.h> ++#include <linux/namei.h> ++#include <linux/slab.h> ++ ++extern struct dentry *cow_break_link(const char *pathname); ++ ++static inline int cow_check_and_break(struct path *path) ++{ ++ struct inode *inode = path->dentry->d_inode; ++ int error = 0; ++ ++ /* do we need this check? */ ++ if (IS_RDONLY(inode)) ++ return -EROFS; ++ ++ if (IS_COW(inode)) { ++ if (IS_COW_LINK(inode)) { ++ struct dentry *new_dentry, *old_dentry = path->dentry; ++ char *pp, *buf; ++ ++ buf = kmalloc(PATH_MAX, GFP_KERNEL); ++ if (!buf) { ++ return -ENOMEM; ++ } ++ pp = d_path(path, buf, PATH_MAX); ++ new_dentry = cow_break_link(pp); ++ kfree(buf); ++ if (!IS_ERR(new_dentry)) { ++ path->dentry = new_dentry; ++ dput(old_dentry); ++ } else ++ error = PTR_ERR(new_dentry); ++ } else { ++ inode->i_flags &= ~(S_IXUNLINK | S_IMMUTABLE); ++ inode->i_ctime = CURRENT_TIME; ++ mark_inode_dirty(inode); ++ } ++ } ++ return error; ++} ++ ++#else ++#warning duplicate inclusion ++#endif +diff -NurpP --minimal linux-3.0.9/include/linux/vs_cvirt.h linux-3.0.9-vs2.3.2.1/include/linux/vs_cvirt.h +--- linux-3.0.9/include/linux/vs_cvirt.h 1970-01-01 01:00:00.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/include/linux/vs_cvirt.h 2011-06-10 22:11:24.000000000 +0200 +@@ -0,0 +1,50 @@ ++#ifndef _VS_CVIRT_H ++#define _VS_CVIRT_H ++ ++#include "vserver/cvirt.h" ++#include "vserver/context.h" ++#include "vserver/base.h" ++#include "vserver/check.h" ++#include "vserver/debug.h" ++ ++ ++static inline void vx_activate_task(struct task_struct *p) ++{ ++ struct vx_info *vxi; ++ ++ if ((vxi = p->vx_info)) { ++ vx_update_load(vxi); ++ atomic_inc(&vxi->cvirt.nr_running); ++ } ++} ++ ++static inline void vx_deactivate_task(struct task_struct *p) ++{ ++ struct vx_info *vxi; ++ ++ if ((vxi = p->vx_info)) { ++ vx_update_load(vxi); ++ atomic_dec(&vxi->cvirt.nr_running); ++ } ++} ++ ++static inline void vx_uninterruptible_inc(struct task_struct *p) ++{ ++ struct vx_info *vxi; ++ ++ if ((vxi = p->vx_info)) ++ atomic_inc(&vxi->cvirt.nr_uninterruptible); ++} ++ ++static inline void vx_uninterruptible_dec(struct task_struct *p) ++{ ++ struct vx_info *vxi; ++ ++ if ((vxi = p->vx_info)) ++ atomic_dec(&vxi->cvirt.nr_uninterruptible); ++} ++ ++ ++#else ++#warning duplicate inclusion ++#endif +diff -NurpP --minimal linux-3.0.9/include/linux/vs_device.h linux-3.0.9-vs2.3.2.1/include/linux/vs_device.h +--- linux-3.0.9/include/linux/vs_device.h 1970-01-01 01:00:00.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/include/linux/vs_device.h 2011-06-10 22:11:24.000000000 +0200 +@@ -0,0 +1,45 @@ ++#ifndef _VS_DEVICE_H ++#define _VS_DEVICE_H ++ ++#include "vserver/base.h" ++#include "vserver/device.h" ++#include "vserver/debug.h" ++ ++ ++#ifdef CONFIG_VSERVER_DEVICE ++ ++int vs_map_device(struct vx_info *, dev_t, dev_t *, umode_t); ++ ++#define vs_device_perm(v, d, m, p) \ ++ ((vs_map_device(current_vx_info(), d, NULL, m) & (p)) == (p)) ++ ++#else ++ ++static inline ++int vs_map_device(struct vx_info *vxi, ++ dev_t device, dev_t *target, umode_t mode) ++{ ++ if (target) ++ *target = device; ++ return ~0; ++} ++ ++#define vs_device_perm(v, d, m, p) ((p) == (p)) ++ ++#endif ++ ++ ++#define vs_map_chrdev(d, t, p) \ ++ ((vs_map_device(current_vx_info(), d, t, S_IFCHR) & (p)) == (p)) ++#define vs_map_blkdev(d, t, p) \ ++ ((vs_map_device(current_vx_info(), d, t, S_IFBLK) & (p)) == (p)) ++ ++#define vs_chrdev_perm(d, p) \ ++ vs_device_perm(current_vx_info(), d, S_IFCHR, p) ++#define vs_blkdev_perm(d, p) \ ++ vs_device_perm(current_vx_info(), d, S_IFBLK, p) ++ ++ ++#else ++#warning duplicate inclusion ++#endif +diff -NurpP --minimal linux-3.0.9/include/linux/vs_dlimit.h linux-3.0.9-vs2.3.2.1/include/linux/vs_dlimit.h +--- linux-3.0.9/include/linux/vs_dlimit.h 1970-01-01 01:00:00.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/include/linux/vs_dlimit.h 2011-06-10 22:11:24.000000000 +0200 +@@ -0,0 +1,215 @@ ++#ifndef _VS_DLIMIT_H ++#define _VS_DLIMIT_H ++ ++#include <linux/fs.h> ++ ++#include "vserver/dlimit.h" ++#include "vserver/base.h" ++#include "vserver/debug.h" ++ ++ ++#define get_dl_info(i) __get_dl_info(i, __FILE__, __LINE__) ++ ++static inline struct dl_info *__get_dl_info(struct dl_info *dli, ++ const char *_file, int _line) ++{ ++ if (!dli) ++ return NULL; ++ vxlprintk(VXD_CBIT(dlim, 4), "get_dl_info(%p[#%d.%d])", ++ dli, dli ? dli->dl_tag : 0, ++ dli ? atomic_read(&dli->dl_usecnt) : 0, ++ _file, _line); ++ atomic_inc(&dli->dl_usecnt); ++ return dli; ++} ++ ++ ++#define free_dl_info(i) \ ++ call_rcu(&(i)->dl_rcu, rcu_free_dl_info) ++ ++#define put_dl_info(i) __put_dl_info(i, __FILE__, __LINE__) ++ ++static inline void __put_dl_info(struct dl_info *dli, ++ const char *_file, int _line) ++{ ++ if (!dli) ++ return; ++ vxlprintk(VXD_CBIT(dlim, 4), "put_dl_info(%p[#%d.%d])", ++ dli, dli ? dli->dl_tag : 0, ++ dli ? atomic_read(&dli->dl_usecnt) : 0, ++ _file, _line); ++ if (atomic_dec_and_test(&dli->dl_usecnt)) ++ free_dl_info(dli); ++} ++ ++ ++#define __dlimit_char(d) ((d) ? '*' : ' ') ++ ++static inline int __dl_alloc_space(struct super_block *sb, ++ tag_t tag, dlsize_t nr, const char *file, int line) ++{ ++ struct dl_info *dli = NULL; ++ int ret = 0; ++ ++ if (nr == 0) ++ goto out; ++ dli = locate_dl_info(sb, tag); ++ if (!dli) ++ goto out; ++ ++ spin_lock(&dli->dl_lock); ++ ret = (dli->dl_space_used + nr > dli->dl_space_total); ++ if (!ret) ++ dli->dl_space_used += nr; ++ spin_unlock(&dli->dl_lock); ++ put_dl_info(dli); ++out: ++ vxlprintk(VXD_CBIT(dlim, 1), ++ "ALLOC (%p,#%d)%c %lld bytes (%d)", ++ sb, tag, __dlimit_char(dli), (long long)nr, ++ ret, file, line); ++ return ret ? -ENOSPC : 0; ++} ++ ++static inline void __dl_free_space(struct super_block *sb, ++ tag_t tag, dlsize_t nr, const char *_file, int _line) ++{ ++ struct dl_info *dli = NULL; ++ ++ if (nr == 0) ++ goto out; ++ dli = locate_dl_info(sb, tag); ++ if (!dli) ++ goto out; ++ ++ spin_lock(&dli->dl_lock); ++ if (dli->dl_space_used > nr) ++ dli->dl_space_used -= nr; ++ else ++ dli->dl_space_used = 0; ++ spin_unlock(&dli->dl_lock); ++ put_dl_info(dli); ++out: ++ vxlprintk(VXD_CBIT(dlim, 1), ++ "FREE (%p,#%d)%c %lld bytes", ++ sb, tag, __dlimit_char(dli), (long long)nr, ++ _file, _line); ++} ++ ++static inline int __dl_alloc_inode(struct super_block *sb, ++ tag_t tag, const char *_file, int _line) ++{ ++ struct dl_info *dli; ++ int ret = 0; ++ ++ dli = locate_dl_info(sb, tag); ++ if (!dli) ++ goto out; ++ ++ spin_lock(&dli->dl_lock); ++ dli->dl_inodes_used++; ++ ret = (dli->dl_inodes_used > dli->dl_inodes_total); ++ spin_unlock(&dli->dl_lock); ++ put_dl_info(dli); ++out: ++ vxlprintk(VXD_CBIT(dlim, 0), ++ "ALLOC (%p,#%d)%c inode (%d)", ++ sb, tag, __dlimit_char(dli), ret, _file, _line); ++ return ret ? -ENOSPC : 0; ++} ++ ++static inline void __dl_free_inode(struct super_block *sb, ++ tag_t tag, const char *_file, int _line) ++{ ++ struct dl_info *dli; ++ ++ dli = locate_dl_info(sb, tag); ++ if (!dli) ++ goto out; ++ ++ spin_lock(&dli->dl_lock); ++ if (dli->dl_inodes_used > 1) ++ dli->dl_inodes_used--; ++ else ++ dli->dl_inodes_used = 0; ++ spin_unlock(&dli->dl_lock); ++ put_dl_info(dli); ++out: ++ vxlprintk(VXD_CBIT(dlim, 0), ++ "FREE (%p,#%d)%c inode", ++ sb, tag, __dlimit_char(dli), _file, _line); ++} ++ ++static inline void __dl_adjust_block(struct super_block *sb, tag_t tag, ++ unsigned long long *free_blocks, unsigned long long *root_blocks, ++ const char *_file, int _line) ++{ ++ struct dl_info *dli; ++ uint64_t broot, bfree; ++ ++ dli = locate_dl_info(sb, tag); ++ if (!dli) ++ return; ++ ++ spin_lock(&dli->dl_lock); ++ broot = (dli->dl_space_total - ++ (dli->dl_space_total >> 10) * dli->dl_nrlmult) ++ >> sb->s_blocksize_bits; ++ bfree = (dli->dl_space_total - dli->dl_space_used) ++ >> sb->s_blocksize_bits; ++ spin_unlock(&dli->dl_lock); ++ ++ vxlprintk(VXD_CBIT(dlim, 2), ++ "ADJUST: %lld,%lld on %lld,%lld [mult=%d]", ++ (long long)bfree, (long long)broot, ++ *free_blocks, *root_blocks, dli->dl_nrlmult, ++ _file, _line); ++ if (free_blocks) { ++ if (*free_blocks > bfree) ++ *free_blocks = bfree; ++ } ++ if (root_blocks) { ++ if (*root_blocks > broot) ++ *root_blocks = broot; ++ } ++ put_dl_info(dli); ++} ++ ++#define dl_prealloc_space(in, bytes) \ ++ __dl_alloc_space((in)->i_sb, (in)->i_tag, (dlsize_t)(bytes), \ ++ __FILE__, __LINE__ ) ++ ++#define dl_alloc_space(in, bytes) \ ++ __dl_alloc_space((in)->i_sb, (in)->i_tag, (dlsize_t)(bytes), \ ++ __FILE__, __LINE__ ) ++ ++#define dl_reserve_space(in, bytes) \ ++ __dl_alloc_space((in)->i_sb, (in)->i_tag, (dlsize_t)(bytes), \ ++ __FILE__, __LINE__ ) ++ ++#define dl_claim_space(in, bytes) (0) ++ ++#define dl_release_space(in, bytes) \ ++ __dl_free_space((in)->i_sb, (in)->i_tag, (dlsize_t)(bytes), \ ++ __FILE__, __LINE__ ) ++ ++#define dl_free_space(in, bytes) \ ++ __dl_free_space((in)->i_sb, (in)->i_tag, (dlsize_t)(bytes), \ ++ __FILE__, __LINE__ ) ++ ++ ++ ++#define dl_alloc_inode(in) \ ++ __dl_alloc_inode((in)->i_sb, (in)->i_tag, __FILE__, __LINE__ ) ++ ++#define dl_free_inode(in) \ ++ __dl_free_inode((in)->i_sb, (in)->i_tag, __FILE__, __LINE__ ) ++ ++ ++#define dl_adjust_block(sb, tag, fb, rb) \ ++ __dl_adjust_block(sb, tag, fb, rb, __FILE__, __LINE__ ) ++ ++ ++#else ++#warning duplicate inclusion ++#endif +diff -NurpP --minimal linux-3.0.9/include/linux/vs_inet.h linux-3.0.9-vs2.3.2.1/include/linux/vs_inet.h +--- linux-3.0.9/include/linux/vs_inet.h 1970-01-01 01:00:00.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/include/linux/vs_inet.h 2011-08-18 17:01:00.000000000 +0200 +@@ -0,0 +1,353 @@ ++#ifndef _VS_INET_H ++#define _VS_INET_H ++ ++#include "vserver/base.h" ++#include "vserver/network.h" ++#include "vserver/debug.h" ++ ++#define IPI_LOOPBACK htonl(INADDR_LOOPBACK) ++ ++#define NXAV4(a) NIPQUAD((a)->ip[0]), NIPQUAD((a)->ip[1]), \ ++ NIPQUAD((a)->mask), (a)->type ++#define NXAV4_FMT "[" NIPQUAD_FMT "-" NIPQUAD_FMT "/" NIPQUAD_FMT ":%04x]" ++ ++#define NIPQUAD(addr) \ ++ ((unsigned char *)&addr)[0], \ ++ ((unsigned char *)&addr)[1], \ ++ ((unsigned char *)&addr)[2], \ ++ ((unsigned char *)&addr)[3] ++ ++#define NIPQUAD_FMT "%u.%u.%u.%u" ++ ++ ++static inline ++int v4_addr_match(struct nx_addr_v4 *nxa, __be32 addr, uint16_t tmask) ++{ ++ __be32 ip = nxa->ip[0].s_addr; ++ __be32 mask = nxa->mask.s_addr; ++ __be32 bcast = ip | ~mask; ++ int ret = 0; ++ ++ switch (nxa->type & tmask) { ++ case NXA_TYPE_MASK: ++ ret = (ip == (addr & mask)); ++ break; ++ case NXA_TYPE_ADDR: ++ ret = 3; ++ if (addr == ip) ++ break; ++ /* fall through to broadcast */ ++ case NXA_MOD_BCAST: ++ ret = ((tmask & NXA_MOD_BCAST) && (addr == bcast)); ++ break; ++ case NXA_TYPE_RANGE: ++ ret = ((nxa->ip[0].s_addr <= addr) && ++ (nxa->ip[1].s_addr > addr)); ++ break; ++ case NXA_TYPE_ANY: ++ ret = 2; ++ break; ++ } ++ ++ vxdprintk(VXD_CBIT(net, 0), ++ "v4_addr_match(%p" NXAV4_FMT "," NIPQUAD_FMT ",%04x) = %d", ++ nxa, NXAV4(nxa), NIPQUAD(addr), tmask, ret); ++ return ret; ++} ++ ++static inline ++int v4_addr_in_nx_info(struct nx_info *nxi, __be32 addr, uint16_t tmask) ++{ ++ struct nx_addr_v4 *nxa; ++ int ret = 1; ++ ++ if (!nxi) ++ goto out; ++ ++ ret = 2; ++ /* allow 127.0.0.1 when remapping lback */ ++ if ((tmask & NXA_LOOPBACK) && ++ (addr == IPI_LOOPBACK) && ++ nx_info_flags(nxi, NXF_LBACK_REMAP, 0)) ++ goto out; ++ ret = 3; ++ /* check for lback address */ ++ if ((tmask & NXA_MOD_LBACK) && ++ (nxi->v4_lback.s_addr == addr)) ++ goto out; ++ ret = 4; ++ /* check for broadcast address */ ++ if ((tmask & NXA_MOD_BCAST) && ++ (nxi->v4_bcast.s_addr == addr)) ++ goto out; ++ ret = 5; ++ /* check for v4 addresses */ ++ for (nxa = &nxi->v4; nxa; nxa = nxa->next) ++ if (v4_addr_match(nxa, addr, tmask)) ++ goto out; ++ ret = 0; ++out: ++ vxdprintk(VXD_CBIT(net, 0), ++ "v4_addr_in_nx_info(%p[#%u]," NIPQUAD_FMT ",%04x) = %d", ++ nxi, nxi ? nxi->nx_id : 0, NIPQUAD(addr), tmask, ret); ++ return ret; ++} ++ ++static inline ++int v4_nx_addr_match(struct nx_addr_v4 *nxa, struct nx_addr_v4 *addr, uint16_t mask) ++{ ++ /* FIXME: needs full range checks */ ++ return v4_addr_match(nxa, addr->ip[0].s_addr, mask); ++} ++ ++static inline ++int v4_nx_addr_in_nx_info(struct nx_info *nxi, struct nx_addr_v4 *nxa, uint16_t mask) ++{ ++ struct nx_addr_v4 *ptr; ++ ++ for (ptr = &nxi->v4; ptr; ptr = ptr->next) ++ if (v4_nx_addr_match(ptr, nxa, mask)) ++ return 1; ++ return 0; ++} ++ ++#include <net/inet_sock.h> ++ ++/* ++ * Check if a given address matches for a socket ++ * ++ * nxi: the socket's nx_info if any ++ * addr: to be verified address ++ */ ++static inline ++int v4_sock_addr_match ( ++ struct nx_info *nxi, ++ struct inet_sock *inet, ++ __be32 addr) ++{ ++ __be32 saddr = inet->inet_rcv_saddr; ++ __be32 bcast = nxi ? nxi->v4_bcast.s_addr : INADDR_BROADCAST; ++ ++ if (addr && (saddr == addr || bcast == addr)) ++ return 1; ++ if (!saddr) ++ return v4_addr_in_nx_info(nxi, addr, NXA_MASK_BIND); ++ return 0; ++} ++ ++ ++/* inet related checks and helpers */ ++ ++ ++struct in_ifaddr; ++struct net_device; ++struct sock; ++ ++#ifdef CONFIG_INET ++ ++#include <linux/netdevice.h> ++#include <linux/inetdevice.h> ++#include <net/inet_sock.h> ++#include <net/inet_timewait_sock.h> ++ ++ ++int dev_in_nx_info(struct net_device *, struct nx_info *); ++int v4_dev_in_nx_info(struct net_device *, struct nx_info *); ++int nx_v4_addr_conflict(struct nx_info *, struct nx_info *); ++ ++ ++/* ++ * check if address is covered by socket ++ * ++ * sk: the socket to check against ++ * addr: the address in question (must be != 0) ++ */ ++ ++static inline ++int __v4_addr_match_socket(const struct sock *sk, struct nx_addr_v4 *nxa) ++{ ++ struct nx_info *nxi = sk->sk_nx_info; ++ __be32 saddr = sk_rcv_saddr(sk); ++ ++ vxdprintk(VXD_CBIT(net, 5), ++ "__v4_addr_in_socket(%p," NXAV4_FMT ") %p:" NIPQUAD_FMT " %p;%lx", ++ sk, NXAV4(nxa), nxi, NIPQUAD(saddr), sk->sk_socket, ++ (sk->sk_socket?sk->sk_socket->flags:0)); ++ ++ if (saddr) { /* direct address match */ ++ return v4_addr_match(nxa, saddr, -1); ++ } else if (nxi) { /* match against nx_info */ ++ return v4_nx_addr_in_nx_info(nxi, nxa, -1); ++ } else { /* unrestricted any socket */ ++ return 1; ++ } ++} ++ ++ ++ ++static inline ++int nx_dev_visible(struct nx_info *nxi, struct net_device *dev) ++{ ++ vxdprintk(VXD_CBIT(net, 1), ++ "nx_dev_visible(%p[#%u],%p " VS_Q("%s") ") %d", ++ nxi, nxi ? nxi->nx_id : 0, dev, dev->name, ++ nxi ? dev_in_nx_info(dev, nxi) : 0); ++ ++ if (!nx_info_flags(nxi, NXF_HIDE_NETIF, 0)) ++ return 1; ++ if (dev_in_nx_info(dev, nxi)) ++ return 1; ++ return 0; ++} ++ ++ ++static inline ++int v4_ifa_in_nx_info(struct in_ifaddr *ifa, struct nx_info *nxi) ++{ ++ if (!nxi) ++ return 1; ++ if (!ifa) ++ return 0; ++ return v4_addr_in_nx_info(nxi, ifa->ifa_local, NXA_MASK_SHOW); ++} ++ ++static inline ++int nx_v4_ifa_visible(struct nx_info *nxi, struct in_ifaddr *ifa) ++{ ++ vxdprintk(VXD_CBIT(net, 1), "nx_v4_ifa_visible(%p[#%u],%p) %d", ++ nxi, nxi ? nxi->nx_id : 0, ifa, ++ nxi ? v4_ifa_in_nx_info(ifa, nxi) : 0); ++ ++ if (!nx_info_flags(nxi, NXF_HIDE_NETIF, 0)) ++ return 1; ++ if (v4_ifa_in_nx_info(ifa, nxi)) ++ return 1; ++ return 0; ++} ++ ++ ++struct nx_v4_sock_addr { ++ __be32 saddr; /* Address used for validation */ ++ __be32 baddr; /* Address used for socket bind */ ++}; ++ ++static inline ++int v4_map_sock_addr(struct inet_sock *inet, struct sockaddr_in *addr, ++ struct nx_v4_sock_addr *nsa) ++{ ++ struct sock *sk = &inet->sk; ++ struct nx_info *nxi = sk->sk_nx_info; ++ __be32 saddr = addr->sin_addr.s_addr; ++ __be32 baddr = saddr; ++ ++ vxdprintk(VXD_CBIT(net, 3), ++ "inet_bind(%p)* %p,%p;%lx " NIPQUAD_FMT, ++ sk, sk->sk_nx_info, sk->sk_socket, ++ (sk->sk_socket ? sk->sk_socket->flags : 0), ++ NIPQUAD(saddr)); ++ ++ if (nxi) { ++ if (saddr == INADDR_ANY) { ++ if (nx_info_flags(nxi, NXF_SINGLE_IP, 0)) ++ baddr = nxi->v4.ip[0].s_addr; ++ } else if (saddr == IPI_LOOPBACK) { ++ if (nx_info_flags(nxi, NXF_LBACK_REMAP, 0)) ++ baddr = nxi->v4_lback.s_addr; ++ } else if (!ipv4_is_multicast(saddr) || ++ !nx_info_ncaps(nxi, NXC_MULTICAST)) { ++ /* normal address bind */ ++ if (!v4_addr_in_nx_info(nxi, saddr, NXA_MASK_BIND)) ++ return -EADDRNOTAVAIL; ++ } ++ } ++ ++ vxdprintk(VXD_CBIT(net, 3), ++ "inet_bind(%p) " NIPQUAD_FMT ", " NIPQUAD_FMT, ++ sk, NIPQUAD(saddr), NIPQUAD(baddr)); ++ ++ nsa->saddr = saddr; ++ nsa->baddr = baddr; ++ return 0; ++} ++ ++static inline ++void v4_set_sock_addr(struct inet_sock *inet, struct nx_v4_sock_addr *nsa) ++{ ++ inet->inet_saddr = nsa->baddr; ++ inet->inet_rcv_saddr = nsa->baddr; ++} ++ ++ ++/* ++ * helper to simplify inet_lookup_listener ++ * ++ * nxi: the socket's nx_info if any ++ * addr: to be verified address ++ * saddr: socket address ++ */ ++static inline int v4_inet_addr_match ( ++ struct nx_info *nxi, ++ __be32 addr, ++ __be32 saddr) ++{ ++ if (addr && (saddr == addr)) ++ return 1; ++ if (!saddr) ++ return nxi ? v4_addr_in_nx_info(nxi, addr, NXA_MASK_BIND) : 1; ++ return 0; ++} ++ ++static inline __be32 nx_map_sock_lback(struct nx_info *nxi, __be32 addr) ++{ ++ if (nx_info_flags(nxi, NXF_HIDE_LBACK, 0) && ++ (addr == nxi->v4_lback.s_addr)) ++ return IPI_LOOPBACK; ++ return addr; ++} ++ ++static inline ++int nx_info_has_v4(struct nx_info *nxi) ++{ ++ if (!nxi) ++ return 1; ++ if (NX_IPV4(nxi)) ++ return 1; ++ if (nx_info_flags(nxi, NXF_LBACK_REMAP, 0)) ++ return 1; ++ return 0; ++} ++ ++#else /* CONFIG_INET */ ++ ++static inline ++int nx_dev_visible(struct nx_info *n, struct net_device *d) ++{ ++ return 1; ++} ++ ++static inline ++int nx_v4_addr_conflict(struct nx_info *n, uint32_t a, const struct sock *s) ++{ ++ return 1; ++} ++ ++static inline ++int v4_ifa_in_nx_info(struct in_ifaddr *a, struct nx_info *n) ++{ ++ return 1; ++} ++ ++static inline ++int nx_info_has_v4(struct nx_info *nxi) ++{ ++ return 0; ++} ++ ++#endif /* CONFIG_INET */ ++ ++#define current_nx_info_has_v4() \ ++ nx_info_has_v4(current_nx_info()) ++ ++#else ++// #warning duplicate inclusion ++#endif +diff -NurpP --minimal linux-3.0.9/include/linux/vs_inet6.h linux-3.0.9-vs2.3.2.1/include/linux/vs_inet6.h +--- linux-3.0.9/include/linux/vs_inet6.h 1970-01-01 01:00:00.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/include/linux/vs_inet6.h 2011-06-10 22:11:24.000000000 +0200 +@@ -0,0 +1,246 @@ ++#ifndef _VS_INET6_H ++#define _VS_INET6_H ++ ++#include "vserver/base.h" ++#include "vserver/network.h" ++#include "vserver/debug.h" ++ ++#include <net/ipv6.h> ++ ++#define NXAV6(a) &(a)->ip, &(a)->mask, (a)->prefix, (a)->type ++#define NXAV6_FMT "[%pI6/%pI6/%d:%04x]" ++ ++ ++#ifdef CONFIG_IPV6 ++ ++static inline ++int v6_addr_match(struct nx_addr_v6 *nxa, ++ const struct in6_addr *addr, uint16_t mask) ++{ ++ int ret = 0; ++ ++ switch (nxa->type & mask) { ++ case NXA_TYPE_MASK: ++ ret = ipv6_masked_addr_cmp(&nxa->ip, &nxa->mask, addr); ++ break; ++ case NXA_TYPE_ADDR: ++ ret = ipv6_addr_equal(&nxa->ip, addr); ++ break; ++ case NXA_TYPE_ANY: ++ ret = 1; ++ break; ++ } ++ vxdprintk(VXD_CBIT(net, 0), ++ "v6_addr_match(%p" NXAV6_FMT ",%pI6,%04x) = %d", ++ nxa, NXAV6(nxa), addr, mask, ret); ++ return ret; ++} ++ ++static inline ++int v6_addr_in_nx_info(struct nx_info *nxi, ++ const struct in6_addr *addr, uint16_t mask) ++{ ++ struct nx_addr_v6 *nxa; ++ int ret = 1; ++ ++ if (!nxi) ++ goto out; ++ for (nxa = &nxi->v6; nxa; nxa = nxa->next) ++ if (v6_addr_match(nxa, addr, mask)) ++ goto out; ++ ret = 0; ++out: ++ vxdprintk(VXD_CBIT(net, 0), ++ "v6_addr_in_nx_info(%p[#%u],%pI6,%04x) = %d", ++ nxi, nxi ? nxi->nx_id : 0, addr, mask, ret); ++ return ret; ++} ++ ++static inline ++int v6_nx_addr_match(struct nx_addr_v6 *nxa, struct nx_addr_v6 *addr, uint16_t mask) ++{ ++ /* FIXME: needs full range checks */ ++ return v6_addr_match(nxa, &addr->ip, mask); ++} ++ ++static inline ++int v6_nx_addr_in_nx_info(struct nx_info *nxi, struct nx_addr_v6 *nxa, uint16_t mask) ++{ ++ struct nx_addr_v6 *ptr; ++ ++ for (ptr = &nxi->v6; ptr; ptr = ptr->next) ++ if (v6_nx_addr_match(ptr, nxa, mask)) ++ return 1; ++ return 0; ++} ++ ++ ++/* ++ * Check if a given address matches for a socket ++ * ++ * nxi: the socket's nx_info if any ++ * addr: to be verified address ++ */ ++static inline ++int v6_sock_addr_match ( ++ struct nx_info *nxi, ++ struct inet_sock *inet, ++ struct in6_addr *addr) ++{ ++ struct sock *sk = &inet->sk; ++ struct in6_addr *saddr = inet6_rcv_saddr(sk); ++ ++ if (!ipv6_addr_any(addr) && ++ ipv6_addr_equal(saddr, addr)) ++ return 1; ++ if (ipv6_addr_any(saddr)) ++ return v6_addr_in_nx_info(nxi, addr, -1); ++ return 0; ++} ++ ++/* ++ * check if address is covered by socket ++ * ++ * sk: the socket to check against ++ * addr: the address in question (must be != 0) ++ */ ++ ++static inline ++int __v6_addr_match_socket(const struct sock *sk, struct nx_addr_v6 *nxa) ++{ ++ struct nx_info *nxi = sk->sk_nx_info; ++ struct in6_addr *saddr = inet6_rcv_saddr(sk); ++ ++ vxdprintk(VXD_CBIT(net, 5), ++ "__v6_addr_in_socket(%p," NXAV6_FMT ") %p:%pI6 %p;%lx", ++ sk, NXAV6(nxa), nxi, saddr, sk->sk_socket, ++ (sk->sk_socket?sk->sk_socket->flags:0)); ++ ++ if (!ipv6_addr_any(saddr)) { /* direct address match */ ++ return v6_addr_match(nxa, saddr, -1); ++ } else if (nxi) { /* match against nx_info */ ++ return v6_nx_addr_in_nx_info(nxi, nxa, -1); ++ } else { /* unrestricted any socket */ ++ return 1; ++ } ++} ++ ++ ++/* inet related checks and helpers */ ++ ++ ++struct in_ifaddr; ++struct net_device; ++struct sock; ++ ++ ++#include <linux/netdevice.h> ++#include <linux/inetdevice.h> ++#include <net/inet_timewait_sock.h> ++ ++ ++int dev_in_nx_info(struct net_device *, struct nx_info *); ++int v6_dev_in_nx_info(struct net_device *, struct nx_info *); ++int nx_v6_addr_conflict(struct nx_info *, struct nx_info *); ++ ++ ++ ++static inline ++int v6_ifa_in_nx_info(struct inet6_ifaddr *ifa, struct nx_info *nxi) ++{ ++ if (!nxi) ++ return 1; ++ if (!ifa) ++ return 0; ++ return v6_addr_in_nx_info(nxi, &ifa->addr, -1); ++} ++ ++static inline ++int nx_v6_ifa_visible(struct nx_info *nxi, struct inet6_ifaddr *ifa) ++{ ++ vxdprintk(VXD_CBIT(net, 1), "nx_v6_ifa_visible(%p[#%u],%p) %d", ++ nxi, nxi ? nxi->nx_id : 0, ifa, ++ nxi ? v6_ifa_in_nx_info(ifa, nxi) : 0); ++ ++ if (!nx_info_flags(nxi, NXF_HIDE_NETIF, 0)) ++ return 1; ++ if (v6_ifa_in_nx_info(ifa, nxi)) ++ return 1; ++ return 0; ++} ++ ++ ++struct nx_v6_sock_addr { ++ struct in6_addr saddr; /* Address used for validation */ ++ struct in6_addr baddr; /* Address used for socket bind */ ++}; ++ ++static inline ++int v6_map_sock_addr(struct inet_sock *inet, struct sockaddr_in6 *addr, ++ struct nx_v6_sock_addr *nsa) ++{ ++ // struct sock *sk = &inet->sk; ++ // struct nx_info *nxi = sk->sk_nx_info; ++ struct in6_addr saddr = addr->sin6_addr; ++ struct in6_addr baddr = saddr; ++ ++ nsa->saddr = saddr; ++ nsa->baddr = baddr; ++ return 0; ++} ++ ++static inline ++void v6_set_sock_addr(struct inet_sock *inet, struct nx_v6_sock_addr *nsa) ++{ ++ // struct sock *sk = &inet->sk; ++ // struct in6_addr *saddr = inet6_rcv_saddr(sk); ++ ++ // *saddr = nsa->baddr; ++ // inet->inet_saddr = nsa->baddr; ++} ++ ++static inline ++int nx_info_has_v6(struct nx_info *nxi) ++{ ++ if (!nxi) ++ return 1; ++ if (NX_IPV6(nxi)) ++ return 1; ++ return 0; ++} ++ ++#else /* CONFIG_IPV6 */ ++ ++static inline ++int nx_v6_dev_visible(struct nx_info *n, struct net_device *d) ++{ ++ return 1; ++} ++ ++ ++static inline ++int nx_v6_addr_conflict(struct nx_info *n, uint32_t a, const struct sock *s) ++{ ++ return 1; ++} ++ ++static inline ++int v6_ifa_in_nx_info(struct in_ifaddr *a, struct nx_info *n) ++{ ++ return 1; ++} ++ ++static inline ++int nx_info_has_v6(struct nx_info *nxi) ++{ ++ return 0; ++} ++ ++#endif /* CONFIG_IPV6 */ ++ ++#define current_nx_info_has_v6() \ ++ nx_info_has_v6(current_nx_info()) ++ ++#else ++#warning duplicate inclusion ++#endif +diff -NurpP --minimal linux-3.0.9/include/linux/vs_limit.h linux-3.0.9-vs2.3.2.1/include/linux/vs_limit.h +--- linux-3.0.9/include/linux/vs_limit.h 1970-01-01 01:00:00.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/include/linux/vs_limit.h 2011-06-10 22:11:24.000000000 +0200 +@@ -0,0 +1,140 @@ ++#ifndef _VS_LIMIT_H ++#define _VS_LIMIT_H ++ ++#include "vserver/limit.h" ++#include "vserver/base.h" ++#include "vserver/context.h" ++#include "vserver/debug.h" ++#include "vserver/context.h" ++#include "vserver/limit_int.h" ++ ++ ++#define vx_acc_cres(v, d, p, r) \ ++ __vx_acc_cres(v, r, d, p, __FILE__, __LINE__) ++ ++#define vx_acc_cres_cond(x, d, p, r) \ ++ __vx_acc_cres(((x) == vx_current_xid()) ? current_vx_info() : 0, \ ++ r, d, p, __FILE__, __LINE__) ++ ++ ++#define vx_add_cres(v, a, p, r) \ ++ __vx_add_cres(v, r, a, p, __FILE__, __LINE__) ++#define vx_sub_cres(v, a, p, r) vx_add_cres(v, -(a), p, r) ++ ++#define vx_add_cres_cond(x, a, p, r) \ ++ __vx_add_cres(((x) == vx_current_xid()) ? current_vx_info() : 0, \ ++ r, a, p, __FILE__, __LINE__) ++#define vx_sub_cres_cond(x, a, p, r) vx_add_cres_cond(x, -(a), p, r) ++ ++ ++/* process and file limits */ ++ ++#define vx_nproc_inc(p) \ ++ vx_acc_cres((p)->vx_info, 1, p, RLIMIT_NPROC) ++ ++#define vx_nproc_dec(p) \ ++ vx_acc_cres((p)->vx_info,-1, p, RLIMIT_NPROC) ++ ++#define vx_files_inc(f) \ ++ vx_acc_cres_cond((f)->f_xid, 1, f, RLIMIT_NOFILE) ++ ++#define vx_files_dec(f) \ ++ vx_acc_cres_cond((f)->f_xid,-1, f, RLIMIT_NOFILE) ++ ++#define vx_locks_inc(l) \ ++ vx_acc_cres_cond((l)->fl_xid, 1, l, RLIMIT_LOCKS) ++ ++#define vx_locks_dec(l) \ ++ vx_acc_cres_cond((l)->fl_xid,-1, l, RLIMIT_LOCKS) ++ ++#define vx_openfd_inc(f) \ ++ vx_acc_cres(current_vx_info(), 1, (void *)(long)(f), VLIMIT_OPENFD) ++ ++#define vx_openfd_dec(f) \ ++ vx_acc_cres(current_vx_info(),-1, (void *)(long)(f), VLIMIT_OPENFD) ++ ++ ++#define vx_cres_avail(v, n, r) \ ++ __vx_cres_avail(v, r, n, __FILE__, __LINE__) ++ ++ ++#define vx_nproc_avail(n) \ ++ vx_cres_avail(current_vx_info(), n, RLIMIT_NPROC) ++ ++#define vx_files_avail(n) \ ++ vx_cres_avail(current_vx_info(), n, RLIMIT_NOFILE) ++ ++#define vx_locks_avail(n) \ ++ vx_cres_avail(current_vx_info(), n, RLIMIT_LOCKS) ++ ++#define vx_openfd_avail(n) \ ++ vx_cres_avail(current_vx_info(), n, VLIMIT_OPENFD) ++ ++ ++/* dentry limits */ ++ ++#define vx_dentry_inc(d) do { \ ++ if ((d)->d_count == 1) \ ++ vx_acc_cres(current_vx_info(), 1, d, VLIMIT_DENTRY); \ ++ } while (0) ++ ++#define vx_dentry_dec(d) do { \ ++ if ((d)->d_count == 0) \ ++ vx_acc_cres(current_vx_info(),-1, d, VLIMIT_DENTRY); \ ++ } while (0) ++ ++#define vx_dentry_avail(n) \ ++ vx_cres_avail(current_vx_info(), n, VLIMIT_DENTRY) ++ ++ ++/* socket limits */ ++ ++#define vx_sock_inc(s) \ ++ vx_acc_cres((s)->sk_vx_info, 1, s, VLIMIT_NSOCK) ++ ++#define vx_sock_dec(s) \ ++ vx_acc_cres((s)->sk_vx_info,-1, s, VLIMIT_NSOCK) ++ ++#define vx_sock_avail(n) \ ++ vx_cres_avail(current_vx_info(), n, VLIMIT_NSOCK) ++ ++ ++/* ipc resource limits */ ++ ++#define vx_ipcmsg_add(v, u, a) \ ++ vx_add_cres(v, a, u, RLIMIT_MSGQUEUE) ++ ++#define vx_ipcmsg_sub(v, u, a) \ ++ vx_sub_cres(v, a, u, RLIMIT_MSGQUEUE) ++ ++#define vx_ipcmsg_avail(v, a) \ ++ vx_cres_avail(v, a, RLIMIT_MSGQUEUE) ++ ++ ++#define vx_ipcshm_add(v, k, a) \ ++ vx_add_cres(v, a, (void *)(long)(k), VLIMIT_SHMEM) ++ ++#define vx_ipcshm_sub(v, k, a) \ ++ vx_sub_cres(v, a, (void *)(long)(k), VLIMIT_SHMEM) ++ ++#define vx_ipcshm_avail(v, a) \ ++ vx_cres_avail(v, a, VLIMIT_SHMEM) ++ ++ ++#define vx_semary_inc(a) \ ++ vx_acc_cres(current_vx_info(), 1, a, VLIMIT_SEMARY) ++ ++#define vx_semary_dec(a) \ ++ vx_acc_cres(current_vx_info(), -1, a, VLIMIT_SEMARY) ++ ++ ++#define vx_nsems_add(a,n) \ ++ vx_add_cres(current_vx_info(), n, a, VLIMIT_NSEMS) ++ ++#define vx_nsems_sub(a,n) \ ++ vx_sub_cres(current_vx_info(), n, a, VLIMIT_NSEMS) ++ ++ ++#else ++#warning duplicate inclusion ++#endif +diff -NurpP --minimal linux-3.0.9/include/linux/vs_memory.h linux-3.0.9-vs2.3.2.1/include/linux/vs_memory.h +--- linux-3.0.9/include/linux/vs_memory.h 1970-01-01 01:00:00.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/include/linux/vs_memory.h 2011-06-10 22:11:24.000000000 +0200 +@@ -0,0 +1,58 @@ ++#ifndef _VS_MEMORY_H ++#define _VS_MEMORY_H ++ ++#include "vserver/limit.h" ++#include "vserver/base.h" ++#include "vserver/context.h" ++#include "vserver/debug.h" ++#include "vserver/context.h" ++#include "vserver/limit_int.h" ++ ++enum { ++ VXPT_UNKNOWN = 0, ++ VXPT_ANON, ++ VXPT_NONE, ++ VXPT_FILE, ++ VXPT_SWAP, ++ VXPT_WRITE ++}; ++ ++#if 0 ++#define vx_page_fault(mm, vma, type, ret) ++#else ++ ++static inline ++void __vx_page_fault(struct mm_struct *mm, ++ struct vm_area_struct *vma, int type, int ret) ++{ ++ struct vx_info *vxi = mm->mm_vx_info; ++ int what; ++/* ++ static char *page_type[6] = ++ { "UNKNOWN", "ANON", "NONE", "FILE", "SWAP", "WRITE" }; ++ static char *page_what[4] = ++ { "FAULT_OOM", "FAULT_SIGBUS", "FAULT_MINOR", "FAULT_MAJOR" }; ++*/ ++ ++ if (!vxi) ++ return; ++ ++ what = (ret & 0x3); ++ ++/* printk("[%d] page[%d][%d] %2x %s %s\n", vxi->vx_id, ++ type, what, ret, page_type[type], page_what[what]); ++*/ ++ if (ret & VM_FAULT_WRITE) ++ what |= 0x4; ++ atomic_inc(&vxi->cacct.page[type][what]); ++} ++ ++#define vx_page_fault(mm, vma, type, ret) __vx_page_fault(mm, vma, type, ret) ++#endif ++ ++ ++extern unsigned long vx_badness(struct task_struct *task, struct mm_struct *mm); ++ ++#else ++#warning duplicate inclusion ++#endif +diff -NurpP --minimal linux-3.0.9/include/linux/vs_network.h linux-3.0.9-vs2.3.2.1/include/linux/vs_network.h +--- linux-3.0.9/include/linux/vs_network.h 1970-01-01 01:00:00.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/include/linux/vs_network.h 2011-06-10 22:11:24.000000000 +0200 +@@ -0,0 +1,169 @@ ++#ifndef _NX_VS_NETWORK_H ++#define _NX_VS_NETWORK_H ++ ++#include "vserver/context.h" ++#include "vserver/network.h" ++#include "vserver/base.h" ++#include "vserver/check.h" ++#include "vserver/debug.h" ++ ++#include <linux/sched.h> ++ ++ ++#define get_nx_info(i) __get_nx_info(i, __FILE__, __LINE__) ++ ++static inline struct nx_info *__get_nx_info(struct nx_info *nxi, ++ const char *_file, int _line) ++{ ++ if (!nxi) ++ return NULL; ++ ++ vxlprintk(VXD_CBIT(nid, 2), "get_nx_info(%p[#%d.%d])", ++ nxi, nxi ? nxi->nx_id : 0, ++ nxi ? atomic_read(&nxi->nx_usecnt) : 0, ++ _file, _line); ++ ++ atomic_inc(&nxi->nx_usecnt); ++ return nxi; ++} ++ ++ ++extern void free_nx_info(struct nx_info *); ++ ++#define put_nx_info(i) __put_nx_info(i, __FILE__, __LINE__) ++ ++static inline void __put_nx_info(struct nx_info *nxi, const char *_file, int _line) ++{ ++ if (!nxi) ++ return; ++ ++ vxlprintk(VXD_CBIT(nid, 2), "put_nx_info(%p[#%d.%d])", ++ nxi, nxi ? nxi->nx_id : 0, ++ nxi ? atomic_read(&nxi->nx_usecnt) : 0, ++ _file, _line); ++ ++ if (atomic_dec_and_test(&nxi->nx_usecnt)) ++ free_nx_info(nxi); ++} ++ ++ ++#define init_nx_info(p, i) __init_nx_info(p, i, __FILE__, __LINE__) ++ ++static inline void __init_nx_info(struct nx_info **nxp, struct nx_info *nxi, ++ const char *_file, int _line) ++{ ++ if (nxi) { ++ vxlprintk(VXD_CBIT(nid, 3), ++ "init_nx_info(%p[#%d.%d])", ++ nxi, nxi ? nxi->nx_id : 0, ++ nxi ? atomic_read(&nxi->nx_usecnt) : 0, ++ _file, _line); ++ ++ atomic_inc(&nxi->nx_usecnt); ++ } ++ *nxp = nxi; ++} ++ ++ ++#define set_nx_info(p, i) __set_nx_info(p, i, __FILE__, __LINE__) ++ ++static inline void __set_nx_info(struct nx_info **nxp, struct nx_info *nxi, ++ const char *_file, int _line) ++{ ++ struct nx_info *nxo; ++ ++ if (!nxi) ++ return; ++ ++ vxlprintk(VXD_CBIT(nid, 3), "set_nx_info(%p[#%d.%d])", ++ nxi, nxi ? nxi->nx_id : 0, ++ nxi ? atomic_read(&nxi->nx_usecnt) : 0, ++ _file, _line); ++ ++ atomic_inc(&nxi->nx_usecnt); ++ nxo = xchg(nxp, nxi); ++ BUG_ON(nxo); ++} ++ ++#define clr_nx_info(p) __clr_nx_info(p, __FILE__, __LINE__) ++ ++static inline void __clr_nx_info(struct nx_info **nxp, ++ const char *_file, int _line) ++{ ++ struct nx_info *nxo; ++ ++ nxo = xchg(nxp, NULL); ++ if (!nxo) ++ return; ++ ++ vxlprintk(VXD_CBIT(nid, 3), "clr_nx_info(%p[#%d.%d])", ++ nxo, nxo ? nxo->nx_id : 0, ++ nxo ? atomic_read(&nxo->nx_usecnt) : 0, ++ _file, _line); ++ ++ if (atomic_dec_and_test(&nxo->nx_usecnt)) ++ free_nx_info(nxo); ++} ++ ++ ++#define claim_nx_info(v, p) __claim_nx_info(v, p, __FILE__, __LINE__) ++ ++static inline void __claim_nx_info(struct nx_info *nxi, ++ struct task_struct *task, const char *_file, int _line) ++{ ++ vxlprintk(VXD_CBIT(nid, 3), "claim_nx_info(%p[#%d.%d.%d]) %p", ++ nxi, nxi ? nxi->nx_id : 0, ++ nxi?atomic_read(&nxi->nx_usecnt):0, ++ nxi?atomic_read(&nxi->nx_tasks):0, ++ task, _file, _line); ++ ++ atomic_inc(&nxi->nx_tasks); ++} ++ ++ ++extern void unhash_nx_info(struct nx_info *); ++ ++#define release_nx_info(v, p) __release_nx_info(v, p, __FILE__, __LINE__) ++ ++static inline void __release_nx_info(struct nx_info *nxi, ++ struct task_struct *task, const char *_file, int _line) ++{ ++ vxlprintk(VXD_CBIT(nid, 3), "release_nx_info(%p[#%d.%d.%d]) %p", ++ nxi, nxi ? nxi->nx_id : 0, ++ nxi ? atomic_read(&nxi->nx_usecnt) : 0, ++ nxi ? atomic_read(&nxi->nx_tasks) : 0, ++ task, _file, _line); ++ ++ might_sleep(); ++ ++ if (atomic_dec_and_test(&nxi->nx_tasks)) ++ unhash_nx_info(nxi); ++} ++ ++ ++#define task_get_nx_info(i) __task_get_nx_info(i, __FILE__, __LINE__) ++ ++static __inline__ struct nx_info *__task_get_nx_info(struct task_struct *p, ++ const char *_file, int _line) ++{ ++ struct nx_info *nxi; ++ ++ task_lock(p); ++ vxlprintk(VXD_CBIT(nid, 5), "task_get_nx_info(%p)", ++ p, _file, _line); ++ nxi = __get_nx_info(p->nx_info, _file, _line); ++ task_unlock(p); ++ return nxi; ++} ++ ++ ++static inline void exit_nx_info(struct task_struct *p) ++{ ++ if (p->nx_info) ++ release_nx_info(p->nx_info, p); ++} ++ ++ ++#else ++#warning duplicate inclusion ++#endif +diff -NurpP --minimal linux-3.0.9/include/linux/vs_pid.h linux-3.0.9-vs2.3.2.1/include/linux/vs_pid.h +--- linux-3.0.9/include/linux/vs_pid.h 1970-01-01 01:00:00.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/include/linux/vs_pid.h 2011-06-10 22:11:24.000000000 +0200 +@@ -0,0 +1,50 @@ ++#ifndef _VS_PID_H ++#define _VS_PID_H ++ ++#include "vserver/base.h" ++#include "vserver/check.h" ++#include "vserver/context.h" ++#include "vserver/debug.h" ++#include "vserver/pid.h" ++#include <linux/pid_namespace.h> ++ ++ ++#define VXF_FAKE_INIT (VXF_INFO_INIT | VXF_STATE_INIT) ++ ++static inline ++int vx_proc_task_visible(struct task_struct *task) ++{ ++ if ((task->pid == 1) && ++ !vx_flags(VXF_FAKE_INIT, VXF_FAKE_INIT)) ++ /* show a blend through init */ ++ goto visible; ++ if (vx_check(vx_task_xid(task), VS_WATCH | VS_IDENT)) ++ goto visible; ++ return 0; ++visible: ++ return 1; ++} ++ ++#define find_task_by_real_pid(pid) find_task_by_pid_ns(pid, &init_pid_ns) ++ ++ ++static inline ++struct task_struct *vx_get_proc_task(struct inode *inode, struct pid *pid) ++{ ++ struct task_struct *task = get_pid_task(pid, PIDTYPE_PID); ++ ++ if (task && !vx_proc_task_visible(task)) { ++ vxdprintk(VXD_CBIT(misc, 6), ++ "dropping task (get) %p[#%u,%u] for %p[#%u,%u]", ++ task, task->xid, task->pid, ++ current, current->xid, current->pid); ++ put_task_struct(task); ++ task = NULL; ++ } ++ return task; ++} ++ ++ ++#else ++#warning duplicate inclusion ++#endif +diff -NurpP --minimal linux-3.0.9/include/linux/vs_sched.h linux-3.0.9-vs2.3.2.1/include/linux/vs_sched.h +--- linux-3.0.9/include/linux/vs_sched.h 1970-01-01 01:00:00.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/include/linux/vs_sched.h 2011-06-10 22:11:24.000000000 +0200 +@@ -0,0 +1,40 @@ ++#ifndef _VS_SCHED_H ++#define _VS_SCHED_H ++ ++#include "vserver/base.h" ++#include "vserver/context.h" ++#include "vserver/sched.h" ++ ++ ++#define MAX_PRIO_BIAS 20 ++#define MIN_PRIO_BIAS -20 ++ ++static inline ++int vx_adjust_prio(struct task_struct *p, int prio, int max_user) ++{ ++ struct vx_info *vxi = p->vx_info; ++ ++ if (vxi) ++ prio += vx_cpu(vxi, sched_pc).prio_bias; ++ return prio; ++} ++ ++static inline void vx_account_user(struct vx_info *vxi, ++ cputime_t cputime, int nice) ++{ ++ if (!vxi) ++ return; ++ vx_cpu(vxi, sched_pc).user_ticks += cputime; ++} ++ ++static inline void vx_account_system(struct vx_info *vxi, ++ cputime_t cputime, int idle) ++{ ++ if (!vxi) ++ return; ++ vx_cpu(vxi, sched_pc).sys_ticks += cputime; ++} ++ ++#else ++#warning duplicate inclusion ++#endif +diff -NurpP --minimal linux-3.0.9/include/linux/vs_socket.h linux-3.0.9-vs2.3.2.1/include/linux/vs_socket.h +--- linux-3.0.9/include/linux/vs_socket.h 1970-01-01 01:00:00.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/include/linux/vs_socket.h 2011-06-10 22:11:24.000000000 +0200 +@@ -0,0 +1,67 @@ ++#ifndef _VS_SOCKET_H ++#define _VS_SOCKET_H ++ ++#include "vserver/debug.h" ++#include "vserver/base.h" ++#include "vserver/cacct.h" ++#include "vserver/context.h" ++#include "vserver/tag.h" ++ ++ ++/* socket accounting */ ++ ++#include <linux/socket.h> ++ ++static inline int vx_sock_type(int family) ++{ ++ switch (family) { ++ case PF_UNSPEC: ++ return VXA_SOCK_UNSPEC; ++ case PF_UNIX: ++ return VXA_SOCK_UNIX; ++ case PF_INET: ++ return VXA_SOCK_INET; ++ case PF_INET6: ++ return VXA_SOCK_INET6; ++ case PF_PACKET: ++ return VXA_SOCK_PACKET; ++ default: ++ return VXA_SOCK_OTHER; ++ } ++} ++ ++#define vx_acc_sock(v, f, p, s) \ ++ __vx_acc_sock(v, f, p, s, __FILE__, __LINE__) ++ ++static inline void __vx_acc_sock(struct vx_info *vxi, ++ int family, int pos, int size, char *file, int line) ++{ ++ if (vxi) { ++ int type = vx_sock_type(family); ++ ++ atomic_long_inc(&vxi->cacct.sock[type][pos].count); ++ atomic_long_add(size, &vxi->cacct.sock[type][pos].total); ++ } ++} ++ ++#define vx_sock_recv(sk, s) \ ++ vx_acc_sock((sk)->sk_vx_info, (sk)->sk_family, 0, s) ++#define vx_sock_send(sk, s) \ ++ vx_acc_sock((sk)->sk_vx_info, (sk)->sk_family, 1, s) ++#define vx_sock_fail(sk, s) \ ++ vx_acc_sock((sk)->sk_vx_info, (sk)->sk_family, 2, s) ++ ++ ++#define sock_vx_init(s) do { \ ++ (s)->sk_xid = 0; \ ++ (s)->sk_vx_info = NULL; \ ++ } while (0) ++ ++#define sock_nx_init(s) do { \ ++ (s)->sk_nid = 0; \ ++ (s)->sk_nx_info = NULL; \ ++ } while (0) ++ ++#else ++#warning duplicate inclusion ++#endif +diff -NurpP --minimal linux-3.0.9/include/linux/vs_tag.h linux-3.0.9-vs2.3.2.1/include/linux/vs_tag.h +--- linux-3.0.9/include/linux/vs_tag.h 1970-01-01 01:00:00.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/include/linux/vs_tag.h 2011-06-10 22:11:24.000000000 +0200 +@@ -0,0 +1,47 @@ ++#ifndef _VS_TAG_H ++#define _VS_TAG_H ++ ++#include <linux/vserver/tag.h> ++ ++/* check conditions */ ++ ++#define DX_ADMIN 0x0001 ++#define DX_WATCH 0x0002 ++#define DX_HOSTID 0x0008 ++ ++#define DX_IDENT 0x0010 ++ ++#define DX_ARG_MASK 0x0010 ++ ++ ++#define dx_task_tag(t) ((t)->tag) ++ ++#define dx_current_tag() dx_task_tag(current) ++ ++#define dx_check(c, m) __dx_check(dx_current_tag(), c, m) ++ ++#define dx_weak_check(c, m) ((m) ? dx_check(c, m) : 1) ++ ++ ++/* ++ * check current context for ADMIN/WATCH and ++ * optionally against supplied argument ++ */ ++static inline int __dx_check(tag_t cid, tag_t id, unsigned int mode) ++{ ++ if (mode & DX_ARG_MASK) { ++ if ((mode & DX_IDENT) && (id == cid)) ++ return 1; ++ } ++ return (((mode & DX_ADMIN) && (cid == 0)) || ++ ((mode & DX_WATCH) && (cid == 1)) || ++ ((mode & DX_HOSTID) && (id == 0))); ++} ++ ++struct inode; ++int dx_permission(const struct inode *inode, int mask); ++ ++ ++#else ++#warning duplicate inclusion ++#endif +diff -NurpP --minimal linux-3.0.9/include/linux/vs_time.h linux-3.0.9-vs2.3.2.1/include/linux/vs_time.h +--- linux-3.0.9/include/linux/vs_time.h 1970-01-01 01:00:00.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/include/linux/vs_time.h 2011-06-13 14:57:45.000000000 +0200 +@@ -0,0 +1,19 @@ ++#ifndef _VS_TIME_H ++#define _VS_TIME_H ++ ++ ++/* time faking stuff */ ++ ++#ifdef CONFIG_VSERVER_VTIME ++ ++extern void vx_adjust_timespec(struct timespec *ts); ++extern int vx_settimeofday(const struct timespec *ts); ++ ++#else ++#define vx_adjust_timespec(t) do { } while (0) ++#define vx_settimeofday(t) do_settimeofday(t) ++#endif ++ ++#else ++#warning duplicate inclusion ++#endif +diff -NurpP --minimal linux-3.0.9/include/linux/vserver/Kbuild linux-3.0.9-vs2.3.2.1/include/linux/vserver/Kbuild +--- linux-3.0.9/include/linux/vserver/Kbuild 1970-01-01 01:00:00.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/include/linux/vserver/Kbuild 2011-08-08 18:02:46.000000000 +0200 +@@ -0,0 +1,8 @@ ++ ++header-y += context_cmd.h network_cmd.h space_cmd.h \ ++ cacct_cmd.h cvirt_cmd.h limit_cmd.h dlimit_cmd.h \ ++ inode_cmd.h tag_cmd.h sched_cmd.h signal_cmd.h \ ++ debug_cmd.h device_cmd.h ++ ++header-y += switch.h network.h monitor.h inode.h device.h ++ +diff -NurpP --minimal linux-3.0.9/include/linux/vserver/base.h linux-3.0.9-vs2.3.2.1/include/linux/vserver/base.h +--- linux-3.0.9/include/linux/vserver/base.h 1970-01-01 01:00:00.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/include/linux/vserver/base.h 2011-08-01 18:26:07.000000000 +0200 +@@ -0,0 +1,178 @@ ++#ifndef _VX_BASE_H ++#define _VX_BASE_H ++ ++ ++/* context state changes */ ++ ++enum { ++ VSC_STARTUP = 1, ++ VSC_SHUTDOWN, ++ ++ VSC_NETUP, ++ VSC_NETDOWN, ++}; ++ ++ ++ ++#define vx_task_xid(t) ((t)->xid) ++ ++#define vx_current_xid() vx_task_xid(current) ++ ++#define current_vx_info() (current->vx_info) ++ ++ ++#define nx_task_nid(t) ((t)->nid) ++ ++#define nx_current_nid() nx_task_nid(current) ++ ++#define current_nx_info() (current->nx_info) ++ ++ ++/* generic flag merging */ ++ ++#define vs_check_flags(v, m, f) (((v) & (m)) ^ (f)) ++ ++#define vs_mask_flags(v, f, m) (((v) & ~(m)) | ((f) & (m))) ++ ++#define vs_mask_mask(v, f, m) (((v) & ~(m)) | ((v) & (f) & (m))) ++ ++#define vs_check_bit(v, n) ((v) & (1LL << (n))) ++ ++ ++/* context flags */ ++ ++#define __vx_flags(v) ((v) ? (v)->vx_flags : 0) ++ ++#define vx_current_flags() __vx_flags(current_vx_info()) ++ ++#define vx_info_flags(v, m, f) \ ++ vs_check_flags(__vx_flags(v), m, f) ++ ++#define task_vx_flags(t, m, f) \ ++ ((t) && vx_info_flags((t)->vx_info, m, f)) ++ ++#define vx_flags(m, f) vx_info_flags(current_vx_info(), m, f) ++ ++ ++/* context caps */ ++ ++#define __vx_ccaps(v) ((v) ? (v)->vx_ccaps : 0) ++ ++#define vx_current_ccaps() __vx_ccaps(current_vx_info()) ++ ++#define vx_info_ccaps(v, c) (__vx_ccaps(v) & (c)) ++ ++#define vx_ccaps(c) vx_info_ccaps(current_vx_info(), (c)) ++ ++ ++ ++/* network flags */ ++ ++#define __nx_flags(n) ((n) ? (n)->nx_flags : 0) ++ ++#define nx_current_flags() __nx_flags(current_nx_info()) ++ ++#define nx_info_flags(n, m, f) \ ++ vs_check_flags(__nx_flags(n), m, f) ++ ++#define task_nx_flags(t, m, f) \ ++ ((t) && nx_info_flags((t)->nx_info, m, f)) ++ ++#define nx_flags(m, f) nx_info_flags(current_nx_info(), m, f) ++ ++ ++/* network caps */ ++ ++#define __nx_ncaps(n) ((n) ? (n)->nx_ncaps : 0) ++ ++#define nx_current_ncaps() __nx_ncaps(current_nx_info()) ++ ++#define nx_info_ncaps(n, c) (__nx_ncaps(n) & (c)) ++ ++#define nx_ncaps(c) nx_info_ncaps(current_nx_info(), c) ++ ++ ++/* context mask capabilities */ ++ ++#define __vx_mcaps(v) ((v) ? (v)->vx_ccaps >> 32UL : ~0 ) ++ ++#define vx_info_mcaps(v, c) (__vx_mcaps(v) & (c)) ++ ++#define vx_mcaps(c) vx_info_mcaps(current_vx_info(), c) ++ ++ ++/* context bcap mask */ ++ ++#define __vx_bcaps(v) ((v)->vx_bcaps) ++ ++#define vx_current_bcaps() __vx_bcaps(current_vx_info()) ++ ++ ++/* mask given bcaps */ ++ ++#define vx_info_mbcaps(v, c) ((v) ? cap_intersect(__vx_bcaps(v), c) : c) ++ ++#define vx_mbcaps(c) vx_info_mbcaps(current_vx_info(), c) ++ ++ ++/* masked cap_bset */ ++ ++#define vx_info_cap_bset(v) vx_info_mbcaps(v, current->cap_bset) ++ ++#define vx_current_cap_bset() vx_info_cap_bset(current_vx_info()) ++ ++#if 0 ++#define vx_info_mbcap(v, b) \ ++ (!vx_info_flags(v, VXF_STATE_SETUP, 0) ? \ ++ vx_info_bcaps(v, b) : (b)) ++ ++#define task_vx_mbcap(t, b) \ ++ vx_info_mbcap((t)->vx_info, (t)->b) ++ ++#define vx_mbcap(b) task_vx_mbcap(current, b) ++#endif ++ ++#define vx_cap_raised(v, c, f) cap_raised(vx_info_mbcaps(v, c), f) ++ ++#define vx_capable(b, c) (capable(b) || \ ++ (cap_raised(current_cap(), b) && vx_ccaps(c))) ++ ++#define vx_ns_capable(n, b, c) (ns_capable(n, b) || \ ++ (cap_raised(current_cap(), b) && vx_ccaps(c))) ++ ++#define nx_capable(b, c) (capable(b) || \ ++ (cap_raised(current_cap(), b) && nx_ncaps(c))) ++ ++#define vx_task_initpid(t, n) \ ++ ((t)->vx_info && \ ++ ((t)->vx_info->vx_initpid == (n))) ++ ++#define vx_current_initpid(n) vx_task_initpid(current, n) ++ ++ ++/* context unshare mask */ ++ ++#define __vx_umask(v) ((v)->vx_umask) ++ ++#define vx_current_umask() __vx_umask(current_vx_info()) ++ ++#define vx_can_unshare(b, f) (capable(b) || \ ++ (cap_raised(current_cap(), b) && \ ++ !((f) & ~vx_current_umask()))) ++ ++ ++#define __vx_wmask(v) ((v)->vx_wmask) ++ ++#define vx_current_wmask() __vx_wmask(current_vx_info()) ++ ++ ++#define __vx_state(v) ((v) ? ((v)->vx_state) : 0) ++ ++#define vx_info_state(v, m) (__vx_state(v) & (m)) ++ ++ ++#define __nx_state(n) ((n) ? ((n)->nx_state) : 0) ++ ++#define nx_info_state(n, m) (__nx_state(n) & (m)) ++ ++#endif +diff -NurpP --minimal linux-3.0.9/include/linux/vserver/cacct.h linux-3.0.9-vs2.3.2.1/include/linux/vserver/cacct.h +--- linux-3.0.9/include/linux/vserver/cacct.h 1970-01-01 01:00:00.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/include/linux/vserver/cacct.h 2011-06-10 22:11:24.000000000 +0200 +@@ -0,0 +1,15 @@ ++#ifndef _VX_CACCT_H ++#define _VX_CACCT_H ++ ++ ++enum sock_acc_field { ++ VXA_SOCK_UNSPEC = 0, ++ VXA_SOCK_UNIX, ++ VXA_SOCK_INET, ++ VXA_SOCK_INET6, ++ VXA_SOCK_PACKET, ++ VXA_SOCK_OTHER, ++ VXA_SOCK_SIZE /* array size */ ++}; ++ ++#endif /* _VX_CACCT_H */ +diff -NurpP --minimal linux-3.0.9/include/linux/vserver/cacct_cmd.h linux-3.0.9-vs2.3.2.1/include/linux/vserver/cacct_cmd.h +--- linux-3.0.9/include/linux/vserver/cacct_cmd.h 1970-01-01 01:00:00.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/include/linux/vserver/cacct_cmd.h 2011-06-10 22:11:24.000000000 +0200 +@@ -0,0 +1,23 @@ ++#ifndef _VX_CACCT_CMD_H ++#define _VX_CACCT_CMD_H ++ ++ ++/* virtual host info name commands */ ++ ++#define VCMD_sock_stat VC_CMD(VSTAT, 5, 0) ++ ++struct vcmd_sock_stat_v0 { ++ uint32_t field; ++ uint32_t count[3]; ++ uint64_t total[3]; ++}; ++ ++ ++#ifdef __KERNEL__ ++ ++#include <linux/compiler.h> ++ ++extern int vc_sock_stat(struct vx_info *, void __user *); ++ ++#endif /* __KERNEL__ */ ++#endif /* _VX_CACCT_CMD_H */ +diff -NurpP --minimal linux-3.0.9/include/linux/vserver/cacct_def.h linux-3.0.9-vs2.3.2.1/include/linux/vserver/cacct_def.h +--- linux-3.0.9/include/linux/vserver/cacct_def.h 1970-01-01 01:00:00.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/include/linux/vserver/cacct_def.h 2011-06-10 22:11:24.000000000 +0200 +@@ -0,0 +1,43 @@ ++#ifndef _VX_CACCT_DEF_H ++#define _VX_CACCT_DEF_H ++ ++#include <asm/atomic.h> ++#include <linux/vserver/cacct.h> ++ ++ ++struct _vx_sock_acc { ++ atomic_long_t count; ++ atomic_long_t total; ++}; ++ ++/* context sub struct */ ++ ++struct _vx_cacct { ++ struct _vx_sock_acc sock[VXA_SOCK_SIZE][3]; ++ atomic_t slab[8]; ++ atomic_t page[6][8]; ++}; ++ ++#ifdef CONFIG_VSERVER_DEBUG ++ ++static inline void __dump_vx_cacct(struct _vx_cacct *cacct) ++{ ++ int i, j; ++ ++ printk("\t_vx_cacct:"); ++ for (i = 0; i < 6; i++) { ++ struct _vx_sock_acc *ptr = cacct->sock[i]; ++ ++ printk("\t [%d] =", i); ++ for (j = 0; j < 3; j++) { ++ printk(" [%d] = %8lu, %8lu", j, ++ atomic_long_read(&ptr[j].count), ++ atomic_long_read(&ptr[j].total)); ++ } ++ printk("\n"); ++ } ++} ++ ++#endif ++ ++#endif /* _VX_CACCT_DEF_H */ +diff -NurpP --minimal linux-3.0.9/include/linux/vserver/cacct_int.h linux-3.0.9-vs2.3.2.1/include/linux/vserver/cacct_int.h +--- linux-3.0.9/include/linux/vserver/cacct_int.h 1970-01-01 01:00:00.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/include/linux/vserver/cacct_int.h 2011-06-10 22:11:24.000000000 +0200 +@@ -0,0 +1,21 @@ ++#ifndef _VX_CACCT_INT_H ++#define _VX_CACCT_INT_H ++ ++ ++#ifdef __KERNEL__ ++ ++static inline ++unsigned long vx_sock_count(struct _vx_cacct *cacct, int type, int pos) ++{ ++ return atomic_long_read(&cacct->sock[type][pos].count); ++} ++ ++ ++static inline ++unsigned long vx_sock_total(struct _vx_cacct *cacct, int type, int pos) ++{ ++ return atomic_long_read(&cacct->sock[type][pos].total); ++} ++ ++#endif /* __KERNEL__ */ ++#endif /* _VX_CACCT_INT_H */ +diff -NurpP --minimal linux-3.0.9/include/linux/vserver/check.h linux-3.0.9-vs2.3.2.1/include/linux/vserver/check.h +--- linux-3.0.9/include/linux/vserver/check.h 1970-01-01 01:00:00.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/include/linux/vserver/check.h 2011-06-10 22:11:24.000000000 +0200 +@@ -0,0 +1,89 @@ ++#ifndef _VS_CHECK_H ++#define _VS_CHECK_H ++ ++ ++#define MAX_S_CONTEXT 65535 /* Arbitrary limit */ ++ ++#ifdef CONFIG_VSERVER_DYNAMIC_IDS ++#define MIN_D_CONTEXT 49152 /* dynamic contexts start here */ ++#else ++#define MIN_D_CONTEXT 65536 ++#endif ++ ++/* check conditions */ ++ ++#define VS_ADMIN 0x0001 ++#define VS_WATCH 0x0002 ++#define VS_HIDE 0x0004 ++#define VS_HOSTID 0x0008 ++ ++#define VS_IDENT 0x0010 ++#define VS_EQUIV 0x0020 ++#define VS_PARENT 0x0040 ++#define VS_CHILD 0x0080 ++ ++#define VS_ARG_MASK 0x00F0 ++ ++#define VS_DYNAMIC 0x0100 ++#define VS_STATIC 0x0200 ++ ++#define VS_ATR_MASK 0x0F00 ++ ++#ifdef CONFIG_VSERVER_PRIVACY ++#define VS_ADMIN_P (0) ++#define VS_WATCH_P (0) ++#else ++#define VS_ADMIN_P VS_ADMIN ++#define VS_WATCH_P VS_WATCH ++#endif ++ ++#define VS_HARDIRQ 0x1000 ++#define VS_SOFTIRQ 0x2000 ++#define VS_IRQ 0x4000 ++ ++#define VS_IRQ_MASK 0xF000 ++ ++#include <linux/hardirq.h> ++ ++/* ++ * check current context for ADMIN/WATCH and ++ * optionally against supplied argument ++ */ ++static inline int __vs_check(int cid, int id, unsigned int mode) ++{ ++ if (mode & VS_ARG_MASK) { ++ if ((mode & VS_IDENT) && (id == cid)) ++ return 1; ++ } ++ if (mode & VS_ATR_MASK) { ++ if ((mode & VS_DYNAMIC) && ++ (id >= MIN_D_CONTEXT) && ++ (id <= MAX_S_CONTEXT)) ++ return 1; ++ if ((mode & VS_STATIC) && ++ (id > 1) && (id < MIN_D_CONTEXT)) ++ return 1; ++ } ++ if (mode & VS_IRQ_MASK) { ++ if ((mode & VS_IRQ) && unlikely(in_interrupt())) ++ return 1; ++ if ((mode & VS_HARDIRQ) && unlikely(in_irq())) ++ return 1; ++ if ((mode & VS_SOFTIRQ) && unlikely(in_softirq())) ++ return 1; ++ } ++ return (((mode & VS_ADMIN) && (cid == 0)) || ++ ((mode & VS_WATCH) && (cid == 1)) || ++ ((mode & VS_HOSTID) && (id == 0))); ++} ++ ++#define vx_check(c, m) __vs_check(vx_current_xid(), c, (m) | VS_IRQ) ++ ++#define vx_weak_check(c, m) ((m) ? vx_check(c, m) : 1) ++ ++ ++#define nx_check(c, m) __vs_check(nx_current_nid(), c, m) ++ ++#define nx_weak_check(c, m) ((m) ? nx_check(c, m) : 1) ++ ++#endif +diff -NurpP --minimal linux-3.0.9/include/linux/vserver/context.h linux-3.0.9-vs2.3.2.1/include/linux/vserver/context.h +--- linux-3.0.9/include/linux/vserver/context.h 1970-01-01 01:00:00.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/include/linux/vserver/context.h 2011-10-27 13:59:29.000000000 +0200 +@@ -0,0 +1,188 @@ ++#ifndef _VX_CONTEXT_H ++#define _VX_CONTEXT_H ++ ++#include <linux/types.h> ++#include <linux/capability.h> ++ ++ ++/* context flags */ ++ ++#define VXF_INFO_SCHED 0x00000002 ++#define VXF_INFO_NPROC 0x00000004 ++#define VXF_INFO_PRIVATE 0x00000008 ++ ++#define VXF_INFO_INIT 0x00000010 ++#define VXF_INFO_HIDE 0x00000020 ++#define VXF_INFO_ULIMIT 0x00000040 ++#define VXF_INFO_NSPACE 0x00000080 ++ ++#define VXF_SCHED_HARD 0x00000100 ++#define VXF_SCHED_PRIO 0x00000200 ++#define VXF_SCHED_PAUSE 0x00000400 ++ ++#define VXF_VIRT_MEM 0x00010000 ++#define VXF_VIRT_UPTIME 0x00020000 ++#define VXF_VIRT_CPU 0x00040000 ++#define VXF_VIRT_LOAD 0x00080000 ++#define VXF_VIRT_TIME 0x00100000 ++ ++#define VXF_HIDE_MOUNT 0x01000000 ++/* was VXF_HIDE_NETIF 0x02000000 */ ++#define VXF_HIDE_VINFO 0x04000000 ++ ++#define VXF_STATE_SETUP (1ULL << 32) ++#define VXF_STATE_INIT (1ULL << 33) ++#define VXF_STATE_ADMIN (1ULL << 34) ++ ++#define VXF_SC_HELPER (1ULL << 36) ++#define VXF_REBOOT_KILL (1ULL << 37) ++#define VXF_PERSISTENT (1ULL << 38) ++ ++#define VXF_FORK_RSS (1ULL << 48) ++#define VXF_PROLIFIC (1ULL << 49) ++ ++#define VXF_IGNEG_NICE (1ULL << 52) ++ ++#define VXF_ONE_TIME (0x0007ULL << 32) ++ ++#define VXF_INIT_SET (VXF_STATE_SETUP | VXF_STATE_INIT | VXF_STATE_ADMIN) ++ ++ ++/* context migration */ ++ ++#define VXM_SET_INIT 0x00000001 ++#define VXM_SET_REAPER 0x00000002 ++ ++/* context caps */ ++ ++#define VXC_SET_UTSNAME 0x00000001 ++#define VXC_SET_RLIMIT 0x00000002 ++#define VXC_FS_SECURITY 0x00000004 ++#define VXC_FS_TRUSTED 0x00000008 ++#define VXC_TIOCSTI 0x00000010 ++ ++/* was VXC_RAW_ICMP 0x00000100 */ ++#define VXC_SYSLOG 0x00001000 ++#define VXC_OOM_ADJUST 0x00002000 ++#define VXC_AUDIT_CONTROL 0x00004000 ++ ++#define VXC_SECURE_MOUNT 0x00010000 ++#define VXC_SECURE_REMOUNT 0x00020000 ++#define VXC_BINARY_MOUNT 0x00040000 ++ ++#define VXC_QUOTA_CTL 0x00100000 ++#define VXC_ADMIN_MAPPER 0x00200000 ++#define VXC_ADMIN_CLOOP 0x00400000 ++ ++#define VXC_KTHREAD 0x01000000 ++#define VXC_NAMESPACE 0x02000000 ++ ++ ++#ifdef __KERNEL__ ++ ++#include <linux/list.h> ++#include <linux/spinlock.h> ++#include <linux/rcupdate.h> ++ ++#include "limit_def.h" ++#include "sched_def.h" ++#include "cvirt_def.h" ++#include "cacct_def.h" ++#include "device_def.h" ++ ++#define VX_SPACES 2 ++ ++struct _vx_info_pc { ++ struct _vx_sched_pc sched_pc; ++ struct _vx_cvirt_pc cvirt_pc; ++}; ++ ++struct _vx_space { ++ unsigned long vx_nsmask; /* assignment mask */ ++ struct nsproxy *vx_nsproxy; /* private namespaces */ ++ struct fs_struct *vx_fs; /* private namespace fs */ ++ const struct cred *vx_cred; /* task credentials */ ++}; ++ ++struct vx_info { ++ struct hlist_node vx_hlist; /* linked list of contexts */ ++ xid_t vx_id; /* context id */ ++ atomic_t vx_usecnt; /* usage count */ ++ atomic_t vx_tasks; /* tasks count */ ++ struct vx_info *vx_parent; /* parent context */ ++ int vx_state; /* context state */ ++ ++ struct _vx_space space[VX_SPACES]; /* namespace store */ ++ ++ uint64_t vx_flags; /* context flags */ ++ uint64_t vx_ccaps; /* context caps (vserver) */ ++ uint64_t vx_umask; /* unshare mask (guest) */ ++ uint64_t vx_wmask; /* warn mask (guest) */ ++ kernel_cap_t vx_bcaps; /* bounding caps (system) */ ++ ++ struct task_struct *vx_reaper; /* guest reaper process */ ++ pid_t vx_initpid; /* PID of guest init */ ++ int64_t vx_badness_bias; /* OOM points bias */ ++ ++ struct _vx_limit limit; /* vserver limits */ ++ struct _vx_sched sched; /* vserver scheduler */ ++ struct _vx_cvirt cvirt; /* virtual/bias stuff */ ++ struct _vx_cacct cacct; /* context accounting */ ++ ++ struct _vx_device dmap; /* default device map targets */ ++ ++#ifndef CONFIG_SMP ++ struct _vx_info_pc info_pc; /* per cpu data */ ++#else ++ struct _vx_info_pc *ptr_pc; /* per cpu array */ ++#endif ++ ++ wait_queue_head_t vx_wait; /* context exit waitqueue */ ++ int reboot_cmd; /* last sys_reboot() cmd */ ++ int exit_code; /* last process exit code */ ++ ++ char vx_name[65]; /* vserver name */ ++}; ++ ++#ifndef CONFIG_SMP ++#define vx_ptr_pc(vxi) (&(vxi)->info_pc) ++#define vx_per_cpu(vxi, v, id) vx_ptr_pc(vxi)->v ++#else ++#define vx_ptr_pc(vxi) ((vxi)->ptr_pc) ++#define vx_per_cpu(vxi, v, id) per_cpu_ptr(vx_ptr_pc(vxi), id)->v ++#endif ++ ++#define vx_cpu(vxi, v) vx_per_cpu(vxi, v, smp_processor_id()) ++ ++ ++struct vx_info_save { ++ struct vx_info *vxi; ++ xid_t xid; ++}; ++ ++ ++/* status flags */ ++ ++#define VXS_HASHED 0x0001 ++#define VXS_PAUSED 0x0010 ++#define VXS_SHUTDOWN 0x0100 ++#define VXS_HELPER 0x1000 ++#define VXS_RELEASED 0x8000 ++ ++ ++extern void claim_vx_info(struct vx_info *, struct task_struct *); ++extern void release_vx_info(struct vx_info *, struct task_struct *); ++ ++extern struct vx_info *lookup_vx_info(int); ++extern struct vx_info *lookup_or_create_vx_info(int); ++ ++extern int get_xid_list(int, unsigned int *, int); ++extern int xid_is_hashed(xid_t); ++ ++extern int vx_migrate_task(struct task_struct *, struct vx_info *, int); ++ ++extern long vs_state_change(struct vx_info *, unsigned int); ++ ++ ++#endif /* __KERNEL__ */ ++#endif /* _VX_CONTEXT_H */ +diff -NurpP --minimal linux-3.0.9/include/linux/vserver/context_cmd.h linux-3.0.9-vs2.3.2.1/include/linux/vserver/context_cmd.h +--- linux-3.0.9/include/linux/vserver/context_cmd.h 1970-01-01 01:00:00.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/include/linux/vserver/context_cmd.h 2011-08-01 18:25:07.000000000 +0200 +@@ -0,0 +1,162 @@ ++#ifndef _VX_CONTEXT_CMD_H ++#define _VX_CONTEXT_CMD_H ++ ++ ++/* vinfo commands */ ++ ++#define VCMD_task_xid VC_CMD(VINFO, 1, 0) ++ ++#ifdef __KERNEL__ ++extern int vc_task_xid(uint32_t); ++ ++#endif /* __KERNEL__ */ ++ ++#define VCMD_vx_info VC_CMD(VINFO, 5, 0) ++ ++struct vcmd_vx_info_v0 { ++ uint32_t xid; ++ uint32_t initpid; ++ /* more to come */ ++}; ++ ++#ifdef __KERNEL__ ++extern int vc_vx_info(struct vx_info *, void __user *); ++ ++#endif /* __KERNEL__ */ ++ ++#define VCMD_ctx_stat VC_CMD(VSTAT, 0, 0) ++ ++struct vcmd_ctx_stat_v0 { ++ uint32_t usecnt; ++ uint32_t tasks; ++ /* more to come */ ++}; ++ ++#ifdef __KERNEL__ ++extern int vc_ctx_stat(struct vx_info *, void __user *); ++ ++#endif /* __KERNEL__ */ ++ ++/* context commands */ ++ ++#define VCMD_ctx_create_v0 VC_CMD(VPROC, 1, 0) ++#define VCMD_ctx_create VC_CMD(VPROC, 1, 1) ++ ++struct vcmd_ctx_create { ++ uint64_t flagword; ++}; ++ ++#define VCMD_ctx_migrate_v0 VC_CMD(PROCMIG, 1, 0) ++#define VCMD_ctx_migrate VC_CMD(PROCMIG, 1, 1) ++ ++struct vcmd_ctx_migrate { ++ uint64_t flagword; ++}; ++ ++#ifdef __KERNEL__ ++extern int vc_ctx_create(uint32_t, void __user *); ++extern int vc_ctx_migrate(struct vx_info *, void __user *); ++ ++#endif /* __KERNEL__ */ ++ ++ ++/* flag commands */ ++ ++#define VCMD_get_cflags VC_CMD(FLAGS, 1, 0) ++#define VCMD_set_cflags VC_CMD(FLAGS, 2, 0) ++ ++struct vcmd_ctx_flags_v0 { ++ uint64_t flagword; ++ uint64_t mask; ++}; ++ ++#ifdef __KERNEL__ ++extern int vc_get_cflags(struct vx_info *, void __user *); ++extern int vc_set_cflags(struct vx_info *, void __user *); ++ ++#endif /* __KERNEL__ */ ++ ++ ++/* context caps commands */ ++ ++#define VCMD_get_ccaps VC_CMD(FLAGS, 3, 1) ++#define VCMD_set_ccaps VC_CMD(FLAGS, 4, 1) ++ ++struct vcmd_ctx_caps_v1 { ++ uint64_t ccaps; ++ uint64_t cmask; ++}; ++ ++#ifdef __KERNEL__ ++extern int vc_get_ccaps(struct vx_info *, void __user *); ++extern int vc_set_ccaps(struct vx_info *, void __user *); ++ ++#endif /* __KERNEL__ */ ++ ++ ++/* bcaps commands */ ++ ++#define VCMD_get_bcaps VC_CMD(FLAGS, 9, 0) ++#define VCMD_set_bcaps VC_CMD(FLAGS, 10, 0) ++ ++struct vcmd_bcaps { ++ uint64_t bcaps; ++ uint64_t bmask; ++}; ++ ++#ifdef __KERNEL__ ++extern int vc_get_bcaps(struct vx_info *, void __user *); ++extern int vc_set_bcaps(struct vx_info *, void __user *); ++ ++#endif /* __KERNEL__ */ ++ ++ ++/* umask commands */ ++ ++#define VCMD_get_umask VC_CMD(FLAGS, 13, 0) ++#define VCMD_set_umask VC_CMD(FLAGS, 14, 0) ++ ++struct vcmd_umask { ++ uint64_t umask; ++ uint64_t mask; ++}; ++ ++#ifdef __KERNEL__ ++extern int vc_get_umask(struct vx_info *, void __user *); ++extern int vc_set_umask(struct vx_info *, void __user *); ++ ++#endif /* __KERNEL__ */ ++ ++ ++/* wmask commands */ ++ ++#define VCMD_get_wmask VC_CMD(FLAGS, 15, 0) ++#define VCMD_set_wmask VC_CMD(FLAGS, 16, 0) ++ ++struct vcmd_wmask { ++ uint64_t wmask; ++ uint64_t mask; ++}; ++ ++#ifdef __KERNEL__ ++extern int vc_get_wmask(struct vx_info *, void __user *); ++extern int vc_set_wmask(struct vx_info *, void __user *); ++ ++#endif /* __KERNEL__ */ ++ ++ ++/* OOM badness */ ++ ++#define VCMD_get_badness VC_CMD(MEMCTRL, 5, 0) ++#define VCMD_set_badness VC_CMD(MEMCTRL, 6, 0) ++ ++struct vcmd_badness_v0 { ++ int64_t bias; ++}; ++ ++#ifdef __KERNEL__ ++extern int vc_get_badness(struct vx_info *, void __user *); ++extern int vc_set_badness(struct vx_info *, void __user *); ++ ++#endif /* __KERNEL__ */ ++#endif /* _VX_CONTEXT_CMD_H */ +diff -NurpP --minimal linux-3.0.9/include/linux/vserver/cvirt.h linux-3.0.9-vs2.3.2.1/include/linux/vserver/cvirt.h +--- linux-3.0.9/include/linux/vserver/cvirt.h 1970-01-01 01:00:00.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/include/linux/vserver/cvirt.h 2011-06-10 22:11:24.000000000 +0200 +@@ -0,0 +1,20 @@ ++#ifndef _VX_CVIRT_H ++#define _VX_CVIRT_H ++ ++ ++#ifdef __KERNEL__ ++ ++struct timespec; ++ ++void vx_vsi_uptime(struct timespec *, struct timespec *); ++ ++ ++struct vx_info; ++ ++void vx_update_load(struct vx_info *); ++ ++ ++int vx_do_syslog(int, char __user *, int); ++ ++#endif /* __KERNEL__ */ ++#endif /* _VX_CVIRT_H */ +diff -NurpP --minimal linux-3.0.9/include/linux/vserver/cvirt_cmd.h linux-3.0.9-vs2.3.2.1/include/linux/vserver/cvirt_cmd.h +--- linux-3.0.9/include/linux/vserver/cvirt_cmd.h 1970-01-01 01:00:00.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/include/linux/vserver/cvirt_cmd.h 2011-06-10 22:11:24.000000000 +0200 +@@ -0,0 +1,53 @@ ++#ifndef _VX_CVIRT_CMD_H ++#define _VX_CVIRT_CMD_H ++ ++ ++/* virtual host info name commands */ ++ ++#define VCMD_set_vhi_name VC_CMD(VHOST, 1, 0) ++#define VCMD_get_vhi_name VC_CMD(VHOST, 2, 0) ++ ++struct vcmd_vhi_name_v0 { ++ uint32_t field; ++ char name[65]; ++}; ++ ++ ++enum vhi_name_field { ++ VHIN_CONTEXT = 0, ++ VHIN_SYSNAME, ++ VHIN_NODENAME, ++ VHIN_RELEASE, ++ VHIN_VERSION, ++ VHIN_MACHINE, ++ VHIN_DOMAINNAME, ++}; ++ ++ ++#ifdef __KERNEL__ ++ ++#include <linux/compiler.h> ++ ++extern int vc_set_vhi_name(struct vx_info *, void __user *); ++extern int vc_get_vhi_name(struct vx_info *, void __user *); ++ ++#endif /* __KERNEL__ */ ++ ++#define VCMD_virt_stat VC_CMD(VSTAT, 3, 0) ++ ++struct vcmd_virt_stat_v0 { ++ uint64_t offset; ++ uint64_t uptime; ++ uint32_t nr_threads; ++ uint32_t nr_running; ++ uint32_t nr_uninterruptible; ++ uint32_t nr_onhold; ++ uint32_t nr_forks; ++ uint32_t load[3]; ++}; ++ ++#ifdef __KERNEL__ ++extern int vc_virt_stat(struct vx_info *, void __user *); ++ ++#endif /* __KERNEL__ */ ++#endif /* _VX_CVIRT_CMD_H */ +diff -NurpP --minimal linux-3.0.9/include/linux/vserver/cvirt_def.h linux-3.0.9-vs2.3.2.1/include/linux/vserver/cvirt_def.h +--- linux-3.0.9/include/linux/vserver/cvirt_def.h 1970-01-01 01:00:00.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/include/linux/vserver/cvirt_def.h 2011-06-10 22:11:24.000000000 +0200 +@@ -0,0 +1,80 @@ ++#ifndef _VX_CVIRT_DEF_H ++#define _VX_CVIRT_DEF_H ++ ++#include <linux/jiffies.h> ++#include <linux/spinlock.h> ++#include <linux/wait.h> ++#include <linux/time.h> ++#include <asm/atomic.h> ++ ++ ++struct _vx_usage_stat { ++ uint64_t user; ++ uint64_t nice; ++ uint64_t system; ++ uint64_t softirq; ++ uint64_t irq; ++ uint64_t idle; ++ uint64_t iowait; ++}; ++ ++struct _vx_syslog { ++ wait_queue_head_t log_wait; ++ spinlock_t logbuf_lock; /* lock for the log buffer */ ++ ++ unsigned long log_start; /* next char to be read by syslog() */ ++ unsigned long con_start; /* next char to be sent to consoles */ ++ unsigned long log_end; /* most-recently-written-char + 1 */ ++ unsigned long logged_chars; /* #chars since last read+clear operation */ ++ ++ char log_buf[1024]; ++}; ++ ++ ++/* context sub struct */ ++ ++struct _vx_cvirt { ++ atomic_t nr_threads; /* number of current threads */ ++ atomic_t nr_running; /* number of running threads */ ++ atomic_t nr_uninterruptible; /* number of uninterruptible threads */ ++ ++ atomic_t nr_onhold; /* processes on hold */ ++ uint32_t onhold_last; /* jiffies when put on hold */ ++ ++ struct timespec bias_ts; /* time offset to the host */ ++ struct timespec bias_idle; ++ struct timespec bias_uptime; /* context creation point */ ++ uint64_t bias_clock; /* offset in clock_t */ ++ ++ spinlock_t load_lock; /* lock for the load averages */ ++ atomic_t load_updates; /* nr of load updates done so far */ ++ uint32_t load_last; /* last time load was calculated */ ++ uint32_t load[3]; /* load averages 1,5,15 */ ++ ++ atomic_t total_forks; /* number of forks so far */ ++ ++ struct _vx_syslog syslog; ++}; ++ ++struct _vx_cvirt_pc { ++ struct _vx_usage_stat cpustat; ++}; ++ ++ ++#ifdef CONFIG_VSERVER_DEBUG ++ ++static inline void __dump_vx_cvirt(struct _vx_cvirt *cvirt) ++{ ++ printk("\t_vx_cvirt:\n"); ++ printk("\t threads: %4d, %4d, %4d, %4d\n", ++ atomic_read(&cvirt->nr_threads), ++ atomic_read(&cvirt->nr_running), ++ atomic_read(&cvirt->nr_uninterruptible), ++ atomic_read(&cvirt->nr_onhold)); ++ /* add rest here */ ++ printk("\t total_forks = %d\n", atomic_read(&cvirt->total_forks)); ++} ++ ++#endif ++ ++#endif /* _VX_CVIRT_DEF_H */ +diff -NurpP --minimal linux-3.0.9/include/linux/vserver/debug.h linux-3.0.9-vs2.3.2.1/include/linux/vserver/debug.h +--- linux-3.0.9/include/linux/vserver/debug.h 1970-01-01 01:00:00.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/include/linux/vserver/debug.h 2011-11-17 21:25:04.000000000 +0100 +@@ -0,0 +1,145 @@ ++#ifndef _VX_DEBUG_H ++#define _VX_DEBUG_H ++ ++ ++#define VXD_CBIT(n, m) (vs_debug_ ## n & (1 << (m))) ++#define VXD_CMIN(n, m) (vs_debug_ ## n > (m)) ++#define VXD_MASK(n, m) (vs_debug_ ## n & (m)) ++ ++#define VXD_DEV(d) (d), (d)->bd_inode->i_ino, \ ++ imajor((d)->bd_inode), iminor((d)->bd_inode) ++#define VXF_DEV "%p[%lu,%d:%d]" ++ ++#if defined(CONFIG_QUOTES_UTF8) ++#define VS_Q_LQM "\xc2\xbb" ++#define VS_Q_RQM "\xc2\xab" ++#elif defined(CONFIG_QUOTES_ASCII) ++#define VS_Q_LQM "\x27" ++#define VS_Q_RQM "\x27" ++#else ++#define VS_Q_LQM "\xbb" ++#define VS_Q_RQM "\xab" ++#endif ++ ++#define VS_Q(f) VS_Q_LQM f VS_Q_RQM ++ ++ ++#define vxd_path(p) \ ++ ({ static char _buffer[PATH_MAX]; \ ++ d_path(p, _buffer, sizeof(_buffer)); }) ++ ++#define vxd_cond_path(n) \ ++ ((n) ? vxd_path(&(n)->path) : "<null>" ) ++ ++ ++#ifdef CONFIG_VSERVER_DEBUG ++ ++extern unsigned int vs_debug_switch; ++extern unsigned int vs_debug_xid; ++extern unsigned int vs_debug_nid; ++extern unsigned int vs_debug_tag; ++extern unsigned int vs_debug_net; ++extern unsigned int vs_debug_limit; ++extern unsigned int vs_debug_cres; ++extern unsigned int vs_debug_dlim; ++extern unsigned int vs_debug_quota; ++extern unsigned int vs_debug_cvirt; ++extern unsigned int vs_debug_space; ++extern unsigned int vs_debug_perm; ++extern unsigned int vs_debug_misc; ++ ++ ++#define VX_LOGLEVEL "vxD: " ++#define VX_PROC_FMT "%p: " ++#define VX_PROCESS current ++ ++#define vxdprintk(c, f, x...) \ ++ do { \ ++ if (c) \ ++ printk(VX_LOGLEVEL VX_PROC_FMT f "\n", \ ++ VX_PROCESS , ##x); \ ++ } while (0) ++ ++#define vxlprintk(c, f, x...) \ ++ do { \ ++ if (c) \ ++ printk(VX_LOGLEVEL f " @%s:%d\n", x); \ ++ } while (0) ++ ++#define vxfprintk(c, f, x...) \ ++ do { \ ++ if (c) \ ++ printk(VX_LOGLEVEL f " %s@%s:%d\n", x); \ ++ } while (0) ++ ++ ++struct vx_info; ++ ++void dump_vx_info(struct vx_info *, int); ++void dump_vx_info_inactive(int); ++ ++#else /* CONFIG_VSERVER_DEBUG */ ++ ++#define vs_debug_switch 0 ++#define vs_debug_xid 0 ++#define vs_debug_nid 0 ++#define vs_debug_tag 0 ++#define vs_debug_net 0 ++#define vs_debug_limit 0 ++#define vs_debug_cres 0 ++#define vs_debug_dlim 0 ++#define vs_debug_quota 0 ++#define vs_debug_cvirt 0 ++#define vs_debug_space 0 ++#define vs_debug_perm 0 ++#define vs_debug_misc 0 ++ ++#define vxdprintk(x...) do { } while (0) ++#define vxlprintk(x...) do { } while (0) ++#define vxfprintk(x...) do { } while (0) ++ ++#endif /* CONFIG_VSERVER_DEBUG */ ++ ++ ++#ifdef CONFIG_VSERVER_WARN ++ ++#define VX_WARNLEVEL KERN_WARNING "vxW: " ++#define VX_WARN_TASK "[" VS_Q("%s") ",%u:#%u|%u|%u] " ++#define VX_WARN_XID "[xid #%u] " ++#define VX_WARN_NID "[nid #%u] " ++#define VX_WARN_TAG "[tag #%u] " ++ ++#define vxwprintk(c, f, x...) \ ++ do { \ ++ if (c) \ ++ printk(VX_WARNLEVEL f "\n", ##x); \ ++ } while (0) ++ ++#else /* CONFIG_VSERVER_WARN */ ++ ++#define vxwprintk(x...) do { } while (0) ++ ++#endif /* CONFIG_VSERVER_WARN */ ++ ++#define vxwprintk_task(c, f, x...) \ ++ vxwprintk(c, VX_WARN_TASK f, \ ++ current->comm, current->pid, \ ++ current->xid, current->nid, current->tag, ##x) ++#define vxwprintk_xid(c, f, x...) \ ++ vxwprintk(c, VX_WARN_XID f, current->xid, x) ++#define vxwprintk_nid(c, f, x...) \ ++ vxwprintk(c, VX_WARN_NID f, current->nid, x) ++#define vxwprintk_tag(c, f, x...) \ ++ vxwprintk(c, VX_WARN_TAG f, current->tag, x) ++ ++#ifdef CONFIG_VSERVER_DEBUG ++#define vxd_assert_lock(l) assert_spin_locked(l) ++#define vxd_assert(c, f, x...) vxlprintk(!(c), \ ++ "assertion [" f "] failed.", ##x, __FILE__, __LINE__) ++#else ++#define vxd_assert_lock(l) do { } while (0) ++#define vxd_assert(c, f, x...) do { } while (0) ++#endif ++ ++ ++#endif /* _VX_DEBUG_H */ +diff -NurpP --minimal linux-3.0.9/include/linux/vserver/debug_cmd.h linux-3.0.9-vs2.3.2.1/include/linux/vserver/debug_cmd.h +--- linux-3.0.9/include/linux/vserver/debug_cmd.h 1970-01-01 01:00:00.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/include/linux/vserver/debug_cmd.h 2011-06-10 22:11:24.000000000 +0200 +@@ -0,0 +1,58 @@ ++#ifndef _VX_DEBUG_CMD_H ++#define _VX_DEBUG_CMD_H ++ ++ ++/* debug commands */ ++ ++#define VCMD_dump_history VC_CMD(DEBUG, 1, 0) ++ ++#define VCMD_read_history VC_CMD(DEBUG, 5, 0) ++#define VCMD_read_monitor VC_CMD(DEBUG, 6, 0) ++ ++struct vcmd_read_history_v0 { ++ uint32_t index; ++ uint32_t count; ++ char __user *data; ++}; ++ ++struct vcmd_read_monitor_v0 { ++ uint32_t index; ++ uint32_t count; ++ char __user *data; ++}; ++ ++ ++#ifdef __KERNEL__ ++ ++#ifdef CONFIG_COMPAT ++ ++#include <asm/compat.h> ++ ++struct vcmd_read_history_v0_x32 { ++ uint32_t index; ++ uint32_t count; ++ compat_uptr_t data_ptr; ++}; ++ ++struct vcmd_read_monitor_v0_x32 { ++ uint32_t index; ++ uint32_t count; ++ compat_uptr_t data_ptr; ++}; ++ ++#endif /* CONFIG_COMPAT */ ++ ++extern int vc_dump_history(uint32_t); ++ ++extern int vc_read_history(uint32_t, void __user *); ++extern int vc_read_monitor(uint32_t, void __user *); ++ ++#ifdef CONFIG_COMPAT ++ ++extern int vc_read_history_x32(uint32_t, void __user *); ++extern int vc_read_monitor_x32(uint32_t, void __user *); ++ ++#endif /* CONFIG_COMPAT */ ++ ++#endif /* __KERNEL__ */ ++#endif /* _VX_DEBUG_CMD_H */ +diff -NurpP --minimal linux-3.0.9/include/linux/vserver/device.h linux-3.0.9-vs2.3.2.1/include/linux/vserver/device.h +--- linux-3.0.9/include/linux/vserver/device.h 1970-01-01 01:00:00.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/include/linux/vserver/device.h 2011-06-10 22:11:24.000000000 +0200 +@@ -0,0 +1,15 @@ ++#ifndef _VX_DEVICE_H ++#define _VX_DEVICE_H ++ ++ ++#define DATTR_CREATE 0x00000001 ++#define DATTR_OPEN 0x00000002 ++ ++#define DATTR_REMAP 0x00000010 ++ ++#define DATTR_MASK 0x00000013 ++ ++ ++#else /* _VX_DEVICE_H */ ++#warning duplicate inclusion ++#endif /* _VX_DEVICE_H */ +diff -NurpP --minimal linux-3.0.9/include/linux/vserver/device_cmd.h linux-3.0.9-vs2.3.2.1/include/linux/vserver/device_cmd.h +--- linux-3.0.9/include/linux/vserver/device_cmd.h 1970-01-01 01:00:00.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/include/linux/vserver/device_cmd.h 2011-06-10 22:11:24.000000000 +0200 +@@ -0,0 +1,44 @@ ++#ifndef _VX_DEVICE_CMD_H ++#define _VX_DEVICE_CMD_H ++ ++ ++/* device vserver commands */ ++ ++#define VCMD_set_mapping VC_CMD(DEVICE, 1, 0) ++#define VCMD_unset_mapping VC_CMD(DEVICE, 2, 0) ++ ++struct vcmd_set_mapping_v0 { ++ const char __user *device; ++ const char __user *target; ++ uint32_t flags; ++}; ++ ++ ++#ifdef __KERNEL__ ++ ++#ifdef CONFIG_COMPAT ++ ++#include <asm/compat.h> ++ ++struct vcmd_set_mapping_v0_x32 { ++ compat_uptr_t device_ptr; ++ compat_uptr_t target_ptr; ++ uint32_t flags; ++}; ++ ++#endif /* CONFIG_COMPAT */ ++ ++#include <linux/compiler.h> ++ ++extern int vc_set_mapping(struct vx_info *, void __user *); ++extern int vc_unset_mapping(struct vx_info *, void __user *); ++ ++#ifdef CONFIG_COMPAT ++ ++extern int vc_set_mapping_x32(struct vx_info *, void __user *); ++extern int vc_unset_mapping_x32(struct vx_info *, void __user *); ++ ++#endif /* CONFIG_COMPAT */ ++ ++#endif /* __KERNEL__ */ ++#endif /* _VX_DEVICE_CMD_H */ +diff -NurpP --minimal linux-3.0.9/include/linux/vserver/device_def.h linux-3.0.9-vs2.3.2.1/include/linux/vserver/device_def.h +--- linux-3.0.9/include/linux/vserver/device_def.h 1970-01-01 01:00:00.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/include/linux/vserver/device_def.h 2011-06-10 22:11:24.000000000 +0200 +@@ -0,0 +1,17 @@ ++#ifndef _VX_DEVICE_DEF_H ++#define _VX_DEVICE_DEF_H ++ ++#include <linux/types.h> ++ ++struct vx_dmap_target { ++ dev_t target; ++ uint32_t flags; ++}; ++ ++struct _vx_device { ++#ifdef CONFIG_VSERVER_DEVICE ++ struct vx_dmap_target targets[2]; ++#endif ++}; ++ ++#endif /* _VX_DEVICE_DEF_H */ +diff -NurpP --minimal linux-3.0.9/include/linux/vserver/dlimit.h linux-3.0.9-vs2.3.2.1/include/linux/vserver/dlimit.h +--- linux-3.0.9/include/linux/vserver/dlimit.h 1970-01-01 01:00:00.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/include/linux/vserver/dlimit.h 2011-06-10 22:11:24.000000000 +0200 +@@ -0,0 +1,54 @@ ++#ifndef _VX_DLIMIT_H ++#define _VX_DLIMIT_H ++ ++#include "switch.h" ++ ++ ++#ifdef __KERNEL__ ++ ++/* keep in sync with CDLIM_INFINITY */ ++ ++#define DLIM_INFINITY (~0ULL) ++ ++#include <linux/spinlock.h> ++#include <linux/rcupdate.h> ++ ++struct super_block; ++ ++struct dl_info { ++ struct hlist_node dl_hlist; /* linked list of contexts */ ++ struct rcu_head dl_rcu; /* the rcu head */ ++ tag_t dl_tag; /* context tag */ ++ atomic_t dl_usecnt; /* usage count */ ++ atomic_t dl_refcnt; /* reference count */ ++ ++ struct super_block *dl_sb; /* associated superblock */ ++ ++ spinlock_t dl_lock; /* protect the values */ ++ ++ unsigned long long dl_space_used; /* used space in bytes */ ++ unsigned long long dl_space_total; /* maximum space in bytes */ ++ unsigned long dl_inodes_used; /* used inodes */ ++ unsigned long dl_inodes_total; /* maximum inodes */ ++ ++ unsigned int dl_nrlmult; /* non root limit mult */ ++}; ++ ++struct rcu_head; ++ ++extern void rcu_free_dl_info(struct rcu_head *); ++extern void unhash_dl_info(struct dl_info *); ++ ++extern struct dl_info *locate_dl_info(struct super_block *, tag_t); ++ ++ ++struct kstatfs; ++ ++extern void vx_vsi_statfs(struct super_block *, struct kstatfs *); ++ ++typedef uint64_t dlsize_t; ++ ++#endif /* __KERNEL__ */ ++#else /* _VX_DLIMIT_H */ ++#warning duplicate inclusion ++#endif /* _VX_DLIMIT_H */ +diff -NurpP --minimal linux-3.0.9/include/linux/vserver/dlimit_cmd.h linux-3.0.9-vs2.3.2.1/include/linux/vserver/dlimit_cmd.h +--- linux-3.0.9/include/linux/vserver/dlimit_cmd.h 1970-01-01 01:00:00.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/include/linux/vserver/dlimit_cmd.h 2011-06-10 22:11:24.000000000 +0200 +@@ -0,0 +1,109 @@ ++#ifndef _VX_DLIMIT_CMD_H ++#define _VX_DLIMIT_CMD_H ++ ++ ++/* dlimit vserver commands */ ++ ++#define VCMD_add_dlimit VC_CMD(DLIMIT, 1, 0) ++#define VCMD_rem_dlimit VC_CMD(DLIMIT, 2, 0) ++ ++#define VCMD_set_dlimit VC_CMD(DLIMIT, 5, 0) ++#define VCMD_get_dlimit VC_CMD(DLIMIT, 6, 0) ++ ++struct vcmd_ctx_dlimit_base_v0 { ++ const char __user *name; ++ uint32_t flags; ++}; ++ ++struct vcmd_ctx_dlimit_v0 { ++ const char __user *name; ++ uint32_t space_used; /* used space in kbytes */ ++ uint32_t space_total; /* maximum space in kbytes */ ++ uint32_t inodes_used; /* used inodes */ ++ uint32_t inodes_total; /* maximum inodes */ ++ uint32_t reserved; /* reserved for root in % */ ++ uint32_t flags; ++}; ++ ++#define CDLIM_UNSET ((uint32_t)0UL) ++#define CDLIM_INFINITY ((uint32_t)~0UL) ++#define CDLIM_KEEP ((uint32_t)~1UL) ++ ++#define DLIME_UNIT 0 ++#define DLIME_KILO 1 ++#define DLIME_MEGA 2 ++#define DLIME_GIGA 3 ++ ++#define DLIMF_SHIFT 0x10 ++ ++#define DLIMS_USED 0 ++#define DLIMS_TOTAL 2 ++ ++static inline ++uint64_t dlimit_space_32to64(uint32_t val, uint32_t flags, int shift) ++{ ++ int exp = (flags & DLIMF_SHIFT) ? ++ (flags >> shift) & DLIME_GIGA : DLIME_KILO; ++ return ((uint64_t)val) << (10 * exp); ++} ++ ++static inline ++uint32_t dlimit_space_64to32(uint64_t val, uint32_t *flags, int shift) ++{ ++ int exp = 0; ++ ++ if (*flags & DLIMF_SHIFT) { ++ while (val > (1LL << 32) && (exp < 3)) { ++ val >>= 10; ++ exp++; ++ } ++ *flags &= ~(DLIME_GIGA << shift); ++ *flags |= exp << shift; ++ } else ++ val >>= 10; ++ return val; ++} ++ ++#ifdef __KERNEL__ ++ ++#ifdef CONFIG_COMPAT ++ ++#include <asm/compat.h> ++ ++struct vcmd_ctx_dlimit_base_v0_x32 { ++ compat_uptr_t name_ptr; ++ uint32_t flags; ++}; ++ ++struct vcmd_ctx_dlimit_v0_x32 { ++ compat_uptr_t name_ptr; ++ uint32_t space_used; /* used space in kbytes */ ++ uint32_t space_total; /* maximum space in kbytes */ ++ uint32_t inodes_used; /* used inodes */ ++ uint32_t inodes_total; /* maximum inodes */ ++ uint32_t reserved; /* reserved for root in % */ ++ uint32_t flags; ++}; ++ ++#endif /* CONFIG_COMPAT */ ++ ++#include <linux/compiler.h> ++ ++extern int vc_add_dlimit(uint32_t, void __user *); ++extern int vc_rem_dlimit(uint32_t, void __user *); ++ ++extern int vc_set_dlimit(uint32_t, void __user *); ++extern int vc_get_dlimit(uint32_t, void __user *); ++ ++#ifdef CONFIG_COMPAT ++ ++extern int vc_add_dlimit_x32(uint32_t, void __user *); ++extern int vc_rem_dlimit_x32(uint32_t, void __user *); ++ ++extern int vc_set_dlimit_x32(uint32_t, void __user *); ++extern int vc_get_dlimit_x32(uint32_t, void __user *); ++ ++#endif /* CONFIG_COMPAT */ ++ ++#endif /* __KERNEL__ */ ++#endif /* _VX_DLIMIT_CMD_H */ +diff -NurpP --minimal linux-3.0.9/include/linux/vserver/global.h linux-3.0.9-vs2.3.2.1/include/linux/vserver/global.h +--- linux-3.0.9/include/linux/vserver/global.h 1970-01-01 01:00:00.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/include/linux/vserver/global.h 2011-06-10 22:11:24.000000000 +0200 +@@ -0,0 +1,19 @@ ++#ifndef _VX_GLOBAL_H ++#define _VX_GLOBAL_H ++ ++ ++extern atomic_t vx_global_ctotal; ++extern atomic_t vx_global_cactive; ++ ++extern atomic_t nx_global_ctotal; ++extern atomic_t nx_global_cactive; ++ ++extern atomic_t vs_global_nsproxy; ++extern atomic_t vs_global_fs; ++extern atomic_t vs_global_mnt_ns; ++extern atomic_t vs_global_uts_ns; ++extern atomic_t vs_global_user_ns; ++extern atomic_t vs_global_pid_ns; ++ ++ ++#endif /* _VX_GLOBAL_H */ +diff -NurpP --minimal linux-3.0.9/include/linux/vserver/history.h linux-3.0.9-vs2.3.2.1/include/linux/vserver/history.h +--- linux-3.0.9/include/linux/vserver/history.h 1970-01-01 01:00:00.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/include/linux/vserver/history.h 2011-06-10 22:11:24.000000000 +0200 +@@ -0,0 +1,197 @@ ++#ifndef _VX_HISTORY_H ++#define _VX_HISTORY_H ++ ++ ++enum { ++ VXH_UNUSED = 0, ++ VXH_THROW_OOPS = 1, ++ ++ VXH_GET_VX_INFO, ++ VXH_PUT_VX_INFO, ++ VXH_INIT_VX_INFO, ++ VXH_SET_VX_INFO, ++ VXH_CLR_VX_INFO, ++ VXH_CLAIM_VX_INFO, ++ VXH_RELEASE_VX_INFO, ++ VXH_ALLOC_VX_INFO, ++ VXH_DEALLOC_VX_INFO, ++ VXH_HASH_VX_INFO, ++ VXH_UNHASH_VX_INFO, ++ VXH_LOC_VX_INFO, ++ VXH_LOOKUP_VX_INFO, ++ VXH_CREATE_VX_INFO, ++}; ++ ++struct _vxhe_vxi { ++ struct vx_info *ptr; ++ unsigned xid; ++ unsigned usecnt; ++ unsigned tasks; ++}; ++ ++struct _vxhe_set_clr { ++ void *data; ++}; ++ ++struct _vxhe_loc_lookup { ++ unsigned arg; ++}; ++ ++struct _vx_hist_entry { ++ void *loc; ++ unsigned short seq; ++ unsigned short type; ++ struct _vxhe_vxi vxi; ++ union { ++ struct _vxhe_set_clr sc; ++ struct _vxhe_loc_lookup ll; ++ }; ++}; ++ ++#ifdef CONFIG_VSERVER_HISTORY ++ ++extern unsigned volatile int vxh_active; ++ ++struct _vx_hist_entry *vxh_advance(void *loc); ++ ++ ++static inline ++void __vxh_copy_vxi(struct _vx_hist_entry *entry, struct vx_info *vxi) ++{ ++ entry->vxi.ptr = vxi; ++ if (vxi) { ++ entry->vxi.usecnt = atomic_read(&vxi->vx_usecnt); ++ entry->vxi.tasks = atomic_read(&vxi->vx_tasks); ++ entry->vxi.xid = vxi->vx_id; ++ } ++} ++ ++ ++#define __HERE__ current_text_addr() ++ ++#define __VXH_BODY(__type, __data, __here) \ ++ struct _vx_hist_entry *entry; \ ++ \ ++ preempt_disable(); \ ++ entry = vxh_advance(__here); \ ++ __data; \ ++ entry->type = __type; \ ++ preempt_enable(); ++ ++ ++ /* pass vxi only */ ++ ++#define __VXH_SMPL \ ++ __vxh_copy_vxi(entry, vxi) ++ ++static inline ++void __vxh_smpl(struct vx_info *vxi, int __type, void *__here) ++{ ++ __VXH_BODY(__type, __VXH_SMPL, __here) ++} ++ ++ /* pass vxi and data (void *) */ ++ ++#define __VXH_DATA \ ++ __vxh_copy_vxi(entry, vxi); \ ++ entry->sc.data = data ++ ++static inline ++void __vxh_data(struct vx_info *vxi, void *data, ++ int __type, void *__here) ++{ ++ __VXH_BODY(__type, __VXH_DATA, __here) ++} ++ ++ /* pass vxi and arg (long) */ ++ ++#define __VXH_LONG \ ++ __vxh_copy_vxi(entry, vxi); \ ++ entry->ll.arg = arg ++ ++static inline ++void __vxh_long(struct vx_info *vxi, long arg, ++ int __type, void *__here) ++{ ++ __VXH_BODY(__type, __VXH_LONG, __here) ++} ++ ++ ++static inline ++void __vxh_throw_oops(void *__here) ++{ ++ __VXH_BODY(VXH_THROW_OOPS, {}, __here); ++ /* prevent further acquisition */ ++ vxh_active = 0; ++} ++ ++ ++#define vxh_throw_oops() __vxh_throw_oops(__HERE__); ++ ++#define __vxh_get_vx_info(v, h) __vxh_smpl(v, VXH_GET_VX_INFO, h); ++#define __vxh_put_vx_info(v, h) __vxh_smpl(v, VXH_PUT_VX_INFO, h); ++ ++#define __vxh_init_vx_info(v, d, h) \ ++ __vxh_data(v, d, VXH_INIT_VX_INFO, h); ++#define __vxh_set_vx_info(v, d, h) \ ++ __vxh_data(v, d, VXH_SET_VX_INFO, h); ++#define __vxh_clr_vx_info(v, d, h) \ ++ __vxh_data(v, d, VXH_CLR_VX_INFO, h); ++ ++#define __vxh_claim_vx_info(v, d, h) \ ++ __vxh_data(v, d, VXH_CLAIM_VX_INFO, h); ++#define __vxh_release_vx_info(v, d, h) \ ++ __vxh_data(v, d, VXH_RELEASE_VX_INFO, h); ++ ++#define vxh_alloc_vx_info(v) \ ++ __vxh_smpl(v, VXH_ALLOC_VX_INFO, __HERE__); ++#define vxh_dealloc_vx_info(v) \ ++ __vxh_smpl(v, VXH_DEALLOC_VX_INFO, __HERE__); ++ ++#define vxh_hash_vx_info(v) \ ++ __vxh_smpl(v, VXH_HASH_VX_INFO, __HERE__); ++#define vxh_unhash_vx_info(v) \ ++ __vxh_smpl(v, VXH_UNHASH_VX_INFO, __HERE__); ++ ++#define vxh_loc_vx_info(v, l) \ ++ __vxh_long(v, l, VXH_LOC_VX_INFO, __HERE__); ++#define vxh_lookup_vx_info(v, l) \ ++ __vxh_long(v, l, VXH_LOOKUP_VX_INFO, __HERE__); ++#define vxh_create_vx_info(v, l) \ ++ __vxh_long(v, l, VXH_CREATE_VX_INFO, __HERE__); ++ ++extern void vxh_dump_history(void); ++ ++ ++#else /* CONFIG_VSERVER_HISTORY */ ++ ++#define __HERE__ 0 ++ ++#define vxh_throw_oops() do { } while (0) ++ ++#define __vxh_get_vx_info(v, h) do { } while (0) ++#define __vxh_put_vx_info(v, h) do { } while (0) ++ ++#define __vxh_init_vx_info(v, d, h) do { } while (0) ++#define __vxh_set_vx_info(v, d, h) do { } while (0) ++#define __vxh_clr_vx_info(v, d, h) do { } while (0) ++ ++#define __vxh_claim_vx_info(v, d, h) do { } while (0) ++#define __vxh_release_vx_info(v, d, h) do { } while (0) ++ ++#define vxh_alloc_vx_info(v) do { } while (0) ++#define vxh_dealloc_vx_info(v) do { } while (0) ++ ++#define vxh_hash_vx_info(v) do { } while (0) ++#define vxh_unhash_vx_info(v) do { } while (0) ++ ++#define vxh_loc_vx_info(v, l) do { } while (0) ++#define vxh_lookup_vx_info(v, l) do { } while (0) ++#define vxh_create_vx_info(v, l) do { } while (0) ++ ++#define vxh_dump_history() do { } while (0) ++ ++ ++#endif /* CONFIG_VSERVER_HISTORY */ ++ ++#endif /* _VX_HISTORY_H */ +diff -NurpP --minimal linux-3.0.9/include/linux/vserver/inode.h linux-3.0.9-vs2.3.2.1/include/linux/vserver/inode.h +--- linux-3.0.9/include/linux/vserver/inode.h 1970-01-01 01:00:00.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/include/linux/vserver/inode.h 2011-06-10 22:11:24.000000000 +0200 +@@ -0,0 +1,39 @@ ++#ifndef _VX_INODE_H ++#define _VX_INODE_H ++ ++ ++#define IATTR_TAG 0x01000000 ++ ++#define IATTR_ADMIN 0x00000001 ++#define IATTR_WATCH 0x00000002 ++#define IATTR_HIDE 0x00000004 ++#define IATTR_FLAGS 0x00000007 ++ ++#define IATTR_BARRIER 0x00010000 ++#define IATTR_IXUNLINK 0x00020000 ++#define IATTR_IMMUTABLE 0x00040000 ++#define IATTR_COW 0x00080000 ++ ++#ifdef __KERNEL__ ++ ++ ++#ifdef CONFIG_VSERVER_PROC_SECURE ++#define IATTR_PROC_DEFAULT ( IATTR_ADMIN | IATTR_HIDE ) ++#define IATTR_PROC_SYMLINK ( IATTR_ADMIN ) ++#else ++#define IATTR_PROC_DEFAULT ( IATTR_ADMIN ) ++#define IATTR_PROC_SYMLINK ( IATTR_ADMIN ) ++#endif ++ ++#define vx_hide_check(c, m) (((m) & IATTR_HIDE) ? vx_check(c, m) : 1) ++ ++#endif /* __KERNEL__ */ ++ ++/* inode ioctls */ ++ ++#define FIOC_GETXFLG _IOR('x', 5, long) ++#define FIOC_SETXFLG _IOW('x', 6, long) ++ ++#else /* _VX_INODE_H */ ++#warning duplicate inclusion ++#endif /* _VX_INODE_H */ +diff -NurpP --minimal linux-3.0.9/include/linux/vserver/inode_cmd.h linux-3.0.9-vs2.3.2.1/include/linux/vserver/inode_cmd.h +--- linux-3.0.9/include/linux/vserver/inode_cmd.h 1970-01-01 01:00:00.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/include/linux/vserver/inode_cmd.h 2011-06-10 22:11:24.000000000 +0200 +@@ -0,0 +1,59 @@ ++#ifndef _VX_INODE_CMD_H ++#define _VX_INODE_CMD_H ++ ++ ++/* inode vserver commands */ ++ ++#define VCMD_get_iattr VC_CMD(INODE, 1, 1) ++#define VCMD_set_iattr VC_CMD(INODE, 2, 1) ++ ++#define VCMD_fget_iattr VC_CMD(INODE, 3, 0) ++#define VCMD_fset_iattr VC_CMD(INODE, 4, 0) ++ ++struct vcmd_ctx_iattr_v1 { ++ const char __user *name; ++ uint32_t tag; ++ uint32_t flags; ++ uint32_t mask; ++}; ++ ++struct vcmd_ctx_fiattr_v0 { ++ uint32_t tag; ++ uint32_t flags; ++ uint32_t mask; ++}; ++ ++ ++#ifdef __KERNEL__ ++ ++ ++#ifdef CONFIG_COMPAT ++ ++#include <asm/compat.h> ++ ++struct vcmd_ctx_iattr_v1_x32 { ++ compat_uptr_t name_ptr; ++ uint32_t tag; ++ uint32_t flags; ++ uint32_t mask; ++}; ++ ++#endif /* CONFIG_COMPAT */ ++ ++#include <linux/compiler.h> ++ ++extern int vc_get_iattr(void __user *); ++extern int vc_set_iattr(void __user *); ++ ++extern int vc_fget_iattr(uint32_t, void __user *); ++extern int vc_fset_iattr(uint32_t, void __user *); ++ ++#ifdef CONFIG_COMPAT ++ ++extern int vc_get_iattr_x32(void __user *); ++extern int vc_set_iattr_x32(void __user *); ++ ++#endif /* CONFIG_COMPAT */ ++ ++#endif /* __KERNEL__ */ ++#endif /* _VX_INODE_CMD_H */ +diff -NurpP --minimal linux-3.0.9/include/linux/vserver/limit.h linux-3.0.9-vs2.3.2.1/include/linux/vserver/limit.h +--- linux-3.0.9/include/linux/vserver/limit.h 1970-01-01 01:00:00.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/include/linux/vserver/limit.h 2011-06-10 22:11:24.000000000 +0200 +@@ -0,0 +1,71 @@ ++#ifndef _VX_LIMIT_H ++#define _VX_LIMIT_H ++ ++#define VLIMIT_NSOCK 16 ++#define VLIMIT_OPENFD 17 ++#define VLIMIT_ANON 18 ++#define VLIMIT_SHMEM 19 ++#define VLIMIT_SEMARY 20 ++#define VLIMIT_NSEMS 21 ++#define VLIMIT_DENTRY 22 ++#define VLIMIT_MAPPED 23 ++ ++ ++#ifdef __KERNEL__ ++ ++#define VLIM_NOCHECK ((1L << VLIMIT_DENTRY) | (1L << RLIMIT_RSS)) ++ ++/* keep in sync with CRLIM_INFINITY */ ++ ++#define VLIM_INFINITY (~0ULL) ++ ++#include <asm/atomic.h> ++#include <asm/resource.h> ++ ++#ifndef RLIM_INFINITY ++#warning RLIM_INFINITY is undefined ++#endif ++ ++#define __rlim_val(l, r, v) ((l)->res[r].v) ++ ++#define __rlim_soft(l, r) __rlim_val(l, r, soft) ++#define __rlim_hard(l, r) __rlim_val(l, r, hard) ++ ++#define __rlim_rcur(l, r) __rlim_val(l, r, rcur) ++#define __rlim_rmin(l, r) __rlim_val(l, r, rmin) ++#define __rlim_rmax(l, r) __rlim_val(l, r, rmax) ++ ++#define __rlim_lhit(l, r) __rlim_val(l, r, lhit) ++#define __rlim_hit(l, r) atomic_inc(&__rlim_lhit(l, r)) ++ ++typedef atomic_long_t rlim_atomic_t; ++typedef unsigned long rlim_t; ++ ++#define __rlim_get(l, r) atomic_long_read(&__rlim_rcur(l, r)) ++#define __rlim_set(l, r, v) atomic_long_set(&__rlim_rcur(l, r), v) ++#define __rlim_inc(l, r) atomic_long_inc(&__rlim_rcur(l, r)) ++#define __rlim_dec(l, r) atomic_long_dec(&__rlim_rcur(l, r)) ++#define __rlim_add(l, r, v) atomic_long_add(v, &__rlim_rcur(l, r)) ++#define __rlim_sub(l, r, v) atomic_long_sub(v, &__rlim_rcur(l, r)) ++ ++ ++#if (RLIM_INFINITY == VLIM_INFINITY) ++#define VX_VLIM(r) ((long long)(long)(r)) ++#define VX_RLIM(v) ((rlim_t)(v)) ++#else ++#define VX_VLIM(r) (((r) == RLIM_INFINITY) \ ++ ? VLIM_INFINITY : (long long)(r)) ++#define VX_RLIM(v) (((v) == VLIM_INFINITY) \ ++ ? RLIM_INFINITY : (rlim_t)(v)) ++#endif ++ ++struct sysinfo; ++ ++void vx_vsi_meminfo(struct sysinfo *); ++void vx_vsi_swapinfo(struct sysinfo *); ++long vx_vsi_cached(struct sysinfo *); ++ ++#define NUM_LIMITS 24 ++ ++#endif /* __KERNEL__ */ ++#endif /* _VX_LIMIT_H */ +diff -NurpP --minimal linux-3.0.9/include/linux/vserver/limit_cmd.h linux-3.0.9-vs2.3.2.1/include/linux/vserver/limit_cmd.h +--- linux-3.0.9/include/linux/vserver/limit_cmd.h 1970-01-01 01:00:00.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/include/linux/vserver/limit_cmd.h 2011-06-10 22:11:24.000000000 +0200 +@@ -0,0 +1,71 @@ ++#ifndef _VX_LIMIT_CMD_H ++#define _VX_LIMIT_CMD_H ++ ++ ++/* rlimit vserver commands */ ++ ++#define VCMD_get_rlimit VC_CMD(RLIMIT, 1, 0) ++#define VCMD_set_rlimit VC_CMD(RLIMIT, 2, 0) ++#define VCMD_get_rlimit_mask VC_CMD(RLIMIT, 3, 0) ++#define VCMD_reset_hits VC_CMD(RLIMIT, 7, 0) ++#define VCMD_reset_minmax VC_CMD(RLIMIT, 9, 0) ++ ++struct vcmd_ctx_rlimit_v0 { ++ uint32_t id; ++ uint64_t minimum; ++ uint64_t softlimit; ++ uint64_t maximum; ++}; ++ ++struct vcmd_ctx_rlimit_mask_v0 { ++ uint32_t minimum; ++ uint32_t softlimit; ++ uint32_t maximum; ++}; ++ ++#define VCMD_rlimit_stat VC_CMD(VSTAT, 1, 0) ++ ++struct vcmd_rlimit_stat_v0 { ++ uint32_t id; ++ uint32_t hits; ++ uint64_t value; ++ uint64_t minimum; ++ uint64_t maximum; ++}; ++ ++#define CRLIM_UNSET (0ULL) ++#define CRLIM_INFINITY (~0ULL) ++#define CRLIM_KEEP (~1ULL) ++ ++#ifdef __KERNEL__ ++ ++#ifdef CONFIG_IA32_EMULATION ++ ++struct vcmd_ctx_rlimit_v0_x32 { ++ uint32_t id; ++ uint64_t minimum; ++ uint64_t softlimit; ++ uint64_t maximum; ++} __attribute__ ((packed)); ++ ++#endif /* CONFIG_IA32_EMULATION */ ++ ++#include <linux/compiler.h> ++ ++extern int vc_get_rlimit_mask(uint32_t, void __user *); ++extern int vc_get_rlimit(struct vx_info *, void __user *); ++extern int vc_set_rlimit(struct vx_info *, void __user *); ++extern int vc_reset_hits(struct vx_info *, void __user *); ++extern int vc_reset_minmax(struct vx_info *, void __user *); ++ ++extern int vc_rlimit_stat(struct vx_info *, void __user *); ++ ++#ifdef CONFIG_IA32_EMULATION ++ ++extern int vc_get_rlimit_x32(struct vx_info *, void __user *); ++extern int vc_set_rlimit_x32(struct vx_info *, void __user *); ++ ++#endif /* CONFIG_IA32_EMULATION */ ++ ++#endif /* __KERNEL__ */ ++#endif /* _VX_LIMIT_CMD_H */ +diff -NurpP --minimal linux-3.0.9/include/linux/vserver/limit_def.h linux-3.0.9-vs2.3.2.1/include/linux/vserver/limit_def.h +--- linux-3.0.9/include/linux/vserver/limit_def.h 1970-01-01 01:00:00.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/include/linux/vserver/limit_def.h 2011-06-10 22:11:24.000000000 +0200 +@@ -0,0 +1,47 @@ ++#ifndef _VX_LIMIT_DEF_H ++#define _VX_LIMIT_DEF_H ++ ++#include <asm/atomic.h> ++#include <asm/resource.h> ++ ++#include "limit.h" ++ ++ ++struct _vx_res_limit { ++ rlim_t soft; /* Context soft limit */ ++ rlim_t hard; /* Context hard limit */ ++ ++ rlim_atomic_t rcur; /* Current value */ ++ rlim_t rmin; /* Context minimum */ ++ rlim_t rmax; /* Context maximum */ ++ ++ atomic_t lhit; /* Limit hits */ ++}; ++ ++/* context sub struct */ ++ ++struct _vx_limit { ++ struct _vx_res_limit res[NUM_LIMITS]; ++}; ++ ++#ifdef CONFIG_VSERVER_DEBUG ++ ++static inline void __dump_vx_limit(struct _vx_limit *limit) ++{ ++ int i; ++ ++ printk("\t_vx_limit:"); ++ for (i = 0; i < NUM_LIMITS; i++) { ++ printk("\t [%2d] = %8lu %8lu/%8lu, %8ld/%8ld, %8d\n", ++ i, (unsigned long)__rlim_get(limit, i), ++ (unsigned long)__rlim_rmin(limit, i), ++ (unsigned long)__rlim_rmax(limit, i), ++ (long)__rlim_soft(limit, i), ++ (long)__rlim_hard(limit, i), ++ atomic_read(&__rlim_lhit(limit, i))); ++ } ++} ++ ++#endif ++ ++#endif /* _VX_LIMIT_DEF_H */ +diff -NurpP --minimal linux-3.0.9/include/linux/vserver/limit_int.h linux-3.0.9-vs2.3.2.1/include/linux/vserver/limit_int.h +--- linux-3.0.9/include/linux/vserver/limit_int.h 1970-01-01 01:00:00.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/include/linux/vserver/limit_int.h 2011-06-10 22:11:24.000000000 +0200 +@@ -0,0 +1,198 @@ ++#ifndef _VX_LIMIT_INT_H ++#define _VX_LIMIT_INT_H ++ ++#include "context.h" ++ ++#ifdef __KERNEL__ ++ ++#define VXD_RCRES_COND(r) VXD_CBIT(cres, r) ++#define VXD_RLIMIT_COND(r) VXD_CBIT(limit, r) ++ ++extern const char *vlimit_name[NUM_LIMITS]; ++ ++static inline void __vx_acc_cres(struct vx_info *vxi, ++ int res, int dir, void *_data, char *_file, int _line) ++{ ++ if (VXD_RCRES_COND(res)) ++ vxlprintk(1, "vx_acc_cres[%5d,%s,%2d]: %5ld%s (%p)", ++ (vxi ? vxi->vx_id : -1), vlimit_name[res], res, ++ (vxi ? (long)__rlim_get(&vxi->limit, res) : 0), ++ (dir > 0) ? "++" : "--", _data, _file, _line); ++ if (!vxi) ++ return; ++ ++ if (dir > 0) ++ __rlim_inc(&vxi->limit, res); ++ else ++ __rlim_dec(&vxi->limit, res); ++} ++ ++static inline void __vx_add_cres(struct vx_info *vxi, ++ int res, int amount, void *_data, char *_file, int _line) ++{ ++ if (VXD_RCRES_COND(res)) ++ vxlprintk(1, "vx_add_cres[%5d,%s,%2d]: %5ld += %5d (%p)", ++ (vxi ? vxi->vx_id : -1), vlimit_name[res], res, ++ (vxi ? (long)__rlim_get(&vxi->limit, res) : 0), ++ amount, _data, _file, _line); ++ if (amount == 0) ++ return; ++ if (!vxi) ++ return; ++ __rlim_add(&vxi->limit, res, amount); ++} ++ ++static inline ++int __vx_cres_adjust_max(struct _vx_limit *limit, int res, rlim_t value) ++{ ++ int cond = (value > __rlim_rmax(limit, res)); ++ ++ if (cond) ++ __rlim_rmax(limit, res) = value; ++ return cond; ++} ++ ++static inline ++int __vx_cres_adjust_min(struct _vx_limit *limit, int res, rlim_t value) ++{ ++ int cond = (value < __rlim_rmin(limit, res)); ++ ++ if (cond) ++ __rlim_rmin(limit, res) = value; ++ return cond; ++} ++ ++static inline ++void __vx_cres_fixup(struct _vx_limit *limit, int res, rlim_t value) ++{ ++ if (!__vx_cres_adjust_max(limit, res, value)) ++ __vx_cres_adjust_min(limit, res, value); ++} ++ ++ ++/* return values: ++ +1 ... no limit hit ++ -1 ... over soft limit ++ 0 ... over hard limit */ ++ ++static inline int __vx_cres_avail(struct vx_info *vxi, ++ int res, int num, char *_file, int _line) ++{ ++ struct _vx_limit *limit; ++ rlim_t value; ++ ++ if (VXD_RLIMIT_COND(res)) ++ vxlprintk(1, "vx_cres_avail[%5d,%s,%2d]: %5ld/%5ld > %5ld + %5d", ++ (vxi ? vxi->vx_id : -1), vlimit_name[res], res, ++ (vxi ? (long)__rlim_soft(&vxi->limit, res) : -1), ++ (vxi ? (long)__rlim_hard(&vxi->limit, res) : -1), ++ (vxi ? (long)__rlim_get(&vxi->limit, res) : 0), ++ num, _file, _line); ++ if (!vxi) ++ return 1; ++ ++ limit = &vxi->limit; ++ value = __rlim_get(limit, res); ++ ++ if (!__vx_cres_adjust_max(limit, res, value)) ++ __vx_cres_adjust_min(limit, res, value); ++ ++ if (num == 0) ++ return 1; ++ ++ if (__rlim_soft(limit, res) == RLIM_INFINITY) ++ return -1; ++ if (value + num <= __rlim_soft(limit, res)) ++ return -1; ++ ++ if (__rlim_hard(limit, res) == RLIM_INFINITY) ++ return 1; ++ if (value + num <= __rlim_hard(limit, res)) ++ return 1; ++ ++ __rlim_hit(limit, res); ++ return 0; ++} ++ ++ ++static const int VLA_RSS[] = { RLIMIT_RSS, VLIMIT_ANON, VLIMIT_MAPPED, 0 }; ++ ++static inline ++rlim_t __vx_cres_array_sum(struct _vx_limit *limit, const int *array) ++{ ++ rlim_t value, sum = 0; ++ int res; ++ ++ while ((res = *array++)) { ++ value = __rlim_get(limit, res); ++ __vx_cres_fixup(limit, res, value); ++ sum += value; ++ } ++ return sum; ++} ++ ++static inline ++rlim_t __vx_cres_array_fixup(struct _vx_limit *limit, const int *array) ++{ ++ rlim_t value = __vx_cres_array_sum(limit, array + 1); ++ int res = *array; ++ ++ if (value == __rlim_get(limit, res)) ++ return value; ++ ++ __rlim_set(limit, res, value); ++ /* now adjust min/max */ ++ if (!__vx_cres_adjust_max(limit, res, value)) ++ __vx_cres_adjust_min(limit, res, value); ++ ++ return value; ++} ++ ++static inline int __vx_cres_array_avail(struct vx_info *vxi, ++ const int *array, int num, char *_file, int _line) ++{ ++ struct _vx_limit *limit; ++ rlim_t value = 0; ++ int res; ++ ++ if (num == 0) ++ return 1; ++ if (!vxi) ++ return 1; ++ ++ limit = &vxi->limit; ++ res = *array; ++ value = __vx_cres_array_sum(limit, array + 1); ++ ++ __rlim_set(limit, res, value); ++ __vx_cres_fixup(limit, res, value); ++ ++ return __vx_cres_avail(vxi, res, num, _file, _line); ++} ++ ++ ++static inline void vx_limit_fixup(struct _vx_limit *limit, int id) ++{ ++ rlim_t value; ++ int res; ++ ++ /* complex resources first */ ++ if ((id < 0) || (id == RLIMIT_RSS)) ++ __vx_cres_array_fixup(limit, VLA_RSS); ++ ++ for (res = 0; res < NUM_LIMITS; res++) { ++ if ((id > 0) && (res != id)) ++ continue; ++ ++ value = __rlim_get(limit, res); ++ __vx_cres_fixup(limit, res, value); ++ ++ /* not supposed to happen, maybe warn? */ ++ if (__rlim_rmax(limit, res) > __rlim_hard(limit, res)) ++ __rlim_rmax(limit, res) = __rlim_hard(limit, res); ++ } ++} ++ ++ ++#endif /* __KERNEL__ */ ++#endif /* _VX_LIMIT_INT_H */ +diff -NurpP --minimal linux-3.0.9/include/linux/vserver/monitor.h linux-3.0.9-vs2.3.2.1/include/linux/vserver/monitor.h +--- linux-3.0.9/include/linux/vserver/monitor.h 1970-01-01 01:00:00.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/include/linux/vserver/monitor.h 2011-06-10 22:11:24.000000000 +0200 +@@ -0,0 +1,96 @@ ++#ifndef _VX_MONITOR_H ++#define _VX_MONITOR_H ++ ++#include <linux/types.h> ++ ++enum { ++ VXM_UNUSED = 0, ++ ++ VXM_SYNC = 0x10, ++ ++ VXM_UPDATE = 0x20, ++ VXM_UPDATE_1, ++ VXM_UPDATE_2, ++ ++ VXM_RQINFO_1 = 0x24, ++ VXM_RQINFO_2, ++ ++ VXM_ACTIVATE = 0x40, ++ VXM_DEACTIVATE, ++ VXM_IDLE, ++ ++ VXM_HOLD = 0x44, ++ VXM_UNHOLD, ++ ++ VXM_MIGRATE = 0x48, ++ VXM_RESCHED, ++ ++ /* all other bits are flags */ ++ VXM_SCHED = 0x80, ++}; ++ ++struct _vxm_update_1 { ++ uint32_t tokens_max; ++ uint32_t fill_rate; ++ uint32_t interval; ++}; ++ ++struct _vxm_update_2 { ++ uint32_t tokens_min; ++ uint32_t fill_rate; ++ uint32_t interval; ++}; ++ ++struct _vxm_rqinfo_1 { ++ uint16_t running; ++ uint16_t onhold; ++ uint16_t iowait; ++ uint16_t uintr; ++ uint32_t idle_tokens; ++}; ++ ++struct _vxm_rqinfo_2 { ++ uint32_t norm_time; ++ uint32_t idle_time; ++ uint32_t idle_skip; ++}; ++ ++struct _vxm_sched { ++ uint32_t tokens; ++ uint32_t norm_time; ++ uint32_t idle_time; ++}; ++ ++struct _vxm_task { ++ uint16_t pid; ++ uint16_t state; ++}; ++ ++struct _vxm_event { ++ uint32_t jif; ++ union { ++ uint32_t seq; ++ uint32_t sec; ++ }; ++ union { ++ uint32_t tokens; ++ uint32_t nsec; ++ struct _vxm_task tsk; ++ }; ++}; ++ ++struct _vx_mon_entry { ++ uint16_t type; ++ uint16_t xid; ++ union { ++ struct _vxm_event ev; ++ struct _vxm_sched sd; ++ struct _vxm_update_1 u1; ++ struct _vxm_update_2 u2; ++ struct _vxm_rqinfo_1 q1; ++ struct _vxm_rqinfo_2 q2; ++ }; ++}; ++ ++ ++#endif /* _VX_MONITOR_H */ +diff -NurpP --minimal linux-3.0.9/include/linux/vserver/network.h linux-3.0.9-vs2.3.2.1/include/linux/vserver/network.h +--- linux-3.0.9/include/linux/vserver/network.h 1970-01-01 01:00:00.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/include/linux/vserver/network.h 2011-08-18 16:30:48.000000000 +0200 +@@ -0,0 +1,148 @@ ++#ifndef _VX_NETWORK_H ++#define _VX_NETWORK_H ++ ++#include <linux/types.h> ++ ++ ++#define MAX_N_CONTEXT 65535 /* Arbitrary limit */ ++ ++ ++/* network flags */ ++ ++#define NXF_INFO_PRIVATE 0x00000008 ++ ++#define NXF_SINGLE_IP 0x00000100 ++#define NXF_LBACK_REMAP 0x00000200 ++#define NXF_LBACK_ALLOW 0x00000400 ++ ++#define NXF_HIDE_NETIF 0x02000000 ++#define NXF_HIDE_LBACK 0x04000000 ++ ++#define NXF_STATE_SETUP (1ULL << 32) ++#define NXF_STATE_ADMIN (1ULL << 34) ++ ++#define NXF_SC_HELPER (1ULL << 36) ++#define NXF_PERSISTENT (1ULL << 38) ++ ++#define NXF_ONE_TIME (0x0005ULL << 32) ++ ++ ++#define NXF_INIT_SET (__nxf_init_set()) ++ ++static inline uint64_t __nxf_init_set(void) { ++ return NXF_STATE_ADMIN ++#ifdef CONFIG_VSERVER_AUTO_LBACK ++ | NXF_LBACK_REMAP ++ | NXF_HIDE_LBACK ++#endif ++#ifdef CONFIG_VSERVER_AUTO_SINGLE ++ | NXF_SINGLE_IP ++#endif ++ | NXF_HIDE_NETIF; ++} ++ ++ ++/* network caps */ ++ ++#define NXC_TUN_CREATE 0x00000001 ++ ++#define NXC_RAW_ICMP 0x00000100 ++ ++#define NXC_MULTICAST 0x00001000 ++ ++ ++/* address types */ ++ ++#define NXA_TYPE_IPV4 0x0001 ++#define NXA_TYPE_IPV6 0x0002 ++ ++#define NXA_TYPE_NONE 0x0000 ++#define NXA_TYPE_ANY 0x00FF ++ ++#define NXA_TYPE_ADDR 0x0010 ++#define NXA_TYPE_MASK 0x0020 ++#define NXA_TYPE_RANGE 0x0040 ++ ++#define NXA_MASK_ALL (NXA_TYPE_ADDR | NXA_TYPE_MASK | NXA_TYPE_RANGE) ++ ++#define NXA_MOD_BCAST 0x0100 ++#define NXA_MOD_LBACK 0x0200 ++ ++#define NXA_LOOPBACK 0x1000 ++ ++#define NXA_MASK_BIND (NXA_MASK_ALL | NXA_MOD_BCAST | NXA_MOD_LBACK) ++#define NXA_MASK_SHOW (NXA_MASK_ALL | NXA_LOOPBACK) ++ ++#ifdef __KERNEL__ ++ ++#include <linux/list.h> ++#include <linux/spinlock.h> ++#include <linux/rcupdate.h> ++#include <linux/in.h> ++#include <linux/in6.h> ++#include <asm/atomic.h> ++ ++struct nx_addr_v4 { ++ struct nx_addr_v4 *next; ++ struct in_addr ip[2]; ++ struct in_addr mask; ++ uint16_t type; ++ uint16_t flags; ++}; ++ ++struct nx_addr_v6 { ++ struct nx_addr_v6 *next; ++ struct in6_addr ip; ++ struct in6_addr mask; ++ uint32_t prefix; ++ uint16_t type; ++ uint16_t flags; ++}; ++ ++struct nx_info { ++ struct hlist_node nx_hlist; /* linked list of nxinfos */ ++ nid_t nx_id; /* vnet id */ ++ atomic_t nx_usecnt; /* usage count */ ++ atomic_t nx_tasks; /* tasks count */ ++ int nx_state; /* context state */ ++ ++ uint64_t nx_flags; /* network flag word */ ++ uint64_t nx_ncaps; /* network capabilities */ ++ ++ struct in_addr v4_lback; /* Loopback address */ ++ struct in_addr v4_bcast; /* Broadcast address */ ++ struct nx_addr_v4 v4; /* First/Single ipv4 address */ ++#ifdef CONFIG_IPV6 ++ struct nx_addr_v6 v6; /* First/Single ipv6 address */ ++#endif ++ char nx_name[65]; /* network context name */ ++}; ++ ++ ++/* status flags */ ++ ++#define NXS_HASHED 0x0001 ++#define NXS_SHUTDOWN 0x0100 ++#define NXS_RELEASED 0x8000 ++ ++extern struct nx_info *lookup_nx_info(int); ++ ++extern int get_nid_list(int, unsigned int *, int); ++extern int nid_is_hashed(nid_t); ++ ++extern int nx_migrate_task(struct task_struct *, struct nx_info *); ++ ++extern long vs_net_change(struct nx_info *, unsigned int); ++ ++struct sock; ++ ++ ++#define NX_IPV4(n) ((n)->v4.type != NXA_TYPE_NONE) ++#ifdef CONFIG_IPV6 ++#define NX_IPV6(n) ((n)->v6.type != NXA_TYPE_NONE) ++#else ++#define NX_IPV6(n) (0) ++#endif ++ ++#endif /* __KERNEL__ */ ++#endif /* _VX_NETWORK_H */ +diff -NurpP --minimal linux-3.0.9/include/linux/vserver/network_cmd.h linux-3.0.9-vs2.3.2.1/include/linux/vserver/network_cmd.h +--- linux-3.0.9/include/linux/vserver/network_cmd.h 1970-01-01 01:00:00.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/include/linux/vserver/network_cmd.h 2011-06-10 22:11:24.000000000 +0200 +@@ -0,0 +1,164 @@ ++#ifndef _VX_NETWORK_CMD_H ++#define _VX_NETWORK_CMD_H ++ ++ ++/* vinfo commands */ ++ ++#define VCMD_task_nid VC_CMD(VINFO, 2, 0) ++ ++#ifdef __KERNEL__ ++extern int vc_task_nid(uint32_t); ++ ++#endif /* __KERNEL__ */ ++ ++#define VCMD_nx_info VC_CMD(VINFO, 6, 0) ++ ++struct vcmd_nx_info_v0 { ++ uint32_t nid; ++ /* more to come */ ++}; ++ ++#ifdef __KERNEL__ ++extern int vc_nx_info(struct nx_info *, void __user *); ++ ++#endif /* __KERNEL__ */ ++ ++#include <linux/in.h> ++#include <linux/in6.h> ++ ++#define VCMD_net_create_v0 VC_CMD(VNET, 1, 0) ++#define VCMD_net_create VC_CMD(VNET, 1, 1) ++ ++struct vcmd_net_create { ++ uint64_t flagword; ++}; ++ ++#define VCMD_net_migrate VC_CMD(NETMIG, 1, 0) ++ ++#define VCMD_net_add VC_CMD(NETALT, 1, 0) ++#define VCMD_net_remove VC_CMD(NETALT, 2, 0) ++ ++struct vcmd_net_addr_v0 { ++ uint16_t type; ++ uint16_t count; ++ struct in_addr ip[4]; ++ struct in_addr mask[4]; ++}; ++ ++#define VCMD_net_add_ipv4_v1 VC_CMD(NETALT, 1, 1) ++#define VCMD_net_rem_ipv4_v1 VC_CMD(NETALT, 2, 1) ++ ++struct vcmd_net_addr_ipv4_v1 { ++ uint16_t type; ++ uint16_t flags; ++ struct in_addr ip; ++ struct in_addr mask; ++}; ++ ++#define VCMD_net_add_ipv4 VC_CMD(NETALT, 1, 2) ++#define VCMD_net_rem_ipv4 VC_CMD(NETALT, 2, 2) ++ ++struct vcmd_net_addr_ipv4_v2 { ++ uint16_t type; ++ uint16_t flags; ++ struct in_addr ip; ++ struct in_addr ip2; ++ struct in_addr mask; ++}; ++ ++#define VCMD_net_add_ipv6 VC_CMD(NETALT, 3, 1) ++#define VCMD_net_remove_ipv6 VC_CMD(NETALT, 4, 1) ++ ++struct vcmd_net_addr_ipv6_v1 { ++ uint16_t type; ++ uint16_t flags; ++ uint32_t prefix; ++ struct in6_addr ip; ++ struct in6_addr mask; ++}; ++ ++#define VCMD_add_match_ipv4 VC_CMD(NETALT, 5, 0) ++#define VCMD_get_match_ipv4 VC_CMD(NETALT, 6, 0) ++ ++struct vcmd_match_ipv4_v0 { ++ uint16_t type; ++ uint16_t flags; ++ uint16_t parent; ++ uint16_t prefix; ++ struct in_addr ip; ++ struct in_addr ip2; ++ struct in_addr mask; ++}; ++ ++#define VCMD_add_match_ipv6 VC_CMD(NETALT, 7, 0) ++#define VCMD_get_match_ipv6 VC_CMD(NETALT, 8, 0) ++ ++struct vcmd_match_ipv6_v0 { ++ uint16_t type; ++ uint16_t flags; ++ uint16_t parent; ++ uint16_t prefix; ++ struct in6_addr ip; ++ struct in6_addr ip2; ++ struct in6_addr mask; ++}; ++ ++ ++#ifdef __KERNEL__ ++extern int vc_net_create(uint32_t, void __user *); ++extern int vc_net_migrate(struct nx_info *, void __user *); ++ ++extern int vc_net_add(struct nx_info *, void __user *); ++extern int vc_net_remove(struct nx_info *, void __user *); ++ ++extern int vc_net_add_ipv4_v1(struct nx_info *, void __user *); ++extern int vc_net_add_ipv4(struct nx_info *, void __user *); ++ ++extern int vc_net_rem_ipv4_v1(struct nx_info *, void __user *); ++extern int vc_net_rem_ipv4(struct nx_info *, void __user *); ++ ++extern int vc_net_add_ipv6(struct nx_info *, void __user *); ++extern int vc_net_remove_ipv6(struct nx_info *, void __user *); ++ ++extern int vc_add_match_ipv4(struct nx_info *, void __user *); ++extern int vc_get_match_ipv4(struct nx_info *, void __user *); ++ ++extern int vc_add_match_ipv6(struct nx_info *, void __user *); ++extern int vc_get_match_ipv6(struct nx_info *, void __user *); ++ ++#endif /* __KERNEL__ */ ++ ++ ++/* flag commands */ ++ ++#define VCMD_get_nflags VC_CMD(FLAGS, 5, 0) ++#define VCMD_set_nflags VC_CMD(FLAGS, 6, 0) ++ ++struct vcmd_net_flags_v0 { ++ uint64_t flagword; ++ uint64_t mask; ++}; ++ ++#ifdef __KERNEL__ ++extern int vc_get_nflags(struct nx_info *, void __user *); ++extern int vc_set_nflags(struct nx_info *, void __user *); ++ ++#endif /* __KERNEL__ */ ++ ++ ++/* network caps commands */ ++ ++#define VCMD_get_ncaps VC_CMD(FLAGS, 7, 0) ++#define VCMD_set_ncaps VC_CMD(FLAGS, 8, 0) ++ ++struct vcmd_net_caps_v0 { ++ uint64_t ncaps; ++ uint64_t cmask; ++}; ++ ++#ifdef __KERNEL__ ++extern int vc_get_ncaps(struct nx_info *, void __user *); ++extern int vc_set_ncaps(struct nx_info *, void __user *); ++ ++#endif /* __KERNEL__ */ ++#endif /* _VX_CONTEXT_CMD_H */ +diff -NurpP --minimal linux-3.0.9/include/linux/vserver/percpu.h linux-3.0.9-vs2.3.2.1/include/linux/vserver/percpu.h +--- linux-3.0.9/include/linux/vserver/percpu.h 1970-01-01 01:00:00.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/include/linux/vserver/percpu.h 2011-06-10 22:11:24.000000000 +0200 +@@ -0,0 +1,14 @@ ++#ifndef _VX_PERCPU_H ++#define _VX_PERCPU_H ++ ++#include "cvirt_def.h" ++#include "sched_def.h" ++ ++struct _vx_percpu { ++ struct _vx_cvirt_pc cvirt; ++ struct _vx_sched_pc sched; ++}; ++ ++#define PERCPU_PERCTX (sizeof(struct _vx_percpu)) ++ ++#endif /* _VX_PERCPU_H */ +diff -NurpP --minimal linux-3.0.9/include/linux/vserver/pid.h linux-3.0.9-vs2.3.2.1/include/linux/vserver/pid.h +--- linux-3.0.9/include/linux/vserver/pid.h 1970-01-01 01:00:00.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/include/linux/vserver/pid.h 2011-06-10 22:11:24.000000000 +0200 +@@ -0,0 +1,51 @@ ++#ifndef _VSERVER_PID_H ++#define _VSERVER_PID_H ++ ++/* pid faking stuff */ ++ ++#define vx_info_map_pid(v, p) \ ++ __vx_info_map_pid((v), (p), __func__, __FILE__, __LINE__) ++#define vx_info_map_tgid(v,p) vx_info_map_pid(v,p) ++#define vx_map_pid(p) vx_info_map_pid(current_vx_info(), p) ++#define vx_map_tgid(p) vx_map_pid(p) ++ ++static inline int __vx_info_map_pid(struct vx_info *vxi, int pid, ++ const char *func, const char *file, int line) ++{ ++ if (vx_info_flags(vxi, VXF_INFO_INIT, 0)) { ++ vxfprintk(VXD_CBIT(cvirt, 2), ++ "vx_map_tgid: %p/%llx: %d -> %d", ++ vxi, (long long)vxi->vx_flags, pid, ++ (pid && pid == vxi->vx_initpid) ? 1 : pid, ++ func, file, line); ++ if (pid == 0) ++ return 0; ++ if (pid == vxi->vx_initpid) ++ return 1; ++ } ++ return pid; ++} ++ ++#define vx_info_rmap_pid(v, p) \ ++ __vx_info_rmap_pid((v), (p), __func__, __FILE__, __LINE__) ++#define vx_rmap_pid(p) vx_info_rmap_pid(current_vx_info(), p) ++#define vx_rmap_tgid(p) vx_rmap_pid(p) ++ ++static inline int __vx_info_rmap_pid(struct vx_info *vxi, int pid, ++ const char *func, const char *file, int line) ++{ ++ if (vx_info_flags(vxi, VXF_INFO_INIT, 0)) { ++ vxfprintk(VXD_CBIT(cvirt, 2), ++ "vx_rmap_tgid: %p/%llx: %d -> %d", ++ vxi, (long long)vxi->vx_flags, pid, ++ (pid == 1) ? vxi->vx_initpid : pid, ++ func, file, line); ++ if ((pid == 1) && vxi->vx_initpid) ++ return vxi->vx_initpid; ++ if (pid == vxi->vx_initpid) ++ return ~0U; ++ } ++ return pid; ++} ++ ++#endif +diff -NurpP --minimal linux-3.0.9/include/linux/vserver/sched.h linux-3.0.9-vs2.3.2.1/include/linux/vserver/sched.h +--- linux-3.0.9/include/linux/vserver/sched.h 1970-01-01 01:00:00.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/include/linux/vserver/sched.h 2011-06-10 22:11:24.000000000 +0200 +@@ -0,0 +1,23 @@ ++#ifndef _VX_SCHED_H ++#define _VX_SCHED_H ++ ++ ++#ifdef __KERNEL__ ++ ++struct timespec; ++ ++void vx_vsi_uptime(struct timespec *, struct timespec *); ++ ++ ++struct vx_info; ++ ++void vx_update_load(struct vx_info *); ++ ++ ++void vx_update_sched_param(struct _vx_sched *sched, ++ struct _vx_sched_pc *sched_pc); ++ ++#endif /* __KERNEL__ */ ++#else /* _VX_SCHED_H */ ++#warning duplicate inclusion ++#endif /* _VX_SCHED_H */ +diff -NurpP --minimal linux-3.0.9/include/linux/vserver/sched_cmd.h linux-3.0.9-vs2.3.2.1/include/linux/vserver/sched_cmd.h +--- linux-3.0.9/include/linux/vserver/sched_cmd.h 1970-01-01 01:00:00.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/include/linux/vserver/sched_cmd.h 2011-06-10 22:11:24.000000000 +0200 +@@ -0,0 +1,21 @@ ++#ifndef _VX_SCHED_CMD_H ++#define _VX_SCHED_CMD_H ++ ++ ++struct vcmd_prio_bias { ++ int32_t cpu_id; ++ int32_t prio_bias; ++}; ++ ++#define VCMD_set_prio_bias VC_CMD(SCHED, 4, 0) ++#define VCMD_get_prio_bias VC_CMD(SCHED, 5, 0) ++ ++#ifdef __KERNEL__ ++ ++#include <linux/compiler.h> ++ ++extern int vc_set_prio_bias(struct vx_info *, void __user *); ++extern int vc_get_prio_bias(struct vx_info *, void __user *); ++ ++#endif /* __KERNEL__ */ ++#endif /* _VX_SCHED_CMD_H */ +diff -NurpP --minimal linux-3.0.9/include/linux/vserver/sched_def.h linux-3.0.9-vs2.3.2.1/include/linux/vserver/sched_def.h +--- linux-3.0.9/include/linux/vserver/sched_def.h 1970-01-01 01:00:00.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/include/linux/vserver/sched_def.h 2011-06-10 22:11:24.000000000 +0200 +@@ -0,0 +1,38 @@ ++#ifndef _VX_SCHED_DEF_H ++#define _VX_SCHED_DEF_H ++ ++#include <linux/spinlock.h> ++#include <linux/jiffies.h> ++#include <linux/cpumask.h> ++#include <asm/atomic.h> ++#include <asm/param.h> ++ ++ ++/* context sub struct */ ++ ++struct _vx_sched { ++ int prio_bias; /* bias offset for priority */ ++ ++ cpumask_t update; /* CPUs which should update */ ++}; ++ ++struct _vx_sched_pc { ++ int prio_bias; /* bias offset for priority */ ++ ++ uint64_t user_ticks; /* token tick events */ ++ uint64_t sys_ticks; /* token tick events */ ++ uint64_t hold_ticks; /* token ticks paused */ ++}; ++ ++ ++#ifdef CONFIG_VSERVER_DEBUG ++ ++static inline void __dump_vx_sched(struct _vx_sched *sched) ++{ ++ printk("\t_vx_sched:\n"); ++ printk("\t priority = %4d\n", sched->prio_bias); ++} ++ ++#endif ++ ++#endif /* _VX_SCHED_DEF_H */ +diff -NurpP --minimal linux-3.0.9/include/linux/vserver/signal.h linux-3.0.9-vs2.3.2.1/include/linux/vserver/signal.h +--- linux-3.0.9/include/linux/vserver/signal.h 1970-01-01 01:00:00.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/include/linux/vserver/signal.h 2011-06-10 22:11:24.000000000 +0200 +@@ -0,0 +1,14 @@ ++#ifndef _VX_SIGNAL_H ++#define _VX_SIGNAL_H ++ ++ ++#ifdef __KERNEL__ ++ ++struct vx_info; ++ ++int vx_info_kill(struct vx_info *, int, int); ++ ++#endif /* __KERNEL__ */ ++#else /* _VX_SIGNAL_H */ ++#warning duplicate inclusion ++#endif /* _VX_SIGNAL_H */ +diff -NurpP --minimal linux-3.0.9/include/linux/vserver/signal_cmd.h linux-3.0.9-vs2.3.2.1/include/linux/vserver/signal_cmd.h +--- linux-3.0.9/include/linux/vserver/signal_cmd.h 1970-01-01 01:00:00.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/include/linux/vserver/signal_cmd.h 2011-06-10 22:11:24.000000000 +0200 +@@ -0,0 +1,43 @@ ++#ifndef _VX_SIGNAL_CMD_H ++#define _VX_SIGNAL_CMD_H ++ ++ ++/* signalling vserver commands */ ++ ++#define VCMD_ctx_kill VC_CMD(PROCTRL, 1, 0) ++#define VCMD_wait_exit VC_CMD(EVENT, 99, 0) ++ ++struct vcmd_ctx_kill_v0 { ++ int32_t pid; ++ int32_t sig; ++}; ++ ++struct vcmd_wait_exit_v0 { ++ int32_t reboot_cmd; ++ int32_t exit_code; ++}; ++ ++#ifdef __KERNEL__ ++ ++extern int vc_ctx_kill(struct vx_info *, void __user *); ++extern int vc_wait_exit(struct vx_info *, void __user *); ++ ++#endif /* __KERNEL__ */ ++ ++/* process alteration commands */ ++ ++#define VCMD_get_pflags VC_CMD(PROCALT, 5, 0) ++#define VCMD_set_pflags VC_CMD(PROCALT, 6, 0) ++ ++struct vcmd_pflags_v0 { ++ uint32_t flagword; ++ uint32_t mask; ++}; ++ ++#ifdef __KERNEL__ ++ ++extern int vc_get_pflags(uint32_t pid, void __user *); ++extern int vc_set_pflags(uint32_t pid, void __user *); ++ ++#endif /* __KERNEL__ */ ++#endif /* _VX_SIGNAL_CMD_H */ +diff -NurpP --minimal linux-3.0.9/include/linux/vserver/space.h linux-3.0.9-vs2.3.2.1/include/linux/vserver/space.h +--- linux-3.0.9/include/linux/vserver/space.h 1970-01-01 01:00:00.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/include/linux/vserver/space.h 2011-06-10 22:11:24.000000000 +0200 +@@ -0,0 +1,12 @@ ++#ifndef _VX_SPACE_H ++#define _VX_SPACE_H ++ ++#include <linux/types.h> ++ ++struct vx_info; ++ ++int vx_set_space(struct vx_info *vxi, unsigned long mask, unsigned index); ++ ++#else /* _VX_SPACE_H */ ++#warning duplicate inclusion ++#endif /* _VX_SPACE_H */ +diff -NurpP --minimal linux-3.0.9/include/linux/vserver/space_cmd.h linux-3.0.9-vs2.3.2.1/include/linux/vserver/space_cmd.h +--- linux-3.0.9/include/linux/vserver/space_cmd.h 1970-01-01 01:00:00.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/include/linux/vserver/space_cmd.h 2011-06-10 22:11:24.000000000 +0200 +@@ -0,0 +1,38 @@ ++#ifndef _VX_SPACE_CMD_H ++#define _VX_SPACE_CMD_H ++ ++ ++#define VCMD_enter_space_v0 VC_CMD(PROCALT, 1, 0) ++#define VCMD_enter_space_v1 VC_CMD(PROCALT, 1, 1) ++#define VCMD_enter_space VC_CMD(PROCALT, 1, 2) ++ ++#define VCMD_set_space_v0 VC_CMD(PROCALT, 3, 0) ++#define VCMD_set_space_v1 VC_CMD(PROCALT, 3, 1) ++#define VCMD_set_space VC_CMD(PROCALT, 3, 2) ++ ++#define VCMD_get_space_mask_v0 VC_CMD(PROCALT, 4, 0) ++ ++#define VCMD_get_space_mask VC_CMD(VSPACE, 0, 1) ++#define VCMD_get_space_default VC_CMD(VSPACE, 1, 0) ++ ++ ++struct vcmd_space_mask_v1 { ++ uint64_t mask; ++}; ++ ++struct vcmd_space_mask_v2 { ++ uint64_t mask; ++ uint32_t index; ++}; ++ ++ ++#ifdef __KERNEL__ ++ ++extern int vc_enter_space_v1(struct vx_info *, void __user *); ++extern int vc_set_space_v1(struct vx_info *, void __user *); ++extern int vc_enter_space(struct vx_info *, void __user *); ++extern int vc_set_space(struct vx_info *, void __user *); ++extern int vc_get_space_mask(void __user *, int); ++ ++#endif /* __KERNEL__ */ ++#endif /* _VX_SPACE_CMD_H */ +diff -NurpP --minimal linux-3.0.9/include/linux/vserver/switch.h linux-3.0.9-vs2.3.2.1/include/linux/vserver/switch.h +--- linux-3.0.9/include/linux/vserver/switch.h 1970-01-01 01:00:00.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/include/linux/vserver/switch.h 2011-06-10 22:11:24.000000000 +0200 +@@ -0,0 +1,98 @@ ++#ifndef _VX_SWITCH_H ++#define _VX_SWITCH_H ++ ++#include <linux/types.h> ++ ++ ++#define VC_CATEGORY(c) (((c) >> 24) & 0x3F) ++#define VC_COMMAND(c) (((c) >> 16) & 0xFF) ++#define VC_VERSION(c) ((c) & 0xFFF) ++ ++#define VC_CMD(c, i, v) ((((VC_CAT_ ## c) & 0x3F) << 24) \ ++ | (((i) & 0xFF) << 16) | ((v) & 0xFFF)) ++ ++/* ++ ++ Syscall Matrix V2.8 ++ ++ |VERSION|CREATE |MODIFY |MIGRATE|CONTROL|EXPERIM| |SPECIAL|SPECIAL| ++ |STATS |DESTROY|ALTER |CHANGE |LIMIT |TEST | | | | ++ |INFO |SETUP | |MOVE | | | | | | ++ -------+-------+-------+-------+-------+-------+-------+ +-------+-------+ ++ SYSTEM |VERSION|VSETUP |VHOST | | | | |DEVICE | | ++ HOST | 00| 01| 02| 03| 04| 05| | 06| 07| ++ -------+-------+-------+-------+-------+-------+-------+ +-------+-------+ ++ CPU | |VPROC |PROCALT|PROCMIG|PROCTRL| | |SCHED. | | ++ PROCESS| 08| 09| 10| 11| 12| 13| | 14| 15| ++ -------+-------+-------+-------+-------+-------+-------+ +-------+-------+ ++ MEMORY | | | | |MEMCTRL| | |SWAP | | ++ | 16| 17| 18| 19| 20| 21| | 22| 23| ++ -------+-------+-------+-------+-------+-------+-------+ +-------+-------+ ++ NETWORK| |VNET |NETALT |NETMIG |NETCTL | | |SERIAL | | ++ | 24| 25| 26| 27| 28| 29| | 30| 31| ++ -------+-------+-------+-------+-------+-------+-------+ +-------+-------+ ++ DISK | | | |TAGMIG |DLIMIT | | |INODE | | ++ VFS | 32| 33| 34| 35| 36| 37| | 38| 39| ++ -------+-------+-------+-------+-------+-------+-------+ +-------+-------+ ++ OTHER |VSTAT | | | | | | |VINFO | | ++ | 40| 41| 42| 43| 44| 45| | 46| 47| ++ =======+=======+=======+=======+=======+=======+=======+ +=======+=======+ ++ SPECIAL|EVENT | | | |FLAGS | | |VSPACE | | ++ | 48| 49| 50| 51| 52| 53| | 54| 55| ++ -------+-------+-------+-------+-------+-------+-------+ +-------+-------+ ++ SPECIAL|DEBUG | | | |RLIMIT |SYSCALL| | |COMPAT | ++ | 56| 57| 58| 59| 60|TEST 61| | 62| 63| ++ -------+-------+-------+-------+-------+-------+-------+ +-------+-------+ ++ ++*/ ++ ++#define VC_CAT_VERSION 0 ++ ++#define VC_CAT_VSETUP 1 ++#define VC_CAT_VHOST 2 ++ ++#define VC_CAT_DEVICE 6 ++ ++#define VC_CAT_VPROC 9 ++#define VC_CAT_PROCALT 10 ++#define VC_CAT_PROCMIG 11 ++#define VC_CAT_PROCTRL 12 ++ ++#define VC_CAT_SCHED 14 ++#define VC_CAT_MEMCTRL 20 ++ ++#define VC_CAT_VNET 25 ++#define VC_CAT_NETALT 26 ++#define VC_CAT_NETMIG 27 ++#define VC_CAT_NETCTRL 28 ++ ++#define VC_CAT_TAGMIG 35 ++#define VC_CAT_DLIMIT 36 ++#define VC_CAT_INODE 38 ++ ++#define VC_CAT_VSTAT 40 ++#define VC_CAT_VINFO 46 ++#define VC_CAT_EVENT 48 ++ ++#define VC_CAT_FLAGS 52 ++#define VC_CAT_VSPACE 54 ++#define VC_CAT_DEBUG 56 ++#define VC_CAT_RLIMIT 60 ++ ++#define VC_CAT_SYSTEST 61 ++#define VC_CAT_COMPAT 63 ++ ++/* query version */ ++ ++#define VCMD_get_version VC_CMD(VERSION, 0, 0) ++#define VCMD_get_vci VC_CMD(VERSION, 1, 0) ++ ++ ++#ifdef __KERNEL__ ++ ++#include <linux/errno.h> ++ ++#endif /* __KERNEL__ */ ++ ++#endif /* _VX_SWITCH_H */ ++ +diff -NurpP --minimal linux-3.0.9/include/linux/vserver/tag.h linux-3.0.9-vs2.3.2.1/include/linux/vserver/tag.h +--- linux-3.0.9/include/linux/vserver/tag.h 1970-01-01 01:00:00.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/include/linux/vserver/tag.h 2011-06-10 22:11:24.000000000 +0200 +@@ -0,0 +1,143 @@ ++#ifndef _DX_TAG_H ++#define _DX_TAG_H ++ ++#include <linux/types.h> ++ ++ ++#define DX_TAG(in) (IS_TAGGED(in)) ++ ++ ++#ifdef CONFIG_TAG_NFSD ++#define DX_TAG_NFSD 1 ++#else ++#define DX_TAG_NFSD 0 ++#endif ++ ++ ++#ifdef CONFIG_TAGGING_NONE ++ ++#define MAX_UID 0xFFFFFFFF ++#define MAX_GID 0xFFFFFFFF ++ ++#define INOTAG_TAG(cond, uid, gid, tag) (0) ++ ++#define TAGINO_UID(cond, uid, tag) (uid) ++#define TAGINO_GID(cond, gid, tag) (gid) ++ ++#endif ++ ++ ++#ifdef CONFIG_TAGGING_GID16 ++ ++#define MAX_UID 0xFFFFFFFF ++#define MAX_GID 0x0000FFFF ++ ++#define INOTAG_TAG(cond, uid, gid, tag) \ ++ ((cond) ? (((gid) >> 16) & 0xFFFF) : 0) ++ ++#define TAGINO_UID(cond, uid, tag) (uid) ++#define TAGINO_GID(cond, gid, tag) \ ++ ((cond) ? (((gid) & 0xFFFF) | ((tag) << 16)) : (gid)) ++ ++#endif ++ ++ ++#ifdef CONFIG_TAGGING_ID24 ++ ++#define MAX_UID 0x00FFFFFF ++#define MAX_GID 0x00FFFFFF ++ ++#define INOTAG_TAG(cond, uid, gid, tag) \ ++ ((cond) ? ((((uid) >> 16) & 0xFF00) | (((gid) >> 24) & 0xFF)) : 0) ++ ++#define TAGINO_UID(cond, uid, tag) \ ++ ((cond) ? (((uid) & 0xFFFFFF) | (((tag) & 0xFF00) << 16)) : (uid)) ++#define TAGINO_GID(cond, gid, tag) \ ++ ((cond) ? (((gid) & 0xFFFFFF) | (((tag) & 0x00FF) << 24)) : (gid)) ++ ++#endif ++ ++ ++#ifdef CONFIG_TAGGING_UID16 ++ ++#define MAX_UID 0x0000FFFF ++#define MAX_GID 0xFFFFFFFF ++ ++#define INOTAG_TAG(cond, uid, gid, tag) \ ++ ((cond) ? (((uid) >> 16) & 0xFFFF) : 0) ++ ++#define TAGINO_UID(cond, uid, tag) \ ++ ((cond) ? (((uid) & 0xFFFF) | ((tag) << 16)) : (uid)) ++#define TAGINO_GID(cond, gid, tag) (gid) ++ ++#endif ++ ++ ++#ifdef CONFIG_TAGGING_INTERN ++ ++#define MAX_UID 0xFFFFFFFF ++#define MAX_GID 0xFFFFFFFF ++ ++#define INOTAG_TAG(cond, uid, gid, tag) \ ++ ((cond) ? (tag) : 0) ++ ++#define TAGINO_UID(cond, uid, tag) (uid) ++#define TAGINO_GID(cond, gid, tag) (gid) ++ ++#endif ++ ++ ++#ifndef CONFIG_TAGGING_NONE ++#define dx_current_fstag(sb) \ ++ ((sb)->s_flags & MS_TAGGED ? dx_current_tag() : 0) ++#else ++#define dx_current_fstag(sb) (0) ++#endif ++ ++#ifndef CONFIG_TAGGING_INTERN ++#define TAGINO_TAG(cond, tag) (0) ++#else ++#define TAGINO_TAG(cond, tag) ((cond) ? (tag) : 0) ++#endif ++ ++#define INOTAG_UID(cond, uid, gid) \ ++ ((cond) ? ((uid) & MAX_UID) : (uid)) ++#define INOTAG_GID(cond, uid, gid) \ ++ ((cond) ? ((gid) & MAX_GID) : (gid)) ++ ++ ++static inline uid_t dx_map_uid(uid_t uid) ++{ ++ if ((uid > MAX_UID) && (uid != -1)) ++ uid = -2; ++ return (uid & MAX_UID); ++} ++ ++static inline gid_t dx_map_gid(gid_t gid) ++{ ++ if ((gid > MAX_GID) && (gid != -1)) ++ gid = -2; ++ return (gid & MAX_GID); ++} ++ ++struct peer_tag { ++ int32_t xid; ++ int32_t nid; ++}; ++ ++#define dx_notagcheck(sb) ((sb) && ((sb)->s_flags & MS_NOTAGCHECK)) ++ ++int dx_parse_tag(char *string, tag_t *tag, int remove, int *mnt_flags, ++ unsigned long *flags); ++ ++#ifdef CONFIG_PROPAGATE ++ ++void __dx_propagate_tag(struct nameidata *nd, struct inode *inode); ++ ++#define dx_propagate_tag(n, i) __dx_propagate_tag(n, i) ++ ++#else ++#define dx_propagate_tag(n, i) do { } while (0) ++#endif ++ ++#endif /* _DX_TAG_H */ +diff -NurpP --minimal linux-3.0.9/include/linux/vserver/tag_cmd.h linux-3.0.9-vs2.3.2.1/include/linux/vserver/tag_cmd.h +--- linux-3.0.9/include/linux/vserver/tag_cmd.h 1970-01-01 01:00:00.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/include/linux/vserver/tag_cmd.h 2011-06-10 22:11:24.000000000 +0200 +@@ -0,0 +1,22 @@ ++#ifndef _VX_TAG_CMD_H ++#define _VX_TAG_CMD_H ++ ++ ++/* vinfo commands */ ++ ++#define VCMD_task_tag VC_CMD(VINFO, 3, 0) ++ ++#ifdef __KERNEL__ ++extern int vc_task_tag(uint32_t); ++ ++#endif /* __KERNEL__ */ ++ ++/* context commands */ ++ ++#define VCMD_tag_migrate VC_CMD(TAGMIG, 1, 0) ++ ++#ifdef __KERNEL__ ++extern int vc_tag_migrate(uint32_t); ++ ++#endif /* __KERNEL__ */ ++#endif /* _VX_TAG_CMD_H */ +diff -NurpP --minimal linux-3.0.9/include/net/addrconf.h linux-3.0.9-vs2.3.2.1/include/net/addrconf.h +--- linux-3.0.9/include/net/addrconf.h 2011-07-22 11:18:11.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/include/net/addrconf.h 2011-06-10 22:11:24.000000000 +0200 +@@ -80,7 +80,8 @@ extern int ipv6_dev_get_saddr(struct n + struct net_device *dev, + const struct in6_addr *daddr, + unsigned int srcprefs, +- struct in6_addr *saddr); ++ struct in6_addr *saddr, ++ struct nx_info *nxi); + extern int ipv6_get_lladdr(struct net_device *dev, + struct in6_addr *addr, + unsigned char banned_flags); +diff -NurpP --minimal linux-3.0.9/include/net/af_unix.h linux-3.0.9-vs2.3.2.1/include/net/af_unix.h +--- linux-3.0.9/include/net/af_unix.h 2011-07-22 11:18:11.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/include/net/af_unix.h 2011-06-10 22:11:24.000000000 +0200 +@@ -4,6 +4,7 @@ + #include <linux/socket.h> + #include <linux/un.h> + #include <linux/mutex.h> ++#include <linux/vs_base.h> + #include <net/sock.h> + + extern void unix_inflight(struct file *fp); +diff -NurpP --minimal linux-3.0.9/include/net/inet_timewait_sock.h linux-3.0.9-vs2.3.2.1/include/net/inet_timewait_sock.h +--- linux-3.0.9/include/net/inet_timewait_sock.h 2011-03-15 18:07:40.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/include/net/inet_timewait_sock.h 2011-06-10 22:11:24.000000000 +0200 +@@ -113,6 +113,10 @@ struct inet_timewait_sock { + #define tw_net __tw_common.skc_net + #define tw_daddr __tw_common.skc_daddr + #define tw_rcv_saddr __tw_common.skc_rcv_saddr ++#define tw_xid __tw_common.skc_xid ++#define tw_vx_info __tw_common.skc_vx_info ++#define tw_nid __tw_common.skc_nid ++#define tw_nx_info __tw_common.skc_nx_info + int tw_timeout; + volatile unsigned char tw_substate; + unsigned char tw_rcv_wscale; +diff -NurpP --minimal linux-3.0.9/include/net/ip6_route.h linux-3.0.9-vs2.3.2.1/include/net/ip6_route.h +--- linux-3.0.9/include/net/ip6_route.h 2011-07-22 11:18:11.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/include/net/ip6_route.h 2011-06-16 14:16:51.000000000 +0200 +@@ -86,7 +86,8 @@ extern int ip6_route_get_saddr(struct + struct rt6_info *rt, + const struct in6_addr *daddr, + unsigned int prefs, +- struct in6_addr *saddr); ++ struct in6_addr *saddr, ++ struct nx_info *nxi); + + extern struct rt6_info *rt6_lookup(struct net *net, + const struct in6_addr *daddr, +diff -NurpP --minimal linux-3.0.9/include/net/route.h linux-3.0.9-vs2.3.2.1/include/net/route.h +--- linux-3.0.9/include/net/route.h 2011-07-22 11:18:11.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/include/net/route.h 2011-07-27 19:42:59.000000000 +0200 +@@ -202,6 +202,9 @@ static inline void ip_rt_put(struct rtab + dst_release(&rt->dst); + } + ++#include <linux/vs_base.h> ++#include <linux/vs_inet.h> ++ + #define IPTOS_RT_MASK (IPTOS_TOS_MASK & ~3) + + extern const __u8 ip_tos2prio[16]; +@@ -253,6 +256,9 @@ static inline void ip_route_connect_init + protocol, flow_flags, dst, src, dport, sport); + } + ++extern struct rtable *ip_v4_find_src(struct net *net, struct nx_info *, ++ struct flowi4 *); ++ + static inline struct rtable *ip_route_connect(struct flowi4 *fl4, + __be32 dst, __be32 src, u32 tos, + int oif, u8 protocol, +@@ -261,11 +267,25 @@ static inline struct rtable *ip_route_co + { + struct net *net = sock_net(sk); + struct rtable *rt; ++ struct nx_info *nx_info = current_nx_info(); + + ip_route_connect_init(fl4, dst, src, tos, oif, protocol, + sport, dport, sk, can_sleep); + +- if (!dst || !src) { ++ if (sk) ++ nx_info = sk->sk_nx_info; ++ ++ vxdprintk(VXD_CBIT(net, 4), ++ "ip_route_connect(%p) %p,%p;%lx", ++ sk, nx_info, sk->sk_socket, ++ (sk->sk_socket?sk->sk_socket->flags:0)); ++ ++ rt = ip_v4_find_src(net, nx_info, fl4); ++ if (IS_ERR(rt)) ++ return rt; ++ ip_rt_put(rt); ++ ++ if (!fl4->daddr || !fl4->saddr) { + rt = __ip_route_output_key(net, fl4); + if (IS_ERR(rt)) + return rt; +diff -NurpP --minimal linux-3.0.9/include/net/sock.h linux-3.0.9-vs2.3.2.1/include/net/sock.h +--- linux-3.0.9/include/net/sock.h 2011-07-22 11:18:12.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/include/net/sock.h 2011-07-01 11:35:35.000000000 +0200 +@@ -149,6 +149,10 @@ struct sock_common { + #ifdef CONFIG_NET_NS + struct net *skc_net; + #endif ++ xid_t skc_xid; ++ struct vx_info *skc_vx_info; ++ nid_t skc_nid; ++ struct nx_info *skc_nx_info; + /* + * fields between dontcopy_begin/dontcopy_end + * are not copied in sock_copy() +@@ -256,6 +260,10 @@ struct sock { + #define sk_bind_node __sk_common.skc_bind_node + #define sk_prot __sk_common.skc_prot + #define sk_net __sk_common.skc_net ++#define sk_xid __sk_common.skc_xid ++#define sk_vx_info __sk_common.skc_vx_info ++#define sk_nid __sk_common.skc_nid ++#define sk_nx_info __sk_common.skc_nx_info + socket_lock_t sk_lock; + struct sk_buff_head sk_receive_queue; + /* +diff -NurpP --minimal linux-3.0.9/init/Kconfig linux-3.0.9-vs2.3.2.1/init/Kconfig +--- linux-3.0.9/init/Kconfig 2011-07-22 11:18:12.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/init/Kconfig 2011-08-08 18:08:57.000000000 +0200 +@@ -574,6 +574,7 @@ config HAVE_UNSTABLE_SCHED_CLOCK + menuconfig CGROUPS + boolean "Control Group support" + depends on EVENTFD ++ default y + help + This option adds support for grouping sets of processes together, for + use with process control subsystems such as Cpusets, CFS, memory +@@ -790,6 +791,7 @@ config IPC_NS + config USER_NS + bool "User namespace (EXPERIMENTAL)" + depends on EXPERIMENTAL ++ depends on VSERVER_DISABLED + default y + help + This allows containers, i.e. vservers, to use user namespaces +diff -NurpP --minimal linux-3.0.9/init/main.c linux-3.0.9-vs2.3.2.1/init/main.c +--- linux-3.0.9/init/main.c 2011-07-22 11:18:12.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/init/main.c 2011-06-22 12:39:15.000000000 +0200 +@@ -68,6 +68,7 @@ + #include <linux/shmem_fs.h> + #include <linux/slab.h> + #include <linux/perf_event.h> ++#include <linux/vserver/percpu.h> + + #include <asm/io.h> + #include <asm/bugs.h> +diff -NurpP --minimal linux-3.0.9/ipc/mqueue.c linux-3.0.9-vs2.3.2.1/ipc/mqueue.c +--- linux-3.0.9/ipc/mqueue.c 2011-11-15 16:40:47.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/ipc/mqueue.c 2011-10-18 14:03:46.000000000 +0200 +@@ -33,6 +33,8 @@ + #include <linux/pid.h> + #include <linux/ipc_namespace.h> + #include <linux/slab.h> ++#include <linux/vs_context.h> ++#include <linux/vs_limit.h> + + #include <net/sock.h> + #include "util.h" +@@ -66,6 +68,7 @@ struct mqueue_inode_info { + struct sigevent notify; + struct pid* notify_owner; + struct user_struct *user; /* user who created, for accounting */ ++ struct vx_info *vxi; + struct sock *notify_sock; + struct sk_buff *notify_cookie; + +@@ -128,6 +131,7 @@ static struct inode *mqueue_get_inode(st + if (S_ISREG(mode)) { + struct mqueue_inode_info *info; + struct task_struct *p = current; ++ struct vx_info *vxi = p->vx_info; + unsigned long mq_bytes, mq_msg_tblsz; + + inode->i_fop = &mqueue_file_operations; +@@ -141,6 +145,7 @@ static struct inode *mqueue_get_inode(st + info->notify_owner = NULL; + info->qsize = 0; + info->user = NULL; /* set when all is ok */ ++ info->vxi = NULL; + memset(&info->attr, 0, sizeof(info->attr)); + info->attr.mq_maxmsg = ipc_ns->mq_msg_max; + info->attr.mq_msgsize = ipc_ns->mq_msgsize_max; +@@ -158,17 +163,20 @@ static struct inode *mqueue_get_inode(st + + spin_lock(&mq_lock); + if (u->mq_bytes + mq_bytes < u->mq_bytes || +- u->mq_bytes + mq_bytes > task_rlimit(p, RLIMIT_MSGQUEUE)) { ++ u->mq_bytes + mq_bytes > task_rlimit(p, RLIMIT_MSGQUEUE) || ++ !vx_ipcmsg_avail(vxi, mq_bytes)) { + spin_unlock(&mq_lock); + /* mqueue_evict_inode() releases info->messages */ + ret = -EMFILE; + goto out_inode; + } + u->mq_bytes += mq_bytes; ++ vx_ipcmsg_add(vxi, u, mq_bytes); + spin_unlock(&mq_lock); + + /* all is ok */ + info->user = get_uid(u); ++ info->vxi = get_vx_info(vxi); + } else if (S_ISDIR(mode)) { + inc_nlink(inode); + /* Some things misbehave if size == 0 on a directory */ +@@ -278,8 +286,11 @@ static void mqueue_evict_inode(struct in + + info->attr.mq_msgsize); + user = info->user; + if (user) { ++ struct vx_info *vxi = info->vxi; ++ + spin_lock(&mq_lock); + user->mq_bytes -= mq_bytes; ++ vx_ipcmsg_sub(vxi, user, mq_bytes); + /* + * get_ns_from_inode() ensures that the + * (ipc_ns = sb->s_fs_info) is either a valid ipc_ns +@@ -289,6 +300,7 @@ static void mqueue_evict_inode(struct in + if (ipc_ns) + ipc_ns->mq_queues_count--; + spin_unlock(&mq_lock); ++ put_vx_info(vxi); + free_uid(user); + } + if (ipc_ns) +diff -NurpP --minimal linux-3.0.9/ipc/msg.c linux-3.0.9-vs2.3.2.1/ipc/msg.c +--- linux-3.0.9/ipc/msg.c 2011-05-22 16:17:59.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/ipc/msg.c 2011-06-10 22:11:24.000000000 +0200 +@@ -37,6 +37,7 @@ + #include <linux/rwsem.h> + #include <linux/nsproxy.h> + #include <linux/ipc_namespace.h> ++#include <linux/vs_base.h> + + #include <asm/current.h> + #include <asm/uaccess.h> +@@ -190,6 +191,7 @@ static int newque(struct ipc_namespace * + + msq->q_perm.mode = msgflg & S_IRWXUGO; + msq->q_perm.key = key; ++ msq->q_perm.xid = vx_current_xid(); + + msq->q_perm.security = NULL; + retval = security_msg_queue_alloc(msq); +diff -NurpP --minimal linux-3.0.9/ipc/namespace.c linux-3.0.9-vs2.3.2.1/ipc/namespace.c +--- linux-3.0.9/ipc/namespace.c 2011-07-22 11:18:12.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/ipc/namespace.c 2011-06-13 14:09:44.000000000 +0200 +@@ -13,11 +13,12 @@ + #include <linux/mount.h> + #include <linux/user_namespace.h> + #include <linux/proc_fs.h> ++#include <linux/vs_base.h> ++#include <linux/vserver/global.h> + + #include "util.h" + +-static struct ipc_namespace *create_ipc_ns(struct task_struct *tsk, +- struct ipc_namespace *old_ns) ++static struct ipc_namespace *create_ipc_ns(struct user_namespace *user_ns) + { + struct ipc_namespace *ns; + int err; +@@ -46,19 +47,18 @@ static struct ipc_namespace *create_ipc_ + ipcns_notify(IPCNS_CREATED); + register_ipcns_notifier(ns); + +- ns->user_ns = get_user_ns(task_cred_xxx(tsk, user)->user_ns); ++ ns->user_ns = get_user_ns(user_ns); + + return ns; + } + + struct ipc_namespace *copy_ipcs(unsigned long flags, +- struct task_struct *tsk) ++ struct ipc_namespace *old_ns, ++ struct user_namespace *user_ns) + { +- struct ipc_namespace *ns = tsk->nsproxy->ipc_ns; +- + if (!(flags & CLONE_NEWIPC)) +- return get_ipc_ns(ns); +- return create_ipc_ns(tsk, ns); ++ return get_ipc_ns(old_ns); ++ return create_ipc_ns(user_ns); + } + + /* +diff -NurpP --minimal linux-3.0.9/ipc/sem.c linux-3.0.9-vs2.3.2.1/ipc/sem.c +--- linux-3.0.9/ipc/sem.c 2011-11-15 16:40:47.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/ipc/sem.c 2011-08-08 23:04:47.000000000 +0200 +@@ -86,6 +86,8 @@ + #include <linux/rwsem.h> + #include <linux/nsproxy.h> + #include <linux/ipc_namespace.h> ++#include <linux/vs_base.h> ++#include <linux/vs_limit.h> + + #include <asm/uaccess.h> + #include "util.h" +@@ -260,6 +262,7 @@ static int newary(struct ipc_namespace * + + sma->sem_perm.mode = (semflg & S_IRWXUGO); + sma->sem_perm.key = key; ++ sma->sem_perm.xid = vx_current_xid(); + + sma->sem_perm.security = NULL; + retval = security_sem_alloc(sma); +@@ -275,6 +278,9 @@ static int newary(struct ipc_namespace * + return id; + } + ns->used_sems += nsems; ++ /* FIXME: obsoleted? */ ++ vx_semary_inc(sma); ++ vx_nsems_add(sma, nsems); + + sma->sem_base = (struct sem *) &sma[1]; + +@@ -730,6 +736,9 @@ static void freeary(struct ipc_namespace + + wake_up_sem_queue_do(&tasks); + ns->used_sems -= sma->sem_nsems; ++ /* FIXME: obsoleted? */ ++ vx_nsems_sub(sma, sma->sem_nsems); ++ vx_semary_dec(sma); + security_sem_free(sma); + ipc_rcu_putref(sma); + } +diff -NurpP --minimal linux-3.0.9/ipc/shm.c linux-3.0.9-vs2.3.2.1/ipc/shm.c +--- linux-3.0.9/ipc/shm.c 2011-07-22 11:18:12.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/ipc/shm.c 2011-06-10 22:11:24.000000000 +0200 +@@ -39,6 +39,8 @@ + #include <linux/nsproxy.h> + #include <linux/mount.h> + #include <linux/ipc_namespace.h> ++#include <linux/vs_context.h> ++#include <linux/vs_limit.h> + + #include <asm/uaccess.h> + +@@ -173,7 +175,12 @@ static void shm_open(struct vm_area_stru + */ + static void shm_destroy(struct ipc_namespace *ns, struct shmid_kernel *shp) + { +- ns->shm_tot -= (shp->shm_segsz + PAGE_SIZE - 1) >> PAGE_SHIFT; ++ struct vx_info *vxi = lookup_vx_info(shp->shm_perm.xid); ++ int numpages = (shp->shm_segsz + PAGE_SIZE - 1) >> PAGE_SHIFT; ++ ++ vx_ipcshm_sub(vxi, shp, numpages); ++ ns->shm_tot -= numpages; ++ + shm_rmid(ns, shp); + shm_unlock(shp); + if (!is_file_hugepages(shp->shm_file)) +@@ -183,6 +190,7 @@ static void shm_destroy(struct ipc_names + shp->mlock_user); + fput (shp->shm_file); + security_shm_free(shp); ++ put_vx_info(vxi); + ipc_rcu_putref(shp); + } + +@@ -355,11 +363,15 @@ static int newseg(struct ipc_namespace * + if (ns->shm_tot + numpages > ns->shm_ctlall) + return -ENOSPC; + ++ if (!vx_ipcshm_avail(current_vx_info(), numpages)) ++ return -ENOSPC; ++ + shp = ipc_rcu_alloc(sizeof(*shp)); + if (!shp) + return -ENOMEM; + + shp->shm_perm.key = key; ++ shp->shm_perm.xid = vx_current_xid(); + shp->shm_perm.mode = (shmflg & S_IRWXUGO); + shp->mlock_user = NULL; + +@@ -413,6 +425,7 @@ static int newseg(struct ipc_namespace * + ns->shm_tot += numpages; + error = shp->shm_perm.id; + shm_unlock(shp); ++ vx_ipcshm_add(current_vx_info(), key, numpages); + return error; + + no_id: +diff -NurpP --minimal linux-3.0.9/kernel/Makefile linux-3.0.9-vs2.3.2.1/kernel/Makefile +--- linux-3.0.9/kernel/Makefile 2011-07-22 11:18:12.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/kernel/Makefile 2011-06-10 22:11:24.000000000 +0200 +@@ -24,6 +24,7 @@ CFLAGS_REMOVE_sched_clock.o = -pg + CFLAGS_REMOVE_irq_work.o = -pg + endif + ++obj-y += vserver/ + obj-$(CONFIG_FREEZER) += freezer.o + obj-$(CONFIG_PROFILING) += profile.o + obj-$(CONFIG_SYSCTL_SYSCALL_CHECK) += sysctl_check.o +diff -NurpP --minimal linux-3.0.9/kernel/capability.c linux-3.0.9-vs2.3.2.1/kernel/capability.c +--- linux-3.0.9/kernel/capability.c 2011-07-22 11:18:12.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/kernel/capability.c 2011-10-27 13:59:20.000000000 +0200 +@@ -15,6 +15,7 @@ + #include <linux/syscalls.h> + #include <linux/pid_namespace.h> + #include <linux/user_namespace.h> ++#include <linux/vs_context.h> + #include <asm/uaccess.h> + + /* +@@ -116,6 +117,7 @@ static int cap_validate_magic(cap_user_h + return 0; + } + ++ + /* + * The only thing that can change the capabilities of the current + * process is the current process. As such, we can't be in this code +@@ -340,6 +342,8 @@ bool has_capability_noaudit(struct task_ + return (ret == 0); + } + ++#include <linux/vserver/base.h> ++ + /** + * capable - Determine if the current task has a superior capability in effect + * @cap: The capability to be tested for +diff -NurpP --minimal linux-3.0.9/kernel/compat.c linux-3.0.9-vs2.3.2.1/kernel/compat.c +--- linux-3.0.9/kernel/compat.c 2011-07-22 11:18:12.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/kernel/compat.c 2011-06-10 22:11:24.000000000 +0200 +@@ -970,7 +970,7 @@ asmlinkage long compat_sys_stime(compat_ + if (err) + return err; + +- do_settimeofday(&tv); ++ vx_settimeofday(&tv); + return 0; + } + +diff -NurpP --minimal linux-3.0.9/kernel/cred.c linux-3.0.9-vs2.3.2.1/kernel/cred.c +--- linux-3.0.9/kernel/cred.c 2011-07-22 11:18:12.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/kernel/cred.c 2011-06-10 22:11:24.000000000 +0200 +@@ -61,31 +61,6 @@ struct cred init_cred = { + #endif + }; + +-static inline void set_cred_subscribers(struct cred *cred, int n) +-{ +-#ifdef CONFIG_DEBUG_CREDENTIALS +- atomic_set(&cred->subscribers, n); +-#endif +-} +- +-static inline int read_cred_subscribers(const struct cred *cred) +-{ +-#ifdef CONFIG_DEBUG_CREDENTIALS +- return atomic_read(&cred->subscribers); +-#else +- return 0; +-#endif +-} +- +-static inline void alter_cred_subscribers(const struct cred *_cred, int n) +-{ +-#ifdef CONFIG_DEBUG_CREDENTIALS +- struct cred *cred = (struct cred *) _cred; +- +- atomic_add(n, &cred->subscribers); +-#endif +-} +- + /* + * Dispose of the shared task group credentials + */ +@@ -281,21 +256,16 @@ error: + * + * Call commit_creds() or abort_creds() to clean up. + */ +-struct cred *prepare_creds(void) ++struct cred *__prepare_creds(const struct cred *old) + { +- struct task_struct *task = current; +- const struct cred *old; + struct cred *new; + +- validate_process_creds(); +- + new = kmem_cache_alloc(cred_jar, GFP_KERNEL); + if (!new) + return NULL; + + kdebug("prepare_creds() alloc %p", new); + +- old = task->cred; + memcpy(new, old, sizeof(struct cred)); + + atomic_set(&new->usage, 1); +@@ -322,6 +292,13 @@ error: + abort_creds(new); + return NULL; + } ++ ++struct cred *prepare_creds(void) ++{ ++ validate_process_creds(); ++ ++ return __prepare_creds(current->cred); ++} + EXPORT_SYMBOL(prepare_creds); + + /* +diff -NurpP --minimal linux-3.0.9/kernel/exit.c linux-3.0.9-vs2.3.2.1/kernel/exit.c +--- linux-3.0.9/kernel/exit.c 2011-07-22 11:18:12.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/kernel/exit.c 2011-06-22 12:39:15.000000000 +0200 +@@ -48,6 +48,10 @@ + #include <linux/fs_struct.h> + #include <linux/init_task.h> + #include <linux/perf_event.h> ++#include <linux/vs_limit.h> ++#include <linux/vs_context.h> ++#include <linux/vs_network.h> ++#include <linux/vs_pid.h> + #include <trace/events/sched.h> + #include <linux/hw_breakpoint.h> + #include <linux/oom.h> +@@ -494,9 +498,11 @@ static void close_files(struct files_str + filp_close(file, files); + cond_resched(); + } ++ vx_openfd_dec(i); + } + i++; + set >>= 1; ++ cond_resched(); + } + } + } +@@ -1047,11 +1053,16 @@ NORET_TYPE void do_exit(long code) + + validate_creds_for_do_exit(tsk); + ++ /* needs to stay after exit_notify() */ ++ exit_vx_info(tsk, code); ++ exit_nx_info(tsk); ++ + preempt_disable(); + exit_rcu(); + /* causes final put_task_struct in finish_task_switch(). */ + tsk->state = TASK_DEAD; + schedule(); ++ printk("bad task: %p [%lx]\n", current, current->state); + BUG(); + /* Avoid "noreturn function does return". */ + for (;;) +diff -NurpP --minimal linux-3.0.9/kernel/fork.c linux-3.0.9-vs2.3.2.1/kernel/fork.c +--- linux-3.0.9/kernel/fork.c 2011-07-22 11:18:12.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/kernel/fork.c 2011-06-10 22:21:02.000000000 +0200 +@@ -67,6 +67,10 @@ + #include <linux/user-return-notifier.h> + #include <linux/oom.h> + #include <linux/khugepaged.h> ++#include <linux/vs_context.h> ++#include <linux/vs_network.h> ++#include <linux/vs_limit.h> ++#include <linux/vs_memory.h> + + #include <asm/pgtable.h> + #include <asm/pgalloc.h> +@@ -167,6 +171,8 @@ void free_task(struct task_struct *tsk) + account_kernel_stack(tsk->stack, -1); + free_thread_info(tsk->stack); + rt_mutex_debug_task_free(tsk); ++ clr_vx_info(&tsk->vx_info); ++ clr_nx_info(&tsk->nx_info); + ftrace_graph_exit_task(tsk); + free_task_struct(tsk); + } +@@ -505,6 +511,7 @@ static struct mm_struct * mm_init(struct + if (likely(!mm_alloc_pgd(mm))) { + mm->def_flags = 0; + mmu_notifier_mm_init(mm); ++ set_vx_info(&mm->mm_vx_info, p->vx_info); + return mm; + } + +@@ -542,6 +549,7 @@ void __mmdrop(struct mm_struct *mm) + #ifdef CONFIG_TRANSPARENT_HUGEPAGE + VM_BUG_ON(mm->pmd_huge_pte); + #endif ++ clr_vx_info(&mm->mm_vx_info); + free_mm(mm); + } + EXPORT_SYMBOL_GPL(__mmdrop); +@@ -729,6 +737,7 @@ struct mm_struct *dup_mm(struct task_str + goto fail_nomem; + + memcpy(mm, oldmm, sizeof(*mm)); ++ mm->mm_vx_info = NULL; + mm_init_cpumask(mm); + + /* Initializing for Swap token stuff */ +@@ -772,6 +781,7 @@ fail_nocontext: + * If init_new_context() failed, we cannot use mmput() to free the mm + * because it calls destroy_context() + */ ++ clr_vx_info(&mm->mm_vx_info); + mm_free_pgd(mm); + free_mm(mm); + return NULL; +@@ -1057,6 +1067,8 @@ static struct task_struct *copy_process( + int retval; + struct task_struct *p; + int cgroup_callbacks_done = 0; ++ struct vx_info *vxi; ++ struct nx_info *nxi; + + if ((clone_flags & (CLONE_NEWNS|CLONE_FS)) == (CLONE_NEWNS|CLONE_FS)) + return ERR_PTR(-EINVAL); +@@ -1103,7 +1115,12 @@ static struct task_struct *copy_process( + DEBUG_LOCKS_WARN_ON(!p->hardirqs_enabled); + DEBUG_LOCKS_WARN_ON(!p->softirqs_enabled); + #endif ++ init_vx_info(&p->vx_info, current_vx_info()); ++ init_nx_info(&p->nx_info, current_nx_info()); ++ + retval = -EAGAIN; ++ if (!vx_nproc_avail(1)) ++ goto bad_fork_free; + if (atomic_read(&p->real_cred->user->processes) >= + task_rlimit(p, RLIMIT_NPROC)) { + if (!capable(CAP_SYS_ADMIN) && !capable(CAP_SYS_RESOURCE) && +@@ -1360,6 +1377,18 @@ static struct task_struct *copy_process( + + total_forks++; + spin_unlock(¤t->sighand->siglock); ++ ++ /* p is copy of current */ ++ vxi = p->vx_info; ++ if (vxi) { ++ claim_vx_info(vxi, p); ++ atomic_inc(&vxi->cvirt.nr_threads); ++ atomic_inc(&vxi->cvirt.total_forks); ++ vx_nproc_inc(p); ++ } ++ nxi = p->nx_info; ++ if (nxi) ++ claim_nx_info(nxi, p); + write_unlock_irq(&tasklist_lock); + proc_fork_connector(p); + cgroup_post_fork(p); +diff -NurpP --minimal linux-3.0.9/kernel/kthread.c linux-3.0.9-vs2.3.2.1/kernel/kthread.c +--- linux-3.0.9/kernel/kthread.c 2011-07-22 11:18:12.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/kernel/kthread.c 2011-06-10 22:11:24.000000000 +0200 +@@ -16,6 +16,7 @@ + #include <linux/mutex.h> + #include <linux/slab.h> + #include <linux/freezer.h> ++#include <linux/vs_pid.h> + #include <trace/events/sched.h> + + static DEFINE_SPINLOCK(kthread_create_lock); +diff -NurpP --minimal linux-3.0.9/kernel/nsproxy.c linux-3.0.9-vs2.3.2.1/kernel/nsproxy.c +--- linux-3.0.9/kernel/nsproxy.c 2011-07-22 11:18:12.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/kernel/nsproxy.c 2011-06-13 14:09:44.000000000 +0200 +@@ -20,6 +20,8 @@ + #include <linux/mnt_namespace.h> + #include <linux/utsname.h> + #include <linux/pid_namespace.h> ++#include <linux/vserver/global.h> ++#include <linux/vserver/debug.h> + #include <net/net_namespace.h> + #include <linux/ipc_namespace.h> + #include <linux/proc_fs.h> +@@ -46,8 +48,11 @@ static inline struct nsproxy *create_nsp + struct nsproxy *nsproxy; + + nsproxy = kmem_cache_alloc(nsproxy_cachep, GFP_KERNEL); +- if (nsproxy) ++ if (nsproxy) { + atomic_set(&nsproxy->count, 1); ++ atomic_inc(&vs_global_nsproxy); ++ } ++ vxdprintk(VXD_CBIT(space, 2), "create_nsproxy = %p[1]", nsproxy); + return nsproxy; + } + +@@ -56,8 +61,11 @@ static inline struct nsproxy *create_nsp + * Return the newly created nsproxy. Do not attach this to the task, + * leave it to the caller to do proper locking and attach it to task. + */ +-static struct nsproxy *create_new_namespaces(unsigned long flags, +- struct task_struct *tsk, struct fs_struct *new_fs) ++static struct nsproxy *unshare_namespaces(unsigned long flags, ++ struct nsproxy *orig, ++ struct fs_struct *new_fs, ++ struct user_namespace *new_user, ++ struct pid_namespace *new_pid) + { + struct nsproxy *new_nsp; + int err; +@@ -66,31 +74,31 @@ static struct nsproxy *create_new_namesp + if (!new_nsp) + return ERR_PTR(-ENOMEM); + +- new_nsp->mnt_ns = copy_mnt_ns(flags, tsk->nsproxy->mnt_ns, new_fs); ++ new_nsp->mnt_ns = copy_mnt_ns(flags, orig->mnt_ns, new_fs); + if (IS_ERR(new_nsp->mnt_ns)) { + err = PTR_ERR(new_nsp->mnt_ns); + goto out_ns; + } + +- new_nsp->uts_ns = copy_utsname(flags, tsk); ++ new_nsp->uts_ns = copy_utsname(flags, orig->uts_ns, new_user); + if (IS_ERR(new_nsp->uts_ns)) { + err = PTR_ERR(new_nsp->uts_ns); + goto out_uts; + } + +- new_nsp->ipc_ns = copy_ipcs(flags, tsk); ++ new_nsp->ipc_ns = copy_ipcs(flags, orig->ipc_ns, new_user); + if (IS_ERR(new_nsp->ipc_ns)) { + err = PTR_ERR(new_nsp->ipc_ns); + goto out_ipc; + } + +- new_nsp->pid_ns = copy_pid_ns(flags, task_active_pid_ns(tsk)); ++ new_nsp->pid_ns = copy_pid_ns(flags, new_pid); + if (IS_ERR(new_nsp->pid_ns)) { + err = PTR_ERR(new_nsp->pid_ns); + goto out_pid; + } + +- new_nsp->net_ns = copy_net_ns(flags, tsk->nsproxy->net_ns); ++ new_nsp->net_ns = copy_net_ns(flags, orig->net_ns); + if (IS_ERR(new_nsp->net_ns)) { + err = PTR_ERR(new_nsp->net_ns); + goto out_net; +@@ -115,6 +123,40 @@ out_ns: + return ERR_PTR(err); + } + ++static struct nsproxy *create_new_namespaces(unsigned long flags, ++ struct task_struct *tsk, struct fs_struct *new_fs) ++{ ++ return unshare_namespaces(flags, tsk->nsproxy, ++ new_fs, task_cred_xxx(tsk, user)->user_ns, ++ task_active_pid_ns(tsk)); ++} ++ ++/* ++ * copies the nsproxy, setting refcount to 1, and grabbing a ++ * reference to all contained namespaces. ++ */ ++struct nsproxy *copy_nsproxy(struct nsproxy *orig) ++{ ++ struct nsproxy *ns = create_nsproxy(); ++ ++ if (ns) { ++ memcpy(ns, orig, sizeof(struct nsproxy)); ++ atomic_set(&ns->count, 1); ++ ++ if (ns->mnt_ns) ++ get_mnt_ns(ns->mnt_ns); ++ if (ns->uts_ns) ++ get_uts_ns(ns->uts_ns); ++ if (ns->ipc_ns) ++ get_ipc_ns(ns->ipc_ns); ++ if (ns->pid_ns) ++ get_pid_ns(ns->pid_ns); ++ if (ns->net_ns) ++ get_net(ns->net_ns); ++ } ++ return ns; ++} ++ + /* + * called from clone. This now handles copy for nsproxy and all + * namespaces therein. +@@ -122,9 +164,12 @@ out_ns: + int copy_namespaces(unsigned long flags, struct task_struct *tsk) + { + struct nsproxy *old_ns = tsk->nsproxy; +- struct nsproxy *new_ns; ++ struct nsproxy *new_ns = NULL; + int err = 0; + ++ vxdprintk(VXD_CBIT(space, 7), "copy_namespaces(0x%08lx,%p[%p])", ++ flags, tsk, old_ns); ++ + if (!old_ns) + return 0; + +@@ -134,7 +179,7 @@ int copy_namespaces(unsigned long flags, + CLONE_NEWPID | CLONE_NEWNET))) + return 0; + +- if (!capable(CAP_SYS_ADMIN)) { ++ if (!vx_can_unshare(CAP_SYS_ADMIN, flags)) { + err = -EPERM; + goto out; + } +@@ -161,6 +206,9 @@ int copy_namespaces(unsigned long flags, + + out: + put_nsproxy(old_ns); ++ vxdprintk(VXD_CBIT(space, 3), ++ "copy_namespaces(0x%08lx,%p[%p]) = %d [%p]", ++ flags, tsk, old_ns, err, new_ns); + return err; + } + +@@ -174,7 +222,9 @@ void free_nsproxy(struct nsproxy *ns) + put_ipc_ns(ns->ipc_ns); + if (ns->pid_ns) + put_pid_ns(ns->pid_ns); +- put_net(ns->net_ns); ++ if (ns->net_ns) ++ put_net(ns->net_ns); ++ atomic_dec(&vs_global_nsproxy); + kmem_cache_free(nsproxy_cachep, ns); + } + +@@ -187,11 +237,15 @@ int unshare_nsproxy_namespaces(unsigned + { + int err = 0; + ++ vxdprintk(VXD_CBIT(space, 4), ++ "unshare_nsproxy_namespaces(0x%08lx,[%p])", ++ unshare_flags, current->nsproxy); ++ + if (!(unshare_flags & (CLONE_NEWNS | CLONE_NEWUTS | CLONE_NEWIPC | + CLONE_NEWNET))) + return 0; + +- if (!capable(CAP_SYS_ADMIN)) ++ if (!vx_can_unshare(CAP_SYS_ADMIN, unshare_flags)) + return -EPERM; + + *new_nsp = create_new_namespaces(unshare_flags, current, +diff -NurpP --minimal linux-3.0.9/kernel/pid.c linux-3.0.9-vs2.3.2.1/kernel/pid.c +--- linux-3.0.9/kernel/pid.c 2011-05-22 16:17:59.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/kernel/pid.c 2011-06-10 22:11:24.000000000 +0200 +@@ -36,6 +36,7 @@ + #include <linux/pid_namespace.h> + #include <linux/init_task.h> + #include <linux/syscalls.h> ++#include <linux/vs_pid.h> + + #define pid_hashfn(nr, ns) \ + hash_long((unsigned long)nr + (unsigned long)ns, pidhash_shift) +@@ -342,7 +343,7 @@ EXPORT_SYMBOL_GPL(find_pid_ns); + + struct pid *find_vpid(int nr) + { +- return find_pid_ns(nr, current->nsproxy->pid_ns); ++ return find_pid_ns(vx_rmap_pid(nr), current->nsproxy->pid_ns); + } + EXPORT_SYMBOL_GPL(find_vpid); + +@@ -402,6 +403,9 @@ void transfer_pid(struct task_struct *ol + struct task_struct *pid_task(struct pid *pid, enum pid_type type) + { + struct task_struct *result = NULL; ++ ++ if (type == PIDTYPE_REALPID) ++ type = PIDTYPE_PID; + if (pid) { + struct hlist_node *first; + first = rcu_dereference_check(hlist_first_rcu(&pid->tasks[type]), +@@ -420,7 +424,7 @@ EXPORT_SYMBOL(pid_task); + struct task_struct *find_task_by_pid_ns(pid_t nr, struct pid_namespace *ns) + { + rcu_lockdep_assert(rcu_read_lock_held()); +- return pid_task(find_pid_ns(nr, ns), PIDTYPE_PID); ++ return pid_task(find_pid_ns(vx_rmap_pid(nr), ns), PIDTYPE_PID); + } + + struct task_struct *find_task_by_vpid(pid_t vnr) +@@ -464,7 +468,7 @@ struct pid *find_get_pid(pid_t nr) + } + EXPORT_SYMBOL_GPL(find_get_pid); + +-pid_t pid_nr_ns(struct pid *pid, struct pid_namespace *ns) ++pid_t pid_unmapped_nr_ns(struct pid *pid, struct pid_namespace *ns) + { + struct upid *upid; + pid_t nr = 0; +@@ -477,6 +481,11 @@ pid_t pid_nr_ns(struct pid *pid, struct + return nr; + } + ++pid_t pid_nr_ns(struct pid *pid, struct pid_namespace *ns) ++{ ++ return vx_map_pid(pid_unmapped_nr_ns(pid, ns)); ++} ++ + pid_t pid_vnr(struct pid *pid) + { + return pid_nr_ns(pid, current->nsproxy->pid_ns); +diff -NurpP --minimal linux-3.0.9/kernel/pid_namespace.c linux-3.0.9-vs2.3.2.1/kernel/pid_namespace.c +--- linux-3.0.9/kernel/pid_namespace.c 2011-05-22 16:17:59.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/kernel/pid_namespace.c 2011-06-10 22:17:45.000000000 +0200 +@@ -15,6 +15,7 @@ + #include <linux/acct.h> + #include <linux/slab.h> + #include <linux/proc_fs.h> ++#include <linux/vserver/global.h> + + #define BITS_PER_PAGE (PAGE_SIZE*8) + +@@ -88,6 +89,7 @@ static struct pid_namespace *create_pid_ + goto out_free_map; + + kref_init(&ns->kref); ++ atomic_inc(&vs_global_pid_ns); + ns->level = level; + ns->parent = get_pid_ns(parent_pid_ns); + +@@ -119,6 +121,7 @@ static void destroy_pid_namespace(struct + + for (i = 0; i < PIDMAP_ENTRIES; i++) + kfree(ns->pidmap[i].page); ++ atomic_dec(&vs_global_pid_ns); + kmem_cache_free(pid_ns_cachep, ns); + } + +diff -NurpP --minimal linux-3.0.9/kernel/posix-timers.c linux-3.0.9-vs2.3.2.1/kernel/posix-timers.c +--- linux-3.0.9/kernel/posix-timers.c 2011-07-22 11:18:12.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/kernel/posix-timers.c 2011-06-10 22:11:24.000000000 +0200 +@@ -47,6 +47,7 @@ + #include <linux/wait.h> + #include <linux/workqueue.h> + #include <linux/module.h> ++#include <linux/vs_context.h> + + /* + * Management arrays for POSIX timers. Timers are kept in slab memory +@@ -340,6 +341,7 @@ int posix_timer_event(struct k_itimer *t + { + struct task_struct *task; + int shared, ret = -1; ++ + /* + * FIXME: if ->sigq is queued we can race with + * dequeue_signal()->do_schedule_next_timer(). +@@ -356,10 +358,18 @@ int posix_timer_event(struct k_itimer *t + rcu_read_lock(); + task = pid_task(timr->it_pid, PIDTYPE_PID); + if (task) { ++ struct vx_info_save vxis; ++ struct vx_info *vxi; ++ ++ vxi = get_vx_info(task->vx_info); ++ enter_vx_info(vxi, &vxis); + shared = !(timr->it_sigev_notify & SIGEV_THREAD_ID); + ret = send_sigqueue(timr->sigq, task, shared); ++ leave_vx_info(&vxis); ++ put_vx_info(vxi); + } + rcu_read_unlock(); ++ + /* If we failed to send the signal the timer stops. */ + return ret > 0; + } +diff -NurpP --minimal linux-3.0.9/kernel/printk.c linux-3.0.9-vs2.3.2.1/kernel/printk.c +--- linux-3.0.9/kernel/printk.c 2011-11-15 16:40:47.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/kernel/printk.c 2011-10-18 13:51:13.000000000 +0200 +@@ -41,6 +41,7 @@ + #include <linux/cpu.h> + #include <linux/notifier.h> + #include <linux/rculist.h> ++#include <linux/vs_cvirt.h> + + #include <asm/uaccess.h> + +@@ -314,7 +315,7 @@ static int check_syslog_permissions(int + return 0; + + if (syslog_action_restricted(type)) { +- if (capable(CAP_SYSLOG)) ++ if (vx_capable(CAP_SYSLOG, VXC_SYSLOG)) + return 0; + /* For historical reasons, accept CAP_SYS_ADMIN too, with a warning */ + if (capable(CAP_SYS_ADMIN)) { +@@ -342,12 +343,9 @@ int do_syslog(int type, char __user *buf + if (error) + return error; + +- switch (type) { +- case SYSLOG_ACTION_CLOSE: /* Close log */ +- break; +- case SYSLOG_ACTION_OPEN: /* Open log */ +- break; +- case SYSLOG_ACTION_READ: /* Read from log */ ++ if ((type == SYSLOG_ACTION_READ) || ++ (type == SYSLOG_ACTION_READ_ALL) || ++ (type == SYSLOG_ACTION_READ_CLEAR)) { + error = -EINVAL; + if (!buf || len < 0) + goto out; +@@ -358,6 +356,16 @@ int do_syslog(int type, char __user *buf + error = -EFAULT; + goto out; + } ++ } ++ if (!vx_check(0, VS_ADMIN|VS_WATCH)) ++ return vx_do_syslog(type, buf, len); ++ ++ switch (type) { ++ case SYSLOG_ACTION_CLOSE: /* Close log */ ++ break; ++ case SYSLOG_ACTION_OPEN: /* Open log */ ++ break; ++ case SYSLOG_ACTION_READ: /* Read from log */ + error = wait_event_interruptible(log_wait, + (log_start - log_end)); + if (error) +@@ -384,16 +392,6 @@ int do_syslog(int type, char __user *buf + /* FALL THRU */ + /* Read last kernel messages */ + case SYSLOG_ACTION_READ_ALL: +- error = -EINVAL; +- if (!buf || len < 0) +- goto out; +- error = 0; +- if (!len) +- goto out; +- if (!access_ok(VERIFY_WRITE, buf, len)) { +- error = -EFAULT; +- goto out; +- } + count = len; + if (count > log_buf_len) + count = log_buf_len; +diff -NurpP --minimal linux-3.0.9/kernel/ptrace.c linux-3.0.9-vs2.3.2.1/kernel/ptrace.c +--- linux-3.0.9/kernel/ptrace.c 2011-07-22 11:18:12.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/kernel/ptrace.c 2011-10-27 16:12:46.000000000 +0200 +@@ -22,6 +22,7 @@ + #include <linux/syscalls.h> + #include <linux/uaccess.h> + #include <linux/regset.h> ++#include <linux/vs_context.h> + #include <linux/hw_breakpoint.h> + + +@@ -145,9 +146,15 @@ int __ptrace_may_access(struct task_stru + * or halting the specified task is impossible. + */ + int dumpable = 0; ++ + /* Don't let security modules deny introspection */ + if (task == current) + return 0; ++ ++ vxdprintk(VXD_CBIT(perm, 8), ++ "__ptrace_may_access(%p[#%d,%d,%d], %d)", ++ task, task->xid, task->pid, task->tgid, mode); ++ + rcu_read_lock(); + tcred = __task_cred(task); + if (cred->user->user_ns == tcred->user->user_ns && +@@ -161,6 +168,8 @@ int __ptrace_may_access(struct task_stru + if (ns_capable(tcred->user->user_ns, CAP_SYS_PTRACE)) + goto ok; + rcu_read_unlock(); ++ vxdprintk(VXD_CBIT(perm, 8), ++ "__ptrace_may_access(%p) cred/cap failed", task); + return -EPERM; + ok: + rcu_read_unlock(); +@@ -169,6 +178,24 @@ ok: + dumpable = get_dumpable(task->mm); + if (!dumpable && !task_ns_capable(task, CAP_SYS_PTRACE)) + return -EPERM; ++ vxdprintk(VXD_CBIT(perm, 8), ++ "__ptrace_may_access(%p) cap/dump ok", task); ++ ++ if (!vx_check(task->xid, VS_ADMIN_P|VS_WATCH_P|VS_IDENT)) ++ return -EPERM; ++ vxdprintk(VXD_CBIT(perm, 8), ++ "__ptrace_may_access(%p) check ok", task); ++ ++ printk("%d,%d %d,%d\n", ++ vx_check(task->xid, VS_IDENT), ++ task_vx_flags(task, VXF_STATE_ADMIN, 0), ++ current->xid, task->xid); ++ ++ if (!vx_check(task->xid, VS_IDENT) && ++ !task_vx_flags(task, VXF_STATE_ADMIN, 0)) ++ return -EACCES; ++ vxdprintk(VXD_CBIT(perm, 8), ++ "__ptrace_may_access(%p) admin ok", task); + + return security_ptrace_access_check(task, mode); + } +diff -NurpP --minimal linux-3.0.9/kernel/sched.c linux-3.0.9-vs2.3.2.1/kernel/sched.c +--- linux-3.0.9/kernel/sched.c 2011-11-15 16:40:47.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/kernel/sched.c 2011-10-18 13:51:13.000000000 +0200 +@@ -71,6 +71,8 @@ + #include <linux/ctype.h> + #include <linux/ftrace.h> + #include <linux/slab.h> ++#include <linux/vs_sched.h> ++#include <linux/vs_cvirt.h> + + #include <asm/tlb.h> + #include <asm/irq_regs.h> +@@ -3461,9 +3463,17 @@ static void calc_global_nohz(unsigned lo + */ + void get_avenrun(unsigned long *loads, unsigned long offset, int shift) + { +- loads[0] = (avenrun[0] + offset) << shift; +- loads[1] = (avenrun[1] + offset) << shift; +- loads[2] = (avenrun[2] + offset) << shift; ++ if (vx_flags(VXF_VIRT_LOAD, 0)) { ++ struct vx_info *vxi = current_vx_info(); ++ ++ loads[0] = (vxi->cvirt.load[0] + offset) << shift; ++ loads[1] = (vxi->cvirt.load[1] + offset) << shift; ++ loads[2] = (vxi->cvirt.load[2] + offset) << shift; ++ } else { ++ loads[0] = (avenrun[0] + offset) << shift; ++ loads[1] = (avenrun[1] + offset) << shift; ++ loads[2] = (avenrun[2] + offset) << shift; ++ } + } + + /* +@@ -3722,16 +3732,19 @@ void account_user_time(struct task_struc + cputime_t cputime_scaled) + { + struct cpu_usage_stat *cpustat = &kstat_this_cpu.cpustat; ++ struct vx_info *vxi = p->vx_info; /* p is _always_ current */ + cputime64_t tmp; ++ int nice = (TASK_NICE(p) > 0); + + /* Add user time to process. */ + p->utime = cputime_add(p->utime, cputime); + p->utimescaled = cputime_add(p->utimescaled, cputime_scaled); ++ vx_account_user(vxi, cputime, nice); + account_group_user_time(p, cputime); + + /* Add user time to cpustat. */ + tmp = cputime_to_cputime64(cputime); +- if (TASK_NICE(p) > 0) ++ if (nice) + cpustat->nice = cputime64_add(cpustat->nice, tmp); + else + cpustat->user = cputime64_add(cpustat->user, tmp); +@@ -3783,10 +3796,12 @@ void __account_system_time(struct task_s + cputime_t cputime_scaled, cputime64_t *target_cputime64) + { + cputime64_t tmp = cputime_to_cputime64(cputime); ++ struct vx_info *vxi = p->vx_info; /* p is _always_ current */ + + /* Add system time to process. */ + p->stime = cputime_add(p->stime, cputime); + p->stimescaled = cputime_add(p->stimescaled, cputime_scaled); ++ vx_account_system(vxi, cputime, 0 /* do we have idle time? */); + account_group_system_time(p, cputime); + + /* Add system time to cpustat. */ +@@ -4954,7 +4969,7 @@ SYSCALL_DEFINE1(nice, int, increment) + nice = 19; + + if (increment < 0 && !can_nice(current, nice)) +- return -EPERM; ++ return vx_flags(VXF_IGNEG_NICE, 0) ? 0 : -EPERM; + + retval = security_task_setnice(current, nice); + if (retval) +diff -NurpP --minimal linux-3.0.9/kernel/sched_fair.c linux-3.0.9-vs2.3.2.1/kernel/sched_fair.c +--- linux-3.0.9/kernel/sched_fair.c 2011-07-22 11:18:12.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/kernel/sched_fair.c 2011-07-22 11:20:39.000000000 +0200 +@@ -998,6 +998,8 @@ enqueue_entity(struct cfs_rq *cfs_rq, st + __enqueue_entity(cfs_rq, se); + se->on_rq = 1; + ++ if (entity_is_task(se)) ++ vx_activate_task(task_of(se)); + if (cfs_rq->nr_running == 1) + list_add_leaf_cfs_rq(cfs_rq); + } +@@ -1074,6 +1076,8 @@ dequeue_entity(struct cfs_rq *cfs_rq, st + if (se != cfs_rq->curr) + __dequeue_entity(cfs_rq, se); + se->on_rq = 0; ++ if (entity_is_task(se)) ++ vx_deactivate_task(task_of(se)); + update_cfs_load(cfs_rq, 0); + account_entity_dequeue(cfs_rq, se); + +diff -NurpP --minimal linux-3.0.9/kernel/signal.c linux-3.0.9-vs2.3.2.1/kernel/signal.c +--- linux-3.0.9/kernel/signal.c 2011-11-15 16:40:47.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/kernel/signal.c 2011-11-15 17:37:07.000000000 +0100 +@@ -28,6 +28,8 @@ + #include <linux/freezer.h> + #include <linux/pid_namespace.h> + #include <linux/nsproxy.h> ++#include <linux/vs_context.h> ++#include <linux/vs_pid.h> + #define CREATE_TRACE_POINTS + #include <trace/events/signal.h> + +@@ -744,9 +746,18 @@ static int check_kill_permission(int sig + struct pid *sid; + int error; + ++ vxdprintk(VXD_CBIT(misc, 7), ++ "check_kill_permission(%d,%p,%p[#%u,%u])", ++ sig, info, t, vx_task_xid(t), t->pid); ++ + if (!valid_signal(sig)) + return -EINVAL; + ++/* FIXME: needed? if so, why? ++ if ((info != SEND_SIG_NOINFO) && ++ (is_si_special(info) || !si_fromuser(info))) ++ goto skip; */ ++ + if (!si_fromuser(info)) + return 0; + +@@ -770,6 +781,20 @@ static int check_kill_permission(int sig + } + } + ++ error = -EPERM; ++ if (t->pid == 1 && current->xid) ++ return error; ++ ++ error = -ESRCH; ++ /* FIXME: we shouldn't return ESRCH ever, to avoid ++ loops, maybe ENOENT or EACCES? */ ++ if (!vx_check(vx_task_xid(t), VS_WATCH_P | VS_IDENT)) { ++ vxdprintk(current->xid || VXD_CBIT(misc, 7), ++ "signal %d[%p] xid mismatch %p[#%u,%u] xid=#%u", ++ sig, info, t, vx_task_xid(t), t->pid, current->xid); ++ return error; ++ } ++/* skip: */ + return security_task_kill(t, info, sig, 0); + } + +@@ -1246,7 +1271,7 @@ int kill_pid_info(int sig, struct siginf + rcu_read_lock(); + retry: + p = pid_task(pid, PIDTYPE_PID); +- if (p) { ++ if (p && vx_check(vx_task_xid(p), VS_IDENT)) { + error = group_send_sig_info(sig, info, p); + if (unlikely(error == -ESRCH)) + /* +@@ -1285,7 +1310,7 @@ int kill_pid_info_as_uid(int sig, struct + + rcu_read_lock(); + p = pid_task(pid, PIDTYPE_PID); +- if (!p) { ++ if (!p || !vx_check(vx_task_xid(p), VS_IDENT)) { + ret = -ESRCH; + goto out_unlock; + } +@@ -1340,8 +1365,10 @@ static int kill_something_info(int sig, + struct task_struct * p; + + for_each_process(p) { +- if (task_pid_vnr(p) > 1 && +- !same_thread_group(p, current)) { ++ if (vx_check(vx_task_xid(p), VS_ADMIN|VS_IDENT) && ++ task_pid_vnr(p) > 1 && ++ !same_thread_group(p, current) && ++ !vx_current_initpid(p->pid)) { + int err = group_send_sig_info(sig, info, p); + ++count; + if (err != -EPERM) +@@ -2138,6 +2165,11 @@ relock: + !sig_kernel_only(signr)) + continue; + ++ /* virtual init is protected against user signals */ ++ if ((info->si_code == SI_USER) && ++ vx_current_initpid(current->pid)) ++ continue; ++ + if (sig_kernel_stop(signr)) { + /* + * The default action is to stop all threads in +diff -NurpP --minimal linux-3.0.9/kernel/softirq.c linux-3.0.9-vs2.3.2.1/kernel/softirq.c +--- linux-3.0.9/kernel/softirq.c 2011-07-22 11:18:12.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/kernel/softirq.c 2011-07-22 11:20:39.000000000 +0200 +@@ -24,6 +24,7 @@ + #include <linux/ftrace.h> + #include <linux/smp.h> + #include <linux/tick.h> ++#include <linux/vs_context.h> + + #define CREATE_TRACE_POINTS + #include <trace/events/irq.h> +diff -NurpP --minimal linux-3.0.9/kernel/sys.c linux-3.0.9-vs2.3.2.1/kernel/sys.c +--- linux-3.0.9/kernel/sys.c 2011-11-15 16:40:47.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/kernel/sys.c 2011-11-15 17:37:07.000000000 +0100 +@@ -45,6 +45,7 @@ + #include <linux/syscalls.h> + #include <linux/kprobes.h> + #include <linux/user_namespace.h> ++#include <linux/vs_pid.h> + + #include <linux/kmsg_dump.h> + /* Move somewhere else to avoid recompiling? */ +@@ -155,7 +156,10 @@ static int set_one_prio(struct task_stru + goto out; + } + if (niceval < task_nice(p) && !can_nice(p, niceval)) { +- error = -EACCES; ++ if (vx_flags(VXF_IGNEG_NICE, 0)) ++ error = 0; ++ else ++ error = -EACCES; + goto out; + } + no_nice = security_task_setnice(p, niceval); +@@ -205,6 +209,8 @@ SYSCALL_DEFINE3(setpriority, int, which, + else + pgrp = task_pgrp(current); + do_each_pid_thread(pgrp, PIDTYPE_PGID, p) { ++ if (!vx_check(p->xid, VS_ADMIN_P | VS_IDENT)) ++ continue; + error = set_one_prio(p, niceval, error); + } while_each_pid_thread(pgrp, PIDTYPE_PGID, p); + break; +@@ -268,6 +274,8 @@ SYSCALL_DEFINE2(getpriority, int, which, + else + pgrp = task_pgrp(current); + do_each_pid_thread(pgrp, PIDTYPE_PGID, p) { ++ if (!vx_check(p->xid, VS_ADMIN_P | VS_IDENT)) ++ continue; + niceval = 20 - task_nice(p); + if (niceval > retval) + retval = niceval; +@@ -387,6 +395,8 @@ EXPORT_SYMBOL_GPL(kernel_power_off); + + static DEFINE_MUTEX(reboot_mutex); + ++long vs_reboot(unsigned int, void __user *); ++ + /* + * Reboot system call: for obvious reasons only root may call it, + * and even root needs to set up some magic numbers in the registers +@@ -419,6 +429,9 @@ SYSCALL_DEFINE4(reboot, int, magic1, int + if ((cmd == LINUX_REBOOT_CMD_POWER_OFF) && !pm_power_off) + cmd = LINUX_REBOOT_CMD_HALT; + ++ if (!vx_check(0, VS_ADMIN|VS_WATCH)) ++ return vs_reboot(cmd, arg); ++ + mutex_lock(&reboot_mutex); + switch (cmd) { + case LINUX_REBOOT_CMD_RESTART: +@@ -1235,7 +1248,8 @@ SYSCALL_DEFINE2(sethostname, char __user + int errno; + char tmp[__NEW_UTS_LEN]; + +- if (!ns_capable(current->nsproxy->uts_ns->user_ns, CAP_SYS_ADMIN)) ++ if (!vx_ns_capable(current->nsproxy->uts_ns->user_ns, ++ CAP_SYS_ADMIN, VXC_SET_UTSNAME)) + return -EPERM; + + if (len < 0 || len > __NEW_UTS_LEN) +@@ -1285,7 +1299,8 @@ SYSCALL_DEFINE2(setdomainname, char __us + int errno; + char tmp[__NEW_UTS_LEN]; + +- if (!ns_capable(current->nsproxy->uts_ns->user_ns, CAP_SYS_ADMIN)) ++ if (!vx_ns_capable(current->nsproxy->uts_ns->user_ns, ++ CAP_SYS_ADMIN, VXC_SET_UTSNAME)) + return -EPERM; + if (len < 0 || len > __NEW_UTS_LEN) + return -EINVAL; +@@ -1403,7 +1418,7 @@ int do_prlimit(struct task_struct *tsk, + /* Keep the capable check against init_user_ns until + cgroups can contain all limits */ + if (new_rlim->rlim_max > rlim->rlim_max && +- !capable(CAP_SYS_RESOURCE)) ++ !vx_capable(CAP_SYS_RESOURCE, VXC_SET_RLIMIT)) + retval = -EPERM; + if (!retval) + retval = security_task_setrlimit(tsk->group_leader, +@@ -1457,7 +1472,8 @@ static int check_prlimit_permission(stru + cred->gid == tcred->sgid && + cred->gid == tcred->gid)) + return 0; +- if (ns_capable(tcred->user->user_ns, CAP_SYS_RESOURCE)) ++ if (vx_ns_capable(tcred->user->user_ns, ++ CAP_SYS_RESOURCE, VXC_SET_RLIMIT)) + return 0; + + return -EPERM; +diff -NurpP --minimal linux-3.0.9/kernel/sysctl.c linux-3.0.9-vs2.3.2.1/kernel/sysctl.c +--- linux-3.0.9/kernel/sysctl.c 2011-07-22 11:18:12.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/kernel/sysctl.c 2011-06-15 02:40:14.000000000 +0200 +@@ -75,6 +75,7 @@ + #if defined(CONFIG_PROVE_LOCKING) || defined(CONFIG_LOCK_STAT) + #include <linux/lockdep.h> + #endif ++extern char vshelper_path[]; + #ifdef CONFIG_CHR_DEV_SG + #include <scsi/sg.h> + #endif +@@ -568,6 +569,13 @@ static struct ctl_table kern_table[] = { + .proc_handler = proc_dostring, + }, + #endif ++ { ++ .procname = "vshelper", ++ .data = &vshelper_path, ++ .maxlen = 256, ++ .mode = 0644, ++ .proc_handler = &proc_dostring, ++ }, + #ifdef CONFIG_CHR_DEV_SG + { + .procname = "sg-big-buff", +diff -NurpP --minimal linux-3.0.9/kernel/sysctl_binary.c linux-3.0.9-vs2.3.2.1/kernel/sysctl_binary.c +--- linux-3.0.9/kernel/sysctl_binary.c 2011-05-22 16:17:59.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/kernel/sysctl_binary.c 2011-06-10 22:11:24.000000000 +0200 +@@ -73,6 +73,7 @@ static const struct bin_table bin_kern_t + + { CTL_INT, KERN_PANIC, "panic" }, + { CTL_INT, KERN_REALROOTDEV, "real-root-dev" }, ++ { CTL_STR, KERN_VSHELPER, "vshelper" }, + + { CTL_STR, KERN_SPARC_REBOOT, "reboot-cmd" }, + { CTL_INT, KERN_CTLALTDEL, "ctrl-alt-del" }, +diff -NurpP --minimal linux-3.0.9/kernel/time/timekeeping.c linux-3.0.9-vs2.3.2.1/kernel/time/timekeeping.c +--- linux-3.0.9/kernel/time/timekeeping.c 2011-07-22 11:18:12.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/kernel/time/timekeeping.c 2011-06-10 22:11:24.000000000 +0200 +@@ -233,6 +233,7 @@ void getnstimeofday(struct timespec *ts) + } while (read_seqretry(&xtime_lock, seq)); + + timespec_add_ns(ts, nsecs); ++ vx_adjust_timespec(ts); + } + + EXPORT_SYMBOL(getnstimeofday); +diff -NurpP --minimal linux-3.0.9/kernel/time.c linux-3.0.9-vs2.3.2.1/kernel/time.c +--- linux-3.0.9/kernel/time.c 2011-11-15 16:40:47.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/kernel/time.c 2011-11-15 17:37:07.000000000 +0100 +@@ -92,7 +92,7 @@ SYSCALL_DEFINE1(stime, time_t __user *, + if (err) + return err; + +- do_settimeofday(&tv); ++ vx_settimeofday(&tv); + return 0; + } + +@@ -177,7 +177,7 @@ int do_sys_settimeofday(const struct tim + /* SMP safe, again the code in arch/foo/time.c should + * globally block out interrupts when it runs. + */ +- return do_settimeofday(tv); ++ return vx_settimeofday(tv); + } + return 0; + } +diff -NurpP --minimal linux-3.0.9/kernel/timer.c linux-3.0.9-vs2.3.2.1/kernel/timer.c +--- linux-3.0.9/kernel/timer.c 2011-07-22 11:18:12.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/kernel/timer.c 2011-06-15 02:40:14.000000000 +0200 +@@ -40,6 +40,10 @@ + #include <linux/irq_work.h> + #include <linux/sched.h> + #include <linux/slab.h> ++#include <linux/vs_base.h> ++#include <linux/vs_cvirt.h> ++#include <linux/vs_pid.h> ++#include <linux/vserver/sched.h> + + #include <asm/uaccess.h> + #include <asm/unistd.h> +@@ -1336,12 +1340,6 @@ SYSCALL_DEFINE1(alarm, unsigned int, sec + + #endif + +-#ifndef __alpha__ +- +-/* +- * The Alpha uses getxpid, getxuid, and getxgid instead. Maybe this +- * should be moved into arch/i386 instead? +- */ + + /** + * sys_getpid - return the thread group id of the current process +@@ -1370,10 +1368,23 @@ SYSCALL_DEFINE0(getppid) + rcu_read_lock(); + pid = task_tgid_vnr(current->real_parent); + rcu_read_unlock(); ++ return vx_map_pid(pid); ++} + +- return pid; ++#ifdef __alpha__ ++ ++/* ++ * The Alpha uses getxpid, getxuid, and getxgid instead. ++ */ ++ ++asmlinkage long do_getxpid(long *ppid) ++{ ++ *ppid = sys_getppid(); ++ return sys_getpid(); + } + ++#else /* _alpha_ */ ++ + SYSCALL_DEFINE0(getuid) + { + /* Only we change this so SMP safe */ +diff -NurpP --minimal linux-3.0.9/kernel/user_namespace.c linux-3.0.9-vs2.3.2.1/kernel/user_namespace.c +--- linux-3.0.9/kernel/user_namespace.c 2011-03-15 18:07:42.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/kernel/user_namespace.c 2011-06-10 22:11:24.000000000 +0200 +@@ -11,6 +11,7 @@ + #include <linux/user_namespace.h> + #include <linux/highuid.h> + #include <linux/cred.h> ++#include <linux/vserver/global.h> + + static struct kmem_cache *user_ns_cachep __read_mostly; + +@@ -33,6 +34,7 @@ int create_user_ns(struct cred *new) + return -ENOMEM; + + kref_init(&ns->kref); ++ atomic_inc(&vs_global_user_ns); + + for (n = 0; n < UIDHASH_SZ; ++n) + INIT_HLIST_HEAD(ns->uidhash_table + n); +@@ -81,6 +83,8 @@ void free_user_ns(struct kref *kref) + struct user_namespace *ns = + container_of(kref, struct user_namespace, kref); + ++ /* FIXME: maybe move into destroyer? */ ++ atomic_dec(&vs_global_user_ns); + INIT_WORK(&ns->destroyer, free_user_ns_work); + schedule_work(&ns->destroyer); + } +diff -NurpP --minimal linux-3.0.9/kernel/utsname.c linux-3.0.9-vs2.3.2.1/kernel/utsname.c +--- linux-3.0.9/kernel/utsname.c 2011-07-22 11:18:12.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/kernel/utsname.c 2011-06-13 14:09:44.000000000 +0200 +@@ -16,14 +16,17 @@ + #include <linux/slab.h> + #include <linux/user_namespace.h> + #include <linux/proc_fs.h> ++#include <linux/vserver/global.h> + + static struct uts_namespace *create_uts_ns(void) + { + struct uts_namespace *uts_ns; + + uts_ns = kmalloc(sizeof(struct uts_namespace), GFP_KERNEL); +- if (uts_ns) ++ if (uts_ns) { + kref_init(&uts_ns->kref); ++ atomic_inc(&vs_global_uts_ns); ++ } + return uts_ns; + } + +@@ -32,8 +35,8 @@ static struct uts_namespace *create_uts_ + * @old_ns: namespace to clone + * Return NULL on error (failure to kmalloc), new ns otherwise + */ +-static struct uts_namespace *clone_uts_ns(struct task_struct *tsk, +- struct uts_namespace *old_ns) ++static struct uts_namespace *clone_uts_ns(struct uts_namespace *old_ns, ++ struct user_namespace *old_user) + { + struct uts_namespace *ns; + +@@ -43,7 +46,7 @@ static struct uts_namespace *clone_uts_n + + down_read(&uts_sem); + memcpy(&ns->name, &old_ns->name, sizeof(ns->name)); +- ns->user_ns = get_user_ns(task_cred_xxx(tsk, user)->user_ns); ++ ns->user_ns = get_user_ns(old_user); + up_read(&uts_sem); + return ns; + } +@@ -55,9 +58,9 @@ static struct uts_namespace *clone_uts_n + * versa. + */ + struct uts_namespace *copy_utsname(unsigned long flags, +- struct task_struct *tsk) ++ struct uts_namespace *old_ns, ++ struct user_namespace *user_ns) + { +- struct uts_namespace *old_ns = tsk->nsproxy->uts_ns; + struct uts_namespace *new_ns; + + BUG_ON(!old_ns); +@@ -66,7 +69,7 @@ struct uts_namespace *copy_utsname(unsig + if (!(flags & CLONE_NEWUTS)) + return old_ns; + +- new_ns = clone_uts_ns(tsk, old_ns); ++ new_ns = clone_uts_ns(old_ns, user_ns); + + put_uts_ns(old_ns); + return new_ns; +@@ -78,6 +81,7 @@ void free_uts_ns(struct kref *kref) + + ns = container_of(kref, struct uts_namespace, kref); + put_user_ns(ns->user_ns); ++ atomic_dec(&vs_global_uts_ns); + kfree(ns); + } + +diff -NurpP --minimal linux-3.0.9/kernel/vserver/Kconfig linux-3.0.9-vs2.3.2.1/kernel/vserver/Kconfig +--- linux-3.0.9/kernel/vserver/Kconfig 1970-01-01 01:00:00.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/kernel/vserver/Kconfig 2011-08-08 18:06:22.000000000 +0200 +@@ -0,0 +1,232 @@ ++# ++# Linux VServer configuration ++# ++ ++menu "Linux VServer" ++ ++config VSERVER_AUTO_LBACK ++ bool "Automatically Assign Loopback IP" ++ default y ++ help ++ Automatically assign a guest specific loopback ++ IP and add it to the kernel network stack on ++ startup. ++ ++config VSERVER_AUTO_SINGLE ++ bool "Automatic Single IP Special Casing" ++ depends on EXPERIMENTAL ++ default y ++ help ++ This allows network contexts with a single IP to ++ automatically remap 0.0.0.0 bindings to that IP, ++ avoiding further network checks and improving ++ performance. ++ ++ (note: such guests do not allow to change the ip ++ on the fly and do not show loopback addresses) ++ ++config VSERVER_COWBL ++ bool "Enable COW Immutable Link Breaking" ++ default y ++ help ++ This enables the COW (Copy-On-Write) link break code. ++ It allows you to treat unified files like normal files ++ when writing to them (which will implicitely break the ++ link and create a copy of the unified file) ++ ++config VSERVER_VTIME ++ bool "Enable Virtualized Guest Time" ++ depends on EXPERIMENTAL ++ default n ++ help ++ This enables per guest time offsets to allow for ++ adjusting the system clock individually per guest. ++ this adds some overhead to the time functions and ++ therefore should not be enabled without good reason. ++ ++config VSERVER_DEVICE ++ bool "Enable Guest Device Mapping" ++ depends on EXPERIMENTAL ++ default n ++ help ++ This enables generic device remapping. ++ ++config VSERVER_PROC_SECURE ++ bool "Enable Proc Security" ++ depends on PROC_FS ++ default y ++ help ++ This configures ProcFS security to initially hide ++ non-process entries for all contexts except the main and ++ spectator context (i.e. for all guests), which is a secure ++ default. ++ ++ (note: on 1.2x the entries were visible by default) ++ ++choice ++ prompt "Persistent Inode Tagging" ++ default TAGGING_ID24 ++ help ++ This adds persistent context information to filesystems ++ mounted with the tagxid option. Tagging is a requirement ++ for per-context disk limits and per-context quota. ++ ++ ++config TAGGING_NONE ++ bool "Disabled" ++ help ++ do not store per-context information in inodes. ++ ++config TAGGING_UID16 ++ bool "UID16/GID32" ++ help ++ reduces UID to 16 bit, but leaves GID at 32 bit. ++ ++config TAGGING_GID16 ++ bool "UID32/GID16" ++ help ++ reduces GID to 16 bit, but leaves UID at 32 bit. ++ ++config TAGGING_ID24 ++ bool "UID24/GID24" ++ help ++ uses the upper 8bit from UID and GID for XID tagging ++ which leaves 24bit for UID/GID each, which should be ++ more than sufficient for normal use. ++ ++config TAGGING_INTERN ++ bool "UID32/GID32" ++ help ++ this uses otherwise reserved inode fields in the on ++ disk representation, which limits the use to a few ++ filesystems (currently ext2 and ext3) ++ ++endchoice ++ ++config TAG_NFSD ++ bool "Tag NFSD User Auth and Files" ++ default n ++ help ++ Enable this if you do want the in-kernel NFS ++ Server to use the tagging specified above. ++ (will require patched clients too) ++ ++config VSERVER_PRIVACY ++ bool "Honor Privacy Aspects of Guests" ++ default n ++ help ++ When enabled, most context checks will disallow ++ access to structures assigned to a specific context, ++ like ptys or loop devices. ++ ++config VSERVER_CONTEXTS ++ int "Maximum number of Contexts (1-65533)" if EMBEDDED ++ range 1 65533 ++ default "768" if 64BIT ++ default "256" ++ help ++ This setting will optimize certain data structures ++ and memory allocations according to the expected ++ maximum. ++ ++ note: this is not a strict upper limit. ++ ++config VSERVER_WARN ++ bool "VServer Warnings" ++ default y ++ help ++ This enables various runtime warnings, which will ++ notify about potential manipulation attempts or ++ resource shortage. It is generally considered to ++ be a good idea to have that enabled. ++ ++config VSERVER_WARN_DEVPTS ++ bool "VServer DevPTS Warnings" ++ depends on VSERVER_WARN ++ default y ++ help ++ This enables DevPTS related warnings, issued when a ++ process inside a context tries to lookup or access ++ a dynamic pts from the host or a different context. ++ ++config VSERVER_DEBUG ++ bool "VServer Debugging Code" ++ default n ++ help ++ Set this to yes if you want to be able to activate ++ debugging output at runtime. It adds a very small ++ overhead to all vserver related functions and ++ increases the kernel size by about 20k. ++ ++config VSERVER_HISTORY ++ bool "VServer History Tracing" ++ depends on VSERVER_DEBUG ++ default n ++ help ++ Set this to yes if you want to record the history of ++ linux-vserver activities, so they can be replayed in ++ the event of a kernel panic or oops. ++ ++config VSERVER_HISTORY_SIZE ++ int "Per-CPU History Size (32-65536)" ++ depends on VSERVER_HISTORY ++ range 32 65536 ++ default 64 ++ help ++ This allows you to specify the number of entries in ++ the per-CPU history buffer. ++ ++config VSERVER_LEGACY_MEM ++ bool "Legacy Memory Limits" ++ default n ++ help ++ This provides fake memory limits to keep ++ older tools happy in the face of memory ++ cgroups ++ ++choice ++ prompt "Quotes used in debug and warn messages" ++ default QUOTES_ISO8859 ++ ++config QUOTES_ISO8859 ++ bool "Extended ASCII (ISO 8859) angle quotes" ++ help ++ This uses the extended ASCII characters \xbb ++ and \xab for quoting file and process names. ++ ++config QUOTES_UTF8 ++ bool "UTF-8 angle quotes" ++ help ++ This uses the the UTF-8 sequences for angle ++ quotes to quote file and process names. ++ ++config QUOTES_ASCII ++ bool "ASCII single quotes" ++ help ++ This uses the ASCII single quote character ++ (\x27) to quote file and process names. ++ ++endchoice ++ ++endmenu ++ ++ ++config VSERVER ++ bool ++ default y ++ select NAMESPACES ++ select UTS_NS ++ select IPC_NS ++# select USER_NS ++ select SYSVIPC ++ ++config VSERVER_SECURITY ++ bool ++ depends on SECURITY ++ default y ++ select SECURITY_CAPABILITIES ++ ++config VSERVER_DISABLED ++ bool ++ default n ++ +diff -NurpP --minimal linux-3.0.9/kernel/vserver/Makefile linux-3.0.9-vs2.3.2.1/kernel/vserver/Makefile +--- linux-3.0.9/kernel/vserver/Makefile 1970-01-01 01:00:00.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/kernel/vserver/Makefile 2011-06-10 22:11:24.000000000 +0200 +@@ -0,0 +1,18 @@ ++# ++# Makefile for the Linux vserver routines. ++# ++ ++ ++obj-y += vserver.o ++ ++vserver-y := switch.o context.o space.o sched.o network.o inode.o \ ++ limit.o cvirt.o cacct.o signal.o helper.o init.o \ ++ dlimit.o tag.o ++ ++vserver-$(CONFIG_INET) += inet.o ++vserver-$(CONFIG_PROC_FS) += proc.o ++vserver-$(CONFIG_VSERVER_DEBUG) += sysctl.o debug.o ++vserver-$(CONFIG_VSERVER_HISTORY) += history.o ++vserver-$(CONFIG_VSERVER_MONITOR) += monitor.o ++vserver-$(CONFIG_VSERVER_DEVICE) += device.o ++ +diff -NurpP --minimal linux-3.0.9/kernel/vserver/cacct.c linux-3.0.9-vs2.3.2.1/kernel/vserver/cacct.c +--- linux-3.0.9/kernel/vserver/cacct.c 1970-01-01 01:00:00.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/kernel/vserver/cacct.c 2011-06-10 22:11:24.000000000 +0200 +@@ -0,0 +1,42 @@ ++/* ++ * linux/kernel/vserver/cacct.c ++ * ++ * Virtual Server: Context Accounting ++ * ++ * Copyright (C) 2006-2007 Herbert Pötzl ++ * ++ * V0.01 added accounting stats ++ * ++ */ ++ ++#include <linux/types.h> ++#include <linux/vs_context.h> ++#include <linux/vserver/cacct_cmd.h> ++#include <linux/vserver/cacct_int.h> ++ ++#include <asm/errno.h> ++#include <asm/uaccess.h> ++ ++ ++int vc_sock_stat(struct vx_info *vxi, void __user *data) ++{ ++ struct vcmd_sock_stat_v0 vc_data; ++ int j, field; ++ ++ if (copy_from_user(&vc_data, data, sizeof(vc_data))) ++ return -EFAULT; ++ ++ field = vc_data.field; ++ if ((field < 0) || (field >= VXA_SOCK_SIZE)) ++ return -EINVAL; ++ ++ for (j = 0; j < 3; j++) { ++ vc_data.count[j] = vx_sock_count(&vxi->cacct, field, j); ++ vc_data.total[j] = vx_sock_total(&vxi->cacct, field, j); ++ } ++ ++ if (copy_to_user(data, &vc_data, sizeof(vc_data))) ++ return -EFAULT; ++ return 0; ++} ++ +diff -NurpP --minimal linux-3.0.9/kernel/vserver/cacct_init.h linux-3.0.9-vs2.3.2.1/kernel/vserver/cacct_init.h +--- linux-3.0.9/kernel/vserver/cacct_init.h 1970-01-01 01:00:00.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/kernel/vserver/cacct_init.h 2011-06-10 22:11:24.000000000 +0200 +@@ -0,0 +1,25 @@ ++ ++ ++static inline void vx_info_init_cacct(struct _vx_cacct *cacct) ++{ ++ int i, j; ++ ++ ++ for (i = 0; i < VXA_SOCK_SIZE; i++) { ++ for (j = 0; j < 3; j++) { ++ atomic_long_set(&cacct->sock[i][j].count, 0); ++ atomic_long_set(&cacct->sock[i][j].total, 0); ++ } ++ } ++ for (i = 0; i < 8; i++) ++ atomic_set(&cacct->slab[i], 0); ++ for (i = 0; i < 5; i++) ++ for (j = 0; j < 4; j++) ++ atomic_set(&cacct->page[i][j], 0); ++} ++ ++static inline void vx_info_exit_cacct(struct _vx_cacct *cacct) ++{ ++ return; ++} ++ +diff -NurpP --minimal linux-3.0.9/kernel/vserver/cacct_proc.h linux-3.0.9-vs2.3.2.1/kernel/vserver/cacct_proc.h +--- linux-3.0.9/kernel/vserver/cacct_proc.h 1970-01-01 01:00:00.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/kernel/vserver/cacct_proc.h 2011-06-10 22:11:24.000000000 +0200 +@@ -0,0 +1,53 @@ ++#ifndef _VX_CACCT_PROC_H ++#define _VX_CACCT_PROC_H ++ ++#include <linux/vserver/cacct_int.h> ++ ++ ++#define VX_SOCKA_TOP \ ++ "Type\t recv #/bytes\t\t send #/bytes\t\t fail #/bytes\n" ++ ++static inline int vx_info_proc_cacct(struct _vx_cacct *cacct, char *buffer) ++{ ++ int i, j, length = 0; ++ static char *type[VXA_SOCK_SIZE] = { ++ "UNSPEC", "UNIX", "INET", "INET6", "PACKET", "OTHER" ++ }; ++ ++ length += sprintf(buffer + length, VX_SOCKA_TOP); ++ for (i = 0; i < VXA_SOCK_SIZE; i++) { ++ length += sprintf(buffer + length, "%s:", type[i]); ++ for (j = 0; j < 3; j++) { ++ length += sprintf(buffer + length, ++ "\t%10lu/%-10lu", ++ vx_sock_count(cacct, i, j), ++ vx_sock_total(cacct, i, j)); ++ } ++ buffer[length++] = '\n'; ++ } ++ ++ length += sprintf(buffer + length, "\n"); ++ length += sprintf(buffer + length, ++ "slab:\t %8u %8u %8u %8u\n", ++ atomic_read(&cacct->slab[1]), ++ atomic_read(&cacct->slab[4]), ++ atomic_read(&cacct->slab[0]), ++ atomic_read(&cacct->slab[2])); ++ ++ length += sprintf(buffer + length, "\n"); ++ for (i = 0; i < 5; i++) { ++ length += sprintf(buffer + length, ++ "page[%d]: %8u %8u %8u %8u\t %8u %8u %8u %8u\n", i, ++ atomic_read(&cacct->page[i][0]), ++ atomic_read(&cacct->page[i][1]), ++ atomic_read(&cacct->page[i][2]), ++ atomic_read(&cacct->page[i][3]), ++ atomic_read(&cacct->page[i][4]), ++ atomic_read(&cacct->page[i][5]), ++ atomic_read(&cacct->page[i][6]), ++ atomic_read(&cacct->page[i][7])); ++ } ++ return length; ++} ++ ++#endif /* _VX_CACCT_PROC_H */ +diff -NurpP --minimal linux-3.0.9/kernel/vserver/context.c linux-3.0.9-vs2.3.2.1/kernel/vserver/context.c +--- linux-3.0.9/kernel/vserver/context.c 1970-01-01 01:00:00.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/kernel/vserver/context.c 2011-08-01 18:28:12.000000000 +0200 +@@ -0,0 +1,1107 @@ ++/* ++ * linux/kernel/vserver/context.c ++ * ++ * Virtual Server: Context Support ++ * ++ * Copyright (C) 2003-2011 Herbert Pötzl ++ * ++ * V0.01 context helper ++ * V0.02 vx_ctx_kill syscall command ++ * V0.03 replaced context_info calls ++ * V0.04 redesign of struct (de)alloc ++ * V0.05 rlimit basic implementation ++ * V0.06 task_xid and info commands ++ * V0.07 context flags and caps ++ * V0.08 switch to RCU based hash ++ * V0.09 revert to non RCU for now ++ * V0.10 and back to working RCU hash ++ * V0.11 and back to locking again ++ * V0.12 referenced context store ++ * V0.13 separate per cpu data ++ * V0.14 changed vcmds to vxi arg ++ * V0.15 added context stat ++ * V0.16 have __create claim() the vxi ++ * V0.17 removed older and legacy stuff ++ * V0.18 added user credentials ++ * V0.19 added warn mask ++ * ++ */ ++ ++#include <linux/slab.h> ++#include <linux/types.h> ++#include <linux/security.h> ++#include <linux/pid_namespace.h> ++#include <linux/capability.h> ++ ++#include <linux/vserver/context.h> ++#include <linux/vserver/network.h> ++#include <linux/vserver/debug.h> ++#include <linux/vserver/limit.h> ++#include <linux/vserver/limit_int.h> ++#include <linux/vserver/space.h> ++#include <linux/init_task.h> ++#include <linux/fs_struct.h> ++#include <linux/cred.h> ++ ++#include <linux/vs_context.h> ++#include <linux/vs_limit.h> ++#include <linux/vs_pid.h> ++#include <linux/vserver/context_cmd.h> ++ ++#include "cvirt_init.h" ++#include "cacct_init.h" ++#include "limit_init.h" ++#include "sched_init.h" ++ ++ ++atomic_t vx_global_ctotal = ATOMIC_INIT(0); ++atomic_t vx_global_cactive = ATOMIC_INIT(0); ++ ++ ++/* now inactive context structures */ ++ ++static struct hlist_head vx_info_inactive = HLIST_HEAD_INIT; ++ ++static DEFINE_SPINLOCK(vx_info_inactive_lock); ++ ++ ++/* __alloc_vx_info() ++ ++ * allocate an initialized vx_info struct ++ * doesn't make it visible (hash) */ ++ ++static struct vx_info *__alloc_vx_info(xid_t xid) ++{ ++ struct vx_info *new = NULL; ++ int cpu, index; ++ ++ vxdprintk(VXD_CBIT(xid, 0), "alloc_vx_info(%d)*", xid); ++ ++ /* would this benefit from a slab cache? */ ++ new = kmalloc(sizeof(struct vx_info), GFP_KERNEL); ++ if (!new) ++ return 0; ++ ++ memset(new, 0, sizeof(struct vx_info)); ++#ifdef CONFIG_SMP ++ new->ptr_pc = alloc_percpu(struct _vx_info_pc); ++ if (!new->ptr_pc) ++ goto error; ++#endif ++ new->vx_id = xid; ++ INIT_HLIST_NODE(&new->vx_hlist); ++ atomic_set(&new->vx_usecnt, 0); ++ atomic_set(&new->vx_tasks, 0); ++ new->vx_parent = NULL; ++ new->vx_state = 0; ++ init_waitqueue_head(&new->vx_wait); ++ ++ /* prepare reaper */ ++ get_task_struct(init_pid_ns.child_reaper); ++ new->vx_reaper = init_pid_ns.child_reaper; ++ new->vx_badness_bias = 0; ++ ++ /* rest of init goes here */ ++ vx_info_init_limit(&new->limit); ++ vx_info_init_sched(&new->sched); ++ vx_info_init_cvirt(&new->cvirt); ++ vx_info_init_cacct(&new->cacct); ++ ++ /* per cpu data structures */ ++ for_each_possible_cpu(cpu) { ++ vx_info_init_sched_pc( ++ &vx_per_cpu(new, sched_pc, cpu), cpu); ++ vx_info_init_cvirt_pc( ++ &vx_per_cpu(new, cvirt_pc, cpu), cpu); ++ } ++ ++ new->vx_flags = VXF_INIT_SET; ++ new->vx_bcaps = CAP_FULL_SET; // maybe ~CAP_SETPCAP ++ new->vx_ccaps = 0; ++ new->vx_umask = 0; ++ new->vx_wmask = 0; ++ ++ new->reboot_cmd = 0; ++ new->exit_code = 0; ++ ++ // preconfig spaces ++ for (index = 0; index < VX_SPACES; index++) { ++ struct _vx_space *space = &new->space[index]; ++ ++ // filesystem ++ spin_lock(&init_fs.lock); ++ init_fs.users++; ++ spin_unlock(&init_fs.lock); ++ space->vx_fs = &init_fs; ++ ++ /* FIXME: do we want defaults? */ ++ // space->vx_real_cred = 0; ++ // space->vx_cred = 0; ++ } ++ ++ ++ vxdprintk(VXD_CBIT(xid, 0), ++ "alloc_vx_info(%d) = %p", xid, new); ++ vxh_alloc_vx_info(new); ++ atomic_inc(&vx_global_ctotal); ++ return new; ++#ifdef CONFIG_SMP ++error: ++ kfree(new); ++ return 0; ++#endif ++} ++ ++/* __dealloc_vx_info() ++ ++ * final disposal of vx_info */ ++ ++static void __dealloc_vx_info(struct vx_info *vxi) ++{ ++#ifdef CONFIG_VSERVER_WARN ++ struct vx_info_save vxis; ++ int cpu; ++#endif ++ vxdprintk(VXD_CBIT(xid, 0), ++ "dealloc_vx_info(%p)", vxi); ++ vxh_dealloc_vx_info(vxi); ++ ++#ifdef CONFIG_VSERVER_WARN ++ enter_vx_info(vxi, &vxis); ++ vx_info_exit_limit(&vxi->limit); ++ vx_info_exit_sched(&vxi->sched); ++ vx_info_exit_cvirt(&vxi->cvirt); ++ vx_info_exit_cacct(&vxi->cacct); ++ ++ for_each_possible_cpu(cpu) { ++ vx_info_exit_sched_pc( ++ &vx_per_cpu(vxi, sched_pc, cpu), cpu); ++ vx_info_exit_cvirt_pc( ++ &vx_per_cpu(vxi, cvirt_pc, cpu), cpu); ++ } ++ leave_vx_info(&vxis); ++#endif ++ ++ vxi->vx_id = -1; ++ vxi->vx_state |= VXS_RELEASED; ++ ++#ifdef CONFIG_SMP ++ free_percpu(vxi->ptr_pc); ++#endif ++ kfree(vxi); ++ atomic_dec(&vx_global_ctotal); ++} ++ ++static void __shutdown_vx_info(struct vx_info *vxi) ++{ ++ struct nsproxy *nsproxy; ++ struct fs_struct *fs; ++ struct cred *cred; ++ int index, kill; ++ ++ might_sleep(); ++ ++ vxi->vx_state |= VXS_SHUTDOWN; ++ vs_state_change(vxi, VSC_SHUTDOWN); ++ ++ for (index = 0; index < VX_SPACES; index++) { ++ struct _vx_space *space = &vxi->space[index]; ++ ++ nsproxy = xchg(&space->vx_nsproxy, NULL); ++ if (nsproxy) ++ put_nsproxy(nsproxy); ++ ++ fs = xchg(&space->vx_fs, NULL); ++ spin_lock(&fs->lock); ++ kill = !--fs->users; ++ spin_unlock(&fs->lock); ++ if (kill) ++ free_fs_struct(fs); ++ ++ cred = (struct cred *)xchg(&space->vx_cred, NULL); ++ if (cred) ++ abort_creds(cred); ++ } ++} ++ ++/* exported stuff */ ++ ++void free_vx_info(struct vx_info *vxi) ++{ ++ unsigned long flags; ++ unsigned index; ++ ++ /* check for reference counts first */ ++ BUG_ON(atomic_read(&vxi->vx_usecnt)); ++ BUG_ON(atomic_read(&vxi->vx_tasks)); ++ ++ /* context must not be hashed */ ++ BUG_ON(vx_info_state(vxi, VXS_HASHED)); ++ ++ /* context shutdown is mandatory */ ++ BUG_ON(!vx_info_state(vxi, VXS_SHUTDOWN)); ++ ++ /* spaces check */ ++ for (index = 0; index < VX_SPACES; index++) { ++ struct _vx_space *space = &vxi->space[index]; ++ ++ BUG_ON(space->vx_nsproxy); ++ BUG_ON(space->vx_fs); ++ // BUG_ON(space->vx_real_cred); ++ // BUG_ON(space->vx_cred); ++ } ++ ++ spin_lock_irqsave(&vx_info_inactive_lock, flags); ++ hlist_del(&vxi->vx_hlist); ++ spin_unlock_irqrestore(&vx_info_inactive_lock, flags); ++ ++ __dealloc_vx_info(vxi); ++} ++ ++ ++/* hash table for vx_info hash */ ++ ++#define VX_HASH_SIZE 13 ++ ++static struct hlist_head vx_info_hash[VX_HASH_SIZE] = ++ { [0 ... VX_HASH_SIZE-1] = HLIST_HEAD_INIT }; ++ ++static DEFINE_SPINLOCK(vx_info_hash_lock); ++ ++ ++static inline unsigned int __hashval(xid_t xid) ++{ ++ return (xid % VX_HASH_SIZE); ++} ++ ++ ++ ++/* __hash_vx_info() ++ ++ * add the vxi to the global hash table ++ * requires the hash_lock to be held */ ++ ++static inline void __hash_vx_info(struct vx_info *vxi) ++{ ++ struct hlist_head *head; ++ ++ vxd_assert_lock(&vx_info_hash_lock); ++ vxdprintk(VXD_CBIT(xid, 4), ++ "__hash_vx_info: %p[#%d]", vxi, vxi->vx_id); ++ vxh_hash_vx_info(vxi); ++ ++ /* context must not be hashed */ ++ BUG_ON(vx_info_state(vxi, VXS_HASHED)); ++ ++ vxi->vx_state |= VXS_HASHED; ++ head = &vx_info_hash[__hashval(vxi->vx_id)]; ++ hlist_add_head(&vxi->vx_hlist, head); ++ atomic_inc(&vx_global_cactive); ++} ++ ++/* __unhash_vx_info() ++ ++ * remove the vxi from the global hash table ++ * requires the hash_lock to be held */ ++ ++static inline void __unhash_vx_info(struct vx_info *vxi) ++{ ++ unsigned long flags; ++ ++ vxd_assert_lock(&vx_info_hash_lock); ++ vxdprintk(VXD_CBIT(xid, 4), ++ "__unhash_vx_info: %p[#%d.%d.%d]", vxi, vxi->vx_id, ++ atomic_read(&vxi->vx_usecnt), atomic_read(&vxi->vx_tasks)); ++ vxh_unhash_vx_info(vxi); ++ ++ /* context must be hashed */ ++ BUG_ON(!vx_info_state(vxi, VXS_HASHED)); ++ /* but without tasks */ ++ BUG_ON(atomic_read(&vxi->vx_tasks)); ++ ++ vxi->vx_state &= ~VXS_HASHED; ++ hlist_del_init(&vxi->vx_hlist); ++ spin_lock_irqsave(&vx_info_inactive_lock, flags); ++ hlist_add_head(&vxi->vx_hlist, &vx_info_inactive); ++ spin_unlock_irqrestore(&vx_info_inactive_lock, flags); ++ atomic_dec(&vx_global_cactive); ++} ++ ++ ++/* __lookup_vx_info() ++ ++ * requires the hash_lock to be held ++ * doesn't increment the vx_refcnt */ ++ ++static inline struct vx_info *__lookup_vx_info(xid_t xid) ++{ ++ struct hlist_head *head = &vx_info_hash[__hashval(xid)]; ++ struct hlist_node *pos; ++ struct vx_info *vxi; ++ ++ vxd_assert_lock(&vx_info_hash_lock); ++ hlist_for_each(pos, head) { ++ vxi = hlist_entry(pos, struct vx_info, vx_hlist); ++ ++ if (vxi->vx_id == xid) ++ goto found; ++ } ++ vxi = NULL; ++found: ++ vxdprintk(VXD_CBIT(xid, 0), ++ "__lookup_vx_info(#%u): %p[#%u]", ++ xid, vxi, vxi ? vxi->vx_id : 0); ++ vxh_lookup_vx_info(vxi, xid); ++ return vxi; ++} ++ ++ ++/* __create_vx_info() ++ ++ * create the requested context ++ * get(), claim() and hash it */ ++ ++static struct vx_info *__create_vx_info(int id) ++{ ++ struct vx_info *new, *vxi = NULL; ++ ++ vxdprintk(VXD_CBIT(xid, 1), "create_vx_info(%d)*", id); ++ ++ if (!(new = __alloc_vx_info(id))) ++ return ERR_PTR(-ENOMEM); ++ ++ /* required to make dynamic xids unique */ ++ spin_lock(&vx_info_hash_lock); ++ ++ /* static context requested */ ++ if ((vxi = __lookup_vx_info(id))) { ++ vxdprintk(VXD_CBIT(xid, 0), ++ "create_vx_info(%d) = %p (already there)", id, vxi); ++ if (vx_info_flags(vxi, VXF_STATE_SETUP, 0)) ++ vxi = ERR_PTR(-EBUSY); ++ else ++ vxi = ERR_PTR(-EEXIST); ++ goto out_unlock; ++ } ++ /* new context */ ++ vxdprintk(VXD_CBIT(xid, 0), ++ "create_vx_info(%d) = %p (new)", id, new); ++ claim_vx_info(new, NULL); ++ __hash_vx_info(get_vx_info(new)); ++ vxi = new, new = NULL; ++ ++out_unlock: ++ spin_unlock(&vx_info_hash_lock); ++ vxh_create_vx_info(IS_ERR(vxi) ? NULL : vxi, id); ++ if (new) ++ __dealloc_vx_info(new); ++ return vxi; ++} ++ ++ ++/* exported stuff */ ++ ++ ++void unhash_vx_info(struct vx_info *vxi) ++{ ++ spin_lock(&vx_info_hash_lock); ++ __unhash_vx_info(vxi); ++ spin_unlock(&vx_info_hash_lock); ++ __shutdown_vx_info(vxi); ++ __wakeup_vx_info(vxi); ++} ++ ++ ++/* lookup_vx_info() ++ ++ * search for a vx_info and get() it ++ * negative id means current */ ++ ++struct vx_info *lookup_vx_info(int id) ++{ ++ struct vx_info *vxi = NULL; ++ ++ if (id < 0) { ++ vxi = get_vx_info(current_vx_info()); ++ } else if (id > 1) { ++ spin_lock(&vx_info_hash_lock); ++ vxi = get_vx_info(__lookup_vx_info(id)); ++ spin_unlock(&vx_info_hash_lock); ++ } ++ return vxi; ++} ++ ++/* xid_is_hashed() ++ ++ * verify that xid is still hashed */ ++ ++int xid_is_hashed(xid_t xid) ++{ ++ int hashed; ++ ++ spin_lock(&vx_info_hash_lock); ++ hashed = (__lookup_vx_info(xid) != NULL); ++ spin_unlock(&vx_info_hash_lock); ++ return hashed; ++} ++ ++#ifdef CONFIG_PROC_FS ++ ++/* get_xid_list() ++ ++ * get a subset of hashed xids for proc ++ * assumes size is at least one */ ++ ++int get_xid_list(int index, unsigned int *xids, int size) ++{ ++ int hindex, nr_xids = 0; ++ ++ /* only show current and children */ ++ if (!vx_check(0, VS_ADMIN | VS_WATCH)) { ++ if (index > 0) ++ return 0; ++ xids[nr_xids] = vx_current_xid(); ++ return 1; ++ } ++ ++ for (hindex = 0; hindex < VX_HASH_SIZE; hindex++) { ++ struct hlist_head *head = &vx_info_hash[hindex]; ++ struct hlist_node *pos; ++ ++ spin_lock(&vx_info_hash_lock); ++ hlist_for_each(pos, head) { ++ struct vx_info *vxi; ++ ++ if (--index > 0) ++ continue; ++ ++ vxi = hlist_entry(pos, struct vx_info, vx_hlist); ++ xids[nr_xids] = vxi->vx_id; ++ if (++nr_xids >= size) { ++ spin_unlock(&vx_info_hash_lock); ++ goto out; ++ } ++ } ++ /* keep the lock time short */ ++ spin_unlock(&vx_info_hash_lock); ++ } ++out: ++ return nr_xids; ++} ++#endif ++ ++#ifdef CONFIG_VSERVER_DEBUG ++ ++void dump_vx_info_inactive(int level) ++{ ++ struct hlist_node *entry, *next; ++ ++ hlist_for_each_safe(entry, next, &vx_info_inactive) { ++ struct vx_info *vxi = ++ list_entry(entry, struct vx_info, vx_hlist); ++ ++ dump_vx_info(vxi, level); ++ } ++} ++ ++#endif ++ ++#if 0 ++int vx_migrate_user(struct task_struct *p, struct vx_info *vxi) ++{ ++ struct user_struct *new_user, *old_user; ++ ++ if (!p || !vxi) ++ BUG(); ++ ++ if (vx_info_flags(vxi, VXF_INFO_PRIVATE, 0)) ++ return -EACCES; ++ ++ new_user = alloc_uid(vxi->vx_id, p->uid); ++ if (!new_user) ++ return -ENOMEM; ++ ++ old_user = p->user; ++ if (new_user != old_user) { ++ atomic_inc(&new_user->processes); ++ atomic_dec(&old_user->processes); ++ p->user = new_user; ++ } ++ free_uid(old_user); ++ return 0; ++} ++#endif ++ ++#if 0 ++void vx_mask_cap_bset(struct vx_info *vxi, struct task_struct *p) ++{ ++ // p->cap_effective &= vxi->vx_cap_bset; ++ p->cap_effective = ++ cap_intersect(p->cap_effective, vxi->cap_bset); ++ // p->cap_inheritable &= vxi->vx_cap_bset; ++ p->cap_inheritable = ++ cap_intersect(p->cap_inheritable, vxi->cap_bset); ++ // p->cap_permitted &= vxi->vx_cap_bset; ++ p->cap_permitted = ++ cap_intersect(p->cap_permitted, vxi->cap_bset); ++} ++#endif ++ ++ ++#include <linux/file.h> ++#include <linux/fdtable.h> ++ ++static int vx_openfd_task(struct task_struct *tsk) ++{ ++ struct files_struct *files = tsk->files; ++ struct fdtable *fdt; ++ const unsigned long *bptr; ++ int count, total; ++ ++ /* no rcu_read_lock() because of spin_lock() */ ++ spin_lock(&files->file_lock); ++ fdt = files_fdtable(files); ++ bptr = fdt->open_fds->fds_bits; ++ count = fdt->max_fds / (sizeof(unsigned long) * 8); ++ for (total = 0; count > 0; count--) { ++ if (*bptr) ++ total += hweight_long(*bptr); ++ bptr++; ++ } ++ spin_unlock(&files->file_lock); ++ return total; ++} ++ ++ ++/* for *space compatibility */ ++ ++asmlinkage long sys_unshare(unsigned long); ++ ++/* ++ * migrate task to new context ++ * gets vxi, puts old_vxi on change ++ * optionally unshares namespaces (hack) ++ */ ++ ++int vx_migrate_task(struct task_struct *p, struct vx_info *vxi, int unshare) ++{ ++ struct vx_info *old_vxi; ++ int ret = 0; ++ ++ if (!p || !vxi) ++ BUG(); ++ ++ vxdprintk(VXD_CBIT(xid, 5), ++ "vx_migrate_task(%p,%p[#%d.%d])", p, vxi, ++ vxi->vx_id, atomic_read(&vxi->vx_usecnt)); ++ ++ if (vx_info_flags(vxi, VXF_INFO_PRIVATE, 0) && ++ !vx_info_flags(vxi, VXF_STATE_SETUP, 0)) ++ return -EACCES; ++ ++ if (vx_info_state(vxi, VXS_SHUTDOWN)) ++ return -EFAULT; ++ ++ old_vxi = task_get_vx_info(p); ++ if (old_vxi == vxi) ++ goto out; ++ ++// if (!(ret = vx_migrate_user(p, vxi))) { ++ { ++ int openfd; ++ ++ task_lock(p); ++ openfd = vx_openfd_task(p); ++ ++ if (old_vxi) { ++ atomic_dec(&old_vxi->cvirt.nr_threads); ++ atomic_dec(&old_vxi->cvirt.nr_running); ++ __rlim_dec(&old_vxi->limit, RLIMIT_NPROC); ++ /* FIXME: what about the struct files here? */ ++ __rlim_sub(&old_vxi->limit, VLIMIT_OPENFD, openfd); ++ /* account for the executable */ ++ __rlim_dec(&old_vxi->limit, VLIMIT_DENTRY); ++ } ++ atomic_inc(&vxi->cvirt.nr_threads); ++ atomic_inc(&vxi->cvirt.nr_running); ++ __rlim_inc(&vxi->limit, RLIMIT_NPROC); ++ /* FIXME: what about the struct files here? */ ++ __rlim_add(&vxi->limit, VLIMIT_OPENFD, openfd); ++ /* account for the executable */ ++ __rlim_inc(&vxi->limit, VLIMIT_DENTRY); ++ ++ if (old_vxi) { ++ release_vx_info(old_vxi, p); ++ clr_vx_info(&p->vx_info); ++ } ++ claim_vx_info(vxi, p); ++ set_vx_info(&p->vx_info, vxi); ++ p->xid = vxi->vx_id; ++ ++ vxdprintk(VXD_CBIT(xid, 5), ++ "moved task %p into vxi:%p[#%d]", ++ p, vxi, vxi->vx_id); ++ ++ // vx_mask_cap_bset(vxi, p); ++ task_unlock(p); ++ ++ /* hack for *spaces to provide compatibility */ ++ if (unshare) { ++ struct nsproxy *old_nsp, *new_nsp; ++ ++ ret = unshare_nsproxy_namespaces( ++ CLONE_NEWUTS | CLONE_NEWIPC | CLONE_NEWUSER, ++ &new_nsp, NULL); ++ if (ret) ++ goto out; ++ ++ old_nsp = xchg(&p->nsproxy, new_nsp); ++ vx_set_space(vxi, ++ CLONE_NEWUTS | CLONE_NEWIPC | CLONE_NEWUSER, 0); ++ put_nsproxy(old_nsp); ++ } ++ } ++out: ++ put_vx_info(old_vxi); ++ return ret; ++} ++ ++int vx_set_reaper(struct vx_info *vxi, struct task_struct *p) ++{ ++ struct task_struct *old_reaper; ++ ++ if (!vxi) ++ return -EINVAL; ++ ++ vxdprintk(VXD_CBIT(xid, 6), ++ "vx_set_reaper(%p[#%d],%p[#%d,%d])", ++ vxi, vxi->vx_id, p, p->xid, p->pid); ++ ++ old_reaper = vxi->vx_reaper; ++ if (old_reaper == p) ++ return 0; ++ ++ /* set new child reaper */ ++ get_task_struct(p); ++ vxi->vx_reaper = p; ++ put_task_struct(old_reaper); ++ return 0; ++} ++ ++int vx_set_init(struct vx_info *vxi, struct task_struct *p) ++{ ++ if (!vxi) ++ return -EINVAL; ++ ++ vxdprintk(VXD_CBIT(xid, 6), ++ "vx_set_init(%p[#%d],%p[#%d,%d,%d])", ++ vxi, vxi->vx_id, p, p->xid, p->pid, p->tgid); ++ ++ vxi->vx_flags &= ~VXF_STATE_INIT; ++ // vxi->vx_initpid = p->tgid; ++ vxi->vx_initpid = p->pid; ++ return 0; ++} ++ ++void vx_exit_init(struct vx_info *vxi, struct task_struct *p, int code) ++{ ++ vxdprintk(VXD_CBIT(xid, 6), ++ "vx_exit_init(%p[#%d],%p[#%d,%d,%d])", ++ vxi, vxi->vx_id, p, p->xid, p->pid, p->tgid); ++ ++ vxi->exit_code = code; ++ vxi->vx_initpid = 0; ++} ++ ++ ++void vx_set_persistent(struct vx_info *vxi) ++{ ++ vxdprintk(VXD_CBIT(xid, 6), ++ "vx_set_persistent(%p[#%d])", vxi, vxi->vx_id); ++ ++ get_vx_info(vxi); ++ claim_vx_info(vxi, NULL); ++} ++ ++void vx_clear_persistent(struct vx_info *vxi) ++{ ++ vxdprintk(VXD_CBIT(xid, 6), ++ "vx_clear_persistent(%p[#%d])", vxi, vxi->vx_id); ++ ++ release_vx_info(vxi, NULL); ++ put_vx_info(vxi); ++} ++ ++void vx_update_persistent(struct vx_info *vxi) ++{ ++ if (vx_info_flags(vxi, VXF_PERSISTENT, 0)) ++ vx_set_persistent(vxi); ++ else ++ vx_clear_persistent(vxi); ++} ++ ++ ++/* task must be current or locked */ ++ ++void exit_vx_info(struct task_struct *p, int code) ++{ ++ struct vx_info *vxi = p->vx_info; ++ ++ if (vxi) { ++ atomic_dec(&vxi->cvirt.nr_threads); ++ vx_nproc_dec(p); ++ ++ vxi->exit_code = code; ++ release_vx_info(vxi, p); ++ } ++} ++ ++void exit_vx_info_early(struct task_struct *p, int code) ++{ ++ struct vx_info *vxi = p->vx_info; ++ ++ if (vxi) { ++ if (vxi->vx_initpid == p->pid) ++ vx_exit_init(vxi, p, code); ++ if (vxi->vx_reaper == p) ++ vx_set_reaper(vxi, init_pid_ns.child_reaper); ++ } ++} ++ ++ ++/* vserver syscall commands below here */ ++ ++/* taks xid and vx_info functions */ ++ ++#include <asm/uaccess.h> ++ ++ ++int vc_task_xid(uint32_t id) ++{ ++ xid_t xid; ++ ++ if (id) { ++ struct task_struct *tsk; ++ ++ rcu_read_lock(); ++ tsk = find_task_by_real_pid(id); ++ xid = (tsk) ? tsk->xid : -ESRCH; ++ rcu_read_unlock(); ++ } else ++ xid = vx_current_xid(); ++ return xid; ++} ++ ++ ++int vc_vx_info(struct vx_info *vxi, void __user *data) ++{ ++ struct vcmd_vx_info_v0 vc_data; ++ ++ vc_data.xid = vxi->vx_id; ++ vc_data.initpid = vxi->vx_initpid; ++ ++ if (copy_to_user(data, &vc_data, sizeof(vc_data))) ++ return -EFAULT; ++ return 0; ++} ++ ++ ++int vc_ctx_stat(struct vx_info *vxi, void __user *data) ++{ ++ struct vcmd_ctx_stat_v0 vc_data; ++ ++ vc_data.usecnt = atomic_read(&vxi->vx_usecnt); ++ vc_data.tasks = atomic_read(&vxi->vx_tasks); ++ ++ if (copy_to_user(data, &vc_data, sizeof(vc_data))) ++ return -EFAULT; ++ return 0; ++} ++ ++ ++/* context functions */ ++ ++int vc_ctx_create(uint32_t xid, void __user *data) ++{ ++ struct vcmd_ctx_create vc_data = { .flagword = VXF_INIT_SET }; ++ struct vx_info *new_vxi; ++ int ret; ++ ++ if (data && copy_from_user(&vc_data, data, sizeof(vc_data))) ++ return -EFAULT; ++ ++ if ((xid > MAX_S_CONTEXT) || (xid < 2)) ++ return -EINVAL; ++ ++ new_vxi = __create_vx_info(xid); ++ if (IS_ERR(new_vxi)) ++ return PTR_ERR(new_vxi); ++ ++ /* initial flags */ ++ new_vxi->vx_flags = vc_data.flagword; ++ ++ ret = -ENOEXEC; ++ if (vs_state_change(new_vxi, VSC_STARTUP)) ++ goto out; ++ ++ ret = vx_migrate_task(current, new_vxi, (!data)); ++ if (ret) ++ goto out; ++ ++ /* return context id on success */ ++ ret = new_vxi->vx_id; ++ ++ /* get a reference for persistent contexts */ ++ if ((vc_data.flagword & VXF_PERSISTENT)) ++ vx_set_persistent(new_vxi); ++out: ++ release_vx_info(new_vxi, NULL); ++ put_vx_info(new_vxi); ++ return ret; ++} ++ ++ ++int vc_ctx_migrate(struct vx_info *vxi, void __user *data) ++{ ++ struct vcmd_ctx_migrate vc_data = { .flagword = 0 }; ++ int ret; ++ ++ if (data && copy_from_user(&vc_data, data, sizeof(vc_data))) ++ return -EFAULT; ++ ++ ret = vx_migrate_task(current, vxi, 0); ++ if (ret) ++ return ret; ++ if (vc_data.flagword & VXM_SET_INIT) ++ ret = vx_set_init(vxi, current); ++ if (ret) ++ return ret; ++ if (vc_data.flagword & VXM_SET_REAPER) ++ ret = vx_set_reaper(vxi, current); ++ return ret; ++} ++ ++ ++int vc_get_cflags(struct vx_info *vxi, void __user *data) ++{ ++ struct vcmd_ctx_flags_v0 vc_data; ++ ++ vc_data.flagword = vxi->vx_flags; ++ ++ /* special STATE flag handling */ ++ vc_data.mask = vs_mask_flags(~0ULL, vxi->vx_flags, VXF_ONE_TIME); ++ ++ if (copy_to_user(data, &vc_data, sizeof(vc_data))) ++ return -EFAULT; ++ return 0; ++} ++ ++int vc_set_cflags(struct vx_info *vxi, void __user *data) ++{ ++ struct vcmd_ctx_flags_v0 vc_data; ++ uint64_t mask, trigger; ++ ++ if (copy_from_user(&vc_data, data, sizeof(vc_data))) ++ return -EFAULT; ++ ++ /* special STATE flag handling */ ++ mask = vs_mask_mask(vc_data.mask, vxi->vx_flags, VXF_ONE_TIME); ++ trigger = (mask & vxi->vx_flags) ^ (mask & vc_data.flagword); ++ ++ if (vxi == current_vx_info()) { ++ /* if (trigger & VXF_STATE_SETUP) ++ vx_mask_cap_bset(vxi, current); */ ++ if (trigger & VXF_STATE_INIT) { ++ int ret; ++ ++ ret = vx_set_init(vxi, current); ++ if (ret) ++ return ret; ++ ret = vx_set_reaper(vxi, current); ++ if (ret) ++ return ret; ++ } ++ } ++ ++ vxi->vx_flags = vs_mask_flags(vxi->vx_flags, ++ vc_data.flagword, mask); ++ if (trigger & VXF_PERSISTENT) ++ vx_update_persistent(vxi); ++ ++ return 0; ++} ++ ++ ++static inline uint64_t caps_from_cap_t(kernel_cap_t c) ++{ ++ uint64_t v = c.cap[0] | ((uint64_t)c.cap[1] << 32); ++ ++ // printk("caps_from_cap_t(%08x:%08x) = %016llx\n", c.cap[1], c.cap[0], v); ++ return v; ++} ++ ++static inline kernel_cap_t cap_t_from_caps(uint64_t v) ++{ ++ kernel_cap_t c = __cap_empty_set; ++ ++ c.cap[0] = v & 0xFFFFFFFF; ++ c.cap[1] = (v >> 32) & 0xFFFFFFFF; ++ ++ // printk("cap_t_from_caps(%016llx) = %08x:%08x\n", v, c.cap[1], c.cap[0]); ++ return c; ++} ++ ++ ++static int do_get_caps(struct vx_info *vxi, uint64_t *bcaps, uint64_t *ccaps) ++{ ++ if (bcaps) ++ *bcaps = caps_from_cap_t(vxi->vx_bcaps); ++ if (ccaps) ++ *ccaps = vxi->vx_ccaps; ++ ++ return 0; ++} ++ ++int vc_get_ccaps(struct vx_info *vxi, void __user *data) ++{ ++ struct vcmd_ctx_caps_v1 vc_data; ++ int ret; ++ ++ ret = do_get_caps(vxi, NULL, &vc_data.ccaps); ++ if (ret) ++ return ret; ++ vc_data.cmask = ~0ULL; ++ ++ if (copy_to_user(data, &vc_data, sizeof(vc_data))) ++ return -EFAULT; ++ return 0; ++} ++ ++static int do_set_caps(struct vx_info *vxi, ++ uint64_t bcaps, uint64_t bmask, uint64_t ccaps, uint64_t cmask) ++{ ++ uint64_t bcold = caps_from_cap_t(vxi->vx_bcaps); ++ ++#if 0 ++ printk("do_set_caps(%16llx, %16llx, %16llx, %16llx)\n", ++ bcaps, bmask, ccaps, cmask); ++#endif ++ vxi->vx_bcaps = cap_t_from_caps( ++ vs_mask_flags(bcold, bcaps, bmask)); ++ vxi->vx_ccaps = vs_mask_flags(vxi->vx_ccaps, ccaps, cmask); ++ ++ return 0; ++} ++ ++int vc_set_ccaps(struct vx_info *vxi, void __user *data) ++{ ++ struct vcmd_ctx_caps_v1 vc_data; ++ ++ if (copy_from_user(&vc_data, data, sizeof(vc_data))) ++ return -EFAULT; ++ ++ return do_set_caps(vxi, 0, 0, vc_data.ccaps, vc_data.cmask); ++} ++ ++int vc_get_bcaps(struct vx_info *vxi, void __user *data) ++{ ++ struct vcmd_bcaps vc_data; ++ int ret; ++ ++ ret = do_get_caps(vxi, &vc_data.bcaps, NULL); ++ if (ret) ++ return ret; ++ vc_data.bmask = ~0ULL; ++ ++ if (copy_to_user(data, &vc_data, sizeof(vc_data))) ++ return -EFAULT; ++ return 0; ++} ++ ++int vc_set_bcaps(struct vx_info *vxi, void __user *data) ++{ ++ struct vcmd_bcaps vc_data; ++ ++ if (copy_from_user(&vc_data, data, sizeof(vc_data))) ++ return -EFAULT; ++ ++ return do_set_caps(vxi, vc_data.bcaps, vc_data.bmask, 0, 0); ++} ++ ++ ++int vc_get_umask(struct vx_info *vxi, void __user *data) ++{ ++ struct vcmd_umask vc_data; ++ ++ vc_data.umask = vxi->vx_umask; ++ vc_data.mask = ~0ULL; ++ ++ if (copy_to_user(data, &vc_data, sizeof(vc_data))) ++ return -EFAULT; ++ return 0; ++} ++ ++int vc_set_umask(struct vx_info *vxi, void __user *data) ++{ ++ struct vcmd_umask vc_data; ++ ++ if (copy_from_user(&vc_data, data, sizeof(vc_data))) ++ return -EFAULT; ++ ++ vxi->vx_umask = vs_mask_flags(vxi->vx_umask, ++ vc_data.umask, vc_data.mask); ++ return 0; ++} ++ ++ ++int vc_get_wmask(struct vx_info *vxi, void __user *data) ++{ ++ struct vcmd_wmask vc_data; ++ ++ vc_data.wmask = vxi->vx_wmask; ++ vc_data.mask = ~0ULL; ++ ++ if (copy_to_user(data, &vc_data, sizeof(vc_data))) ++ return -EFAULT; ++ return 0; ++} ++ ++int vc_set_wmask(struct vx_info *vxi, void __user *data) ++{ ++ struct vcmd_wmask vc_data; ++ ++ if (copy_from_user(&vc_data, data, sizeof(vc_data))) ++ return -EFAULT; ++ ++ vxi->vx_wmask = vs_mask_flags(vxi->vx_wmask, ++ vc_data.wmask, vc_data.mask); ++ return 0; ++} ++ ++ ++int vc_get_badness(struct vx_info *vxi, void __user *data) ++{ ++ struct vcmd_badness_v0 vc_data; ++ ++ vc_data.bias = vxi->vx_badness_bias; ++ ++ if (copy_to_user(data, &vc_data, sizeof(vc_data))) ++ return -EFAULT; ++ return 0; ++} ++ ++int vc_set_badness(struct vx_info *vxi, void __user *data) ++{ ++ struct vcmd_badness_v0 vc_data; ++ ++ if (copy_from_user(&vc_data, data, sizeof(vc_data))) ++ return -EFAULT; ++ ++ vxi->vx_badness_bias = vc_data.bias; ++ return 0; ++} ++ ++#include <linux/module.h> ++ ++EXPORT_SYMBOL_GPL(free_vx_info); ++ +diff -NurpP --minimal linux-3.0.9/kernel/vserver/cvirt.c linux-3.0.9-vs2.3.2.1/kernel/vserver/cvirt.c +--- linux-3.0.9/kernel/vserver/cvirt.c 1970-01-01 01:00:00.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/kernel/vserver/cvirt.c 2011-06-13 14:57:26.000000000 +0200 +@@ -0,0 +1,303 @@ ++/* ++ * linux/kernel/vserver/cvirt.c ++ * ++ * Virtual Server: Context Virtualization ++ * ++ * Copyright (C) 2004-2007 Herbert Pötzl ++ * ++ * V0.01 broken out from limit.c ++ * V0.02 added utsname stuff ++ * V0.03 changed vcmds to vxi arg ++ * ++ */ ++ ++#include <linux/types.h> ++#include <linux/utsname.h> ++#include <linux/vs_cvirt.h> ++#include <linux/vserver/switch.h> ++#include <linux/vserver/cvirt_cmd.h> ++ ++#include <asm/uaccess.h> ++ ++ ++void vx_vsi_uptime(struct timespec *uptime, struct timespec *idle) ++{ ++ struct vx_info *vxi = current_vx_info(); ++ ++ set_normalized_timespec(uptime, ++ uptime->tv_sec - vxi->cvirt.bias_uptime.tv_sec, ++ uptime->tv_nsec - vxi->cvirt.bias_uptime.tv_nsec); ++ if (!idle) ++ return; ++ set_normalized_timespec(idle, ++ idle->tv_sec - vxi->cvirt.bias_idle.tv_sec, ++ idle->tv_nsec - vxi->cvirt.bias_idle.tv_nsec); ++ return; ++} ++ ++uint64_t vx_idle_jiffies(void) ++{ ++ return init_task.utime + init_task.stime; ++} ++ ++ ++ ++static inline uint32_t __update_loadavg(uint32_t load, ++ int wsize, int delta, int n) ++{ ++ unsigned long long calc, prev; ++ ++ /* just set it to n */ ++ if (unlikely(delta >= wsize)) ++ return (n << FSHIFT); ++ ++ calc = delta * n; ++ calc <<= FSHIFT; ++ prev = (wsize - delta); ++ prev *= load; ++ calc += prev; ++ do_div(calc, wsize); ++ return calc; ++} ++ ++ ++void vx_update_load(struct vx_info *vxi) ++{ ++ uint32_t now, last, delta; ++ unsigned int nr_running, nr_uninterruptible; ++ unsigned int total; ++ unsigned long flags; ++ ++ spin_lock_irqsave(&vxi->cvirt.load_lock, flags); ++ ++ now = jiffies; ++ last = vxi->cvirt.load_last; ++ delta = now - last; ++ ++ if (delta < 5*HZ) ++ goto out; ++ ++ nr_running = atomic_read(&vxi->cvirt.nr_running); ++ nr_uninterruptible = atomic_read(&vxi->cvirt.nr_uninterruptible); ++ total = nr_running + nr_uninterruptible; ++ ++ vxi->cvirt.load[0] = __update_loadavg(vxi->cvirt.load[0], ++ 60*HZ, delta, total); ++ vxi->cvirt.load[1] = __update_loadavg(vxi->cvirt.load[1], ++ 5*60*HZ, delta, total); ++ vxi->cvirt.load[2] = __update_loadavg(vxi->cvirt.load[2], ++ 15*60*HZ, delta, total); ++ ++ vxi->cvirt.load_last = now; ++out: ++ atomic_inc(&vxi->cvirt.load_updates); ++ spin_unlock_irqrestore(&vxi->cvirt.load_lock, flags); ++} ++ ++ ++/* ++ * Commands to do_syslog: ++ * ++ * 0 -- Close the log. Currently a NOP. ++ * 1 -- Open the log. Currently a NOP. ++ * 2 -- Read from the log. ++ * 3 -- Read all messages remaining in the ring buffer. ++ * 4 -- Read and clear all messages remaining in the ring buffer ++ * 5 -- Clear ring buffer. ++ * 6 -- Disable printk's to console ++ * 7 -- Enable printk's to console ++ * 8 -- Set level of messages printed to console ++ * 9 -- Return number of unread characters in the log buffer ++ * 10 -- Return size of the log buffer ++ */ ++int vx_do_syslog(int type, char __user *buf, int len) ++{ ++ int error = 0; ++ int do_clear = 0; ++ struct vx_info *vxi = current_vx_info(); ++ struct _vx_syslog *log; ++ ++ if (!vxi) ++ return -EINVAL; ++ log = &vxi->cvirt.syslog; ++ ++ switch (type) { ++ case 0: /* Close log */ ++ case 1: /* Open log */ ++ break; ++ case 2: /* Read from log */ ++ error = wait_event_interruptible(log->log_wait, ++ (log->log_start - log->log_end)); ++ if (error) ++ break; ++ spin_lock_irq(&log->logbuf_lock); ++ spin_unlock_irq(&log->logbuf_lock); ++ break; ++ case 4: /* Read/clear last kernel messages */ ++ do_clear = 1; ++ /* fall through */ ++ case 3: /* Read last kernel messages */ ++ return 0; ++ ++ case 5: /* Clear ring buffer */ ++ return 0; ++ ++ case 6: /* Disable logging to console */ ++ case 7: /* Enable logging to console */ ++ case 8: /* Set level of messages printed to console */ ++ break; ++ ++ case 9: /* Number of chars in the log buffer */ ++ return 0; ++ case 10: /* Size of the log buffer */ ++ return 0; ++ default: ++ error = -EINVAL; ++ break; ++ } ++ return error; ++} ++ ++ ++/* virtual host info names */ ++ ++static char *vx_vhi_name(struct vx_info *vxi, int id) ++{ ++ struct nsproxy *nsproxy; ++ struct uts_namespace *uts; ++ ++ if (id == VHIN_CONTEXT) ++ return vxi->vx_name; ++ ++ nsproxy = vxi->space[0].vx_nsproxy; ++ if (!nsproxy) ++ return NULL; ++ ++ uts = nsproxy->uts_ns; ++ if (!uts) ++ return NULL; ++ ++ switch (id) { ++ case VHIN_SYSNAME: ++ return uts->name.sysname; ++ case VHIN_NODENAME: ++ return uts->name.nodename; ++ case VHIN_RELEASE: ++ return uts->name.release; ++ case VHIN_VERSION: ++ return uts->name.version; ++ case VHIN_MACHINE: ++ return uts->name.machine; ++ case VHIN_DOMAINNAME: ++ return uts->name.domainname; ++ default: ++ return NULL; ++ } ++ return NULL; ++} ++ ++int vc_set_vhi_name(struct vx_info *vxi, void __user *data) ++{ ++ struct vcmd_vhi_name_v0 vc_data; ++ char *name; ++ ++ if (copy_from_user(&vc_data, data, sizeof(vc_data))) ++ return -EFAULT; ++ ++ name = vx_vhi_name(vxi, vc_data.field); ++ if (!name) ++ return -EINVAL; ++ ++ memcpy(name, vc_data.name, 65); ++ return 0; ++} ++ ++int vc_get_vhi_name(struct vx_info *vxi, void __user *data) ++{ ++ struct vcmd_vhi_name_v0 vc_data; ++ char *name; ++ ++ if (copy_from_user(&vc_data, data, sizeof(vc_data))) ++ return -EFAULT; ++ ++ name = vx_vhi_name(vxi, vc_data.field); ++ if (!name) ++ return -EINVAL; ++ ++ memcpy(vc_data.name, name, 65); ++ if (copy_to_user(data, &vc_data, sizeof(vc_data))) ++ return -EFAULT; ++ return 0; ++} ++ ++ ++int vc_virt_stat(struct vx_info *vxi, void __user *data) ++{ ++ struct vcmd_virt_stat_v0 vc_data; ++ struct _vx_cvirt *cvirt = &vxi->cvirt; ++ struct timespec uptime; ++ ++ do_posix_clock_monotonic_gettime(&uptime); ++ set_normalized_timespec(&uptime, ++ uptime.tv_sec - cvirt->bias_uptime.tv_sec, ++ uptime.tv_nsec - cvirt->bias_uptime.tv_nsec); ++ ++ vc_data.offset = timespec_to_ns(&cvirt->bias_ts); ++ vc_data.uptime = timespec_to_ns(&uptime); ++ vc_data.nr_threads = atomic_read(&cvirt->nr_threads); ++ vc_data.nr_running = atomic_read(&cvirt->nr_running); ++ vc_data.nr_uninterruptible = atomic_read(&cvirt->nr_uninterruptible); ++ vc_data.nr_onhold = atomic_read(&cvirt->nr_onhold); ++ vc_data.nr_forks = atomic_read(&cvirt->total_forks); ++ vc_data.load[0] = cvirt->load[0]; ++ vc_data.load[1] = cvirt->load[1]; ++ vc_data.load[2] = cvirt->load[2]; ++ ++ if (copy_to_user(data, &vc_data, sizeof(vc_data))) ++ return -EFAULT; ++ return 0; ++} ++ ++ ++#ifdef CONFIG_VSERVER_VTIME ++ ++/* virtualized time base */ ++ ++void vx_adjust_timespec(struct timespec *ts) ++{ ++ struct vx_info *vxi; ++ ++ if (!vx_flags(VXF_VIRT_TIME, 0)) ++ return; ++ ++ vxi = current_vx_info(); ++ ts->tv_sec += vxi->cvirt.bias_ts.tv_sec; ++ ts->tv_nsec += vxi->cvirt.bias_ts.tv_nsec; ++ ++ if (ts->tv_nsec >= NSEC_PER_SEC) { ++ ts->tv_sec++; ++ ts->tv_nsec -= NSEC_PER_SEC; ++ } else if (ts->tv_nsec < 0) { ++ ts->tv_sec--; ++ ts->tv_nsec += NSEC_PER_SEC; ++ } ++} ++ ++int vx_settimeofday(const struct timespec *ts) ++{ ++ struct timespec ats, delta; ++ struct vx_info *vxi; ++ ++ if (!vx_flags(VXF_VIRT_TIME, 0)) ++ return do_settimeofday(ts); ++ ++ getnstimeofday(&ats); ++ delta = timespec_sub(*ts, ats); ++ ++ vxi = current_vx_info(); ++ vxi->cvirt.bias_ts = timespec_add(vxi->cvirt.bias_ts, delta); ++ return 0; ++} ++ ++#endif ++ +diff -NurpP --minimal linux-3.0.9/kernel/vserver/cvirt_init.h linux-3.0.9-vs2.3.2.1/kernel/vserver/cvirt_init.h +--- linux-3.0.9/kernel/vserver/cvirt_init.h 1970-01-01 01:00:00.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/kernel/vserver/cvirt_init.h 2011-06-10 22:11:24.000000000 +0200 +@@ -0,0 +1,70 @@ ++ ++ ++extern uint64_t vx_idle_jiffies(void); ++ ++static inline void vx_info_init_cvirt(struct _vx_cvirt *cvirt) ++{ ++ uint64_t idle_jiffies = vx_idle_jiffies(); ++ uint64_t nsuptime; ++ ++ do_posix_clock_monotonic_gettime(&cvirt->bias_uptime); ++ nsuptime = (unsigned long long)cvirt->bias_uptime.tv_sec ++ * NSEC_PER_SEC + cvirt->bias_uptime.tv_nsec; ++ cvirt->bias_clock = nsec_to_clock_t(nsuptime); ++ cvirt->bias_ts.tv_sec = 0; ++ cvirt->bias_ts.tv_nsec = 0; ++ ++ jiffies_to_timespec(idle_jiffies, &cvirt->bias_idle); ++ atomic_set(&cvirt->nr_threads, 0); ++ atomic_set(&cvirt->nr_running, 0); ++ atomic_set(&cvirt->nr_uninterruptible, 0); ++ atomic_set(&cvirt->nr_onhold, 0); ++ ++ spin_lock_init(&cvirt->load_lock); ++ cvirt->load_last = jiffies; ++ atomic_set(&cvirt->load_updates, 0); ++ cvirt->load[0] = 0; ++ cvirt->load[1] = 0; ++ cvirt->load[2] = 0; ++ atomic_set(&cvirt->total_forks, 0); ++ ++ spin_lock_init(&cvirt->syslog.logbuf_lock); ++ init_waitqueue_head(&cvirt->syslog.log_wait); ++ cvirt->syslog.log_start = 0; ++ cvirt->syslog.log_end = 0; ++ cvirt->syslog.con_start = 0; ++ cvirt->syslog.logged_chars = 0; ++} ++ ++static inline ++void vx_info_init_cvirt_pc(struct _vx_cvirt_pc *cvirt_pc, int cpu) ++{ ++ // cvirt_pc->cpustat = { 0 }; ++} ++ ++static inline void vx_info_exit_cvirt(struct _vx_cvirt *cvirt) ++{ ++#ifdef CONFIG_VSERVER_WARN ++ int value; ++#endif ++ vxwprintk_xid((value = atomic_read(&cvirt->nr_threads)), ++ "!!! cvirt: %p[nr_threads] = %d on exit.", ++ cvirt, value); ++ vxwprintk_xid((value = atomic_read(&cvirt->nr_running)), ++ "!!! cvirt: %p[nr_running] = %d on exit.", ++ cvirt, value); ++ vxwprintk_xid((value = atomic_read(&cvirt->nr_uninterruptible)), ++ "!!! cvirt: %p[nr_uninterruptible] = %d on exit.", ++ cvirt, value); ++ vxwprintk_xid((value = atomic_read(&cvirt->nr_onhold)), ++ "!!! cvirt: %p[nr_onhold] = %d on exit.", ++ cvirt, value); ++ return; ++} ++ ++static inline ++void vx_info_exit_cvirt_pc(struct _vx_cvirt_pc *cvirt_pc, int cpu) ++{ ++ return; ++} ++ +diff -NurpP --minimal linux-3.0.9/kernel/vserver/cvirt_proc.h linux-3.0.9-vs2.3.2.1/kernel/vserver/cvirt_proc.h +--- linux-3.0.9/kernel/vserver/cvirt_proc.h 1970-01-01 01:00:00.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/kernel/vserver/cvirt_proc.h 2011-06-10 22:11:24.000000000 +0200 +@@ -0,0 +1,135 @@ ++#ifndef _VX_CVIRT_PROC_H ++#define _VX_CVIRT_PROC_H ++ ++#include <linux/nsproxy.h> ++#include <linux/mnt_namespace.h> ++#include <linux/ipc_namespace.h> ++#include <linux/utsname.h> ++#include <linux/ipc.h> ++ ++ ++static inline ++int vx_info_proc_nsproxy(struct nsproxy *nsproxy, char *buffer) ++{ ++ struct mnt_namespace *ns; ++ struct uts_namespace *uts; ++ struct ipc_namespace *ipc; ++ struct path path; ++ char *pstr, *root; ++ int length = 0; ++ ++ if (!nsproxy) ++ goto out; ++ ++ length += sprintf(buffer + length, ++ "NSProxy:\t%p [%p,%p,%p]\n", ++ nsproxy, nsproxy->mnt_ns, ++ nsproxy->uts_ns, nsproxy->ipc_ns); ++ ++ ns = nsproxy->mnt_ns; ++ if (!ns) ++ goto skip_ns; ++ ++ pstr = kmalloc(PATH_MAX, GFP_KERNEL); ++ if (!pstr) ++ goto skip_ns; ++ ++ path.mnt = ns->root; ++ path.dentry = ns->root->mnt_root; ++ root = d_path(&path, pstr, PATH_MAX - 2); ++ length += sprintf(buffer + length, ++ "Namespace:\t%p [#%u]\n" ++ "RootPath:\t%s\n", ++ ns, atomic_read(&ns->count), ++ root); ++ kfree(pstr); ++skip_ns: ++ ++ uts = nsproxy->uts_ns; ++ if (!uts) ++ goto skip_uts; ++ ++ length += sprintf(buffer + length, ++ "SysName:\t%.*s\n" ++ "NodeName:\t%.*s\n" ++ "Release:\t%.*s\n" ++ "Version:\t%.*s\n" ++ "Machine:\t%.*s\n" ++ "DomainName:\t%.*s\n", ++ __NEW_UTS_LEN, uts->name.sysname, ++ __NEW_UTS_LEN, uts->name.nodename, ++ __NEW_UTS_LEN, uts->name.release, ++ __NEW_UTS_LEN, uts->name.version, ++ __NEW_UTS_LEN, uts->name.machine, ++ __NEW_UTS_LEN, uts->name.domainname); ++skip_uts: ++ ++ ipc = nsproxy->ipc_ns; ++ if (!ipc) ++ goto skip_ipc; ++ ++ length += sprintf(buffer + length, ++ "SEMS:\t\t%d %d %d %d %d\n" ++ "MSG:\t\t%d %d %d\n" ++ "SHM:\t\t%lu %lu %d %d\n", ++ ipc->sem_ctls[0], ipc->sem_ctls[1], ++ ipc->sem_ctls[2], ipc->sem_ctls[3], ++ ipc->used_sems, ++ ipc->msg_ctlmax, ipc->msg_ctlmnb, ipc->msg_ctlmni, ++ (unsigned long)ipc->shm_ctlmax, ++ (unsigned long)ipc->shm_ctlall, ++ ipc->shm_ctlmni, ipc->shm_tot); ++skip_ipc: ++out: ++ return length; ++} ++ ++ ++#include <linux/sched.h> ++ ++#define LOAD_INT(x) ((x) >> FSHIFT) ++#define LOAD_FRAC(x) LOAD_INT(((x) & (FIXED_1 - 1)) * 100) ++ ++static inline ++int vx_info_proc_cvirt(struct _vx_cvirt *cvirt, char *buffer) ++{ ++ int length = 0; ++ int a, b, c; ++ ++ length += sprintf(buffer + length, ++ "BiasUptime:\t%lu.%02lu\n", ++ (unsigned long)cvirt->bias_uptime.tv_sec, ++ (cvirt->bias_uptime.tv_nsec / (NSEC_PER_SEC / 100))); ++ ++ a = cvirt->load[0] + (FIXED_1 / 200); ++ b = cvirt->load[1] + (FIXED_1 / 200); ++ c = cvirt->load[2] + (FIXED_1 / 200); ++ length += sprintf(buffer + length, ++ "nr_threads:\t%d\n" ++ "nr_running:\t%d\n" ++ "nr_unintr:\t%d\n" ++ "nr_onhold:\t%d\n" ++ "load_updates:\t%d\n" ++ "loadavg:\t%d.%02d %d.%02d %d.%02d\n" ++ "total_forks:\t%d\n", ++ atomic_read(&cvirt->nr_threads), ++ atomic_read(&cvirt->nr_running), ++ atomic_read(&cvirt->nr_uninterruptible), ++ atomic_read(&cvirt->nr_onhold), ++ atomic_read(&cvirt->load_updates), ++ LOAD_INT(a), LOAD_FRAC(a), ++ LOAD_INT(b), LOAD_FRAC(b), ++ LOAD_INT(c), LOAD_FRAC(c), ++ atomic_read(&cvirt->total_forks)); ++ return length; ++} ++ ++static inline ++int vx_info_proc_cvirt_pc(struct _vx_cvirt_pc *cvirt_pc, ++ char *buffer, int cpu) ++{ ++ int length = 0; ++ return length; ++} ++ ++#endif /* _VX_CVIRT_PROC_H */ +diff -NurpP --minimal linux-3.0.9/kernel/vserver/debug.c linux-3.0.9-vs2.3.2.1/kernel/vserver/debug.c +--- linux-3.0.9/kernel/vserver/debug.c 1970-01-01 01:00:00.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/kernel/vserver/debug.c 2011-06-10 22:11:24.000000000 +0200 +@@ -0,0 +1,32 @@ ++/* ++ * kernel/vserver/debug.c ++ * ++ * Copyright (C) 2005-2007 Herbert Pötzl ++ * ++ * V0.01 vx_info dump support ++ * ++ */ ++ ++#include <linux/module.h> ++ ++#include <linux/vserver/context.h> ++ ++ ++void dump_vx_info(struct vx_info *vxi, int level) ++{ ++ printk("vx_info %p[#%d, %d.%d, %4x]\n", vxi, vxi->vx_id, ++ atomic_read(&vxi->vx_usecnt), ++ atomic_read(&vxi->vx_tasks), ++ vxi->vx_state); ++ if (level > 0) { ++ __dump_vx_limit(&vxi->limit); ++ __dump_vx_sched(&vxi->sched); ++ __dump_vx_cvirt(&vxi->cvirt); ++ __dump_vx_cacct(&vxi->cacct); ++ } ++ printk("---\n"); ++} ++ ++ ++EXPORT_SYMBOL_GPL(dump_vx_info); ++ +diff -NurpP --minimal linux-3.0.9/kernel/vserver/device.c linux-3.0.9-vs2.3.2.1/kernel/vserver/device.c +--- linux-3.0.9/kernel/vserver/device.c 1970-01-01 01:00:00.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/kernel/vserver/device.c 2011-06-10 23:20:56.000000000 +0200 +@@ -0,0 +1,443 @@ ++/* ++ * linux/kernel/vserver/device.c ++ * ++ * Linux-VServer: Device Support ++ * ++ * Copyright (C) 2006 Herbert Pötzl ++ * Copyright (C) 2007 Daniel Hokka Zakrisson ++ * ++ * V0.01 device mapping basics ++ * V0.02 added defaults ++ * ++ */ ++ ++#include <linux/slab.h> ++#include <linux/rcupdate.h> ++#include <linux/fs.h> ++#include <linux/namei.h> ++#include <linux/hash.h> ++ ++#include <asm/errno.h> ++#include <asm/uaccess.h> ++#include <linux/vserver/base.h> ++#include <linux/vserver/debug.h> ++#include <linux/vserver/context.h> ++#include <linux/vserver/device.h> ++#include <linux/vserver/device_cmd.h> ++ ++ ++#define DMAP_HASH_BITS 4 ++ ++ ++struct vs_mapping { ++ union { ++ struct hlist_node hlist; ++ struct list_head list; ++ } u; ++#define dm_hlist u.hlist ++#define dm_list u.list ++ xid_t xid; ++ dev_t device; ++ struct vx_dmap_target target; ++}; ++ ++ ++static struct hlist_head dmap_main_hash[1 << DMAP_HASH_BITS]; ++ ++static DEFINE_SPINLOCK(dmap_main_hash_lock); ++ ++static struct vx_dmap_target dmap_defaults[2] = { ++ { .flags = DATTR_OPEN }, ++ { .flags = DATTR_OPEN }, ++}; ++ ++ ++struct kmem_cache *dmap_cachep __read_mostly; ++ ++int __init dmap_cache_init(void) ++{ ++ dmap_cachep = kmem_cache_create("dmap_cache", ++ sizeof(struct vs_mapping), 0, ++ SLAB_HWCACHE_ALIGN|SLAB_PANIC, NULL); ++ return 0; ++} ++ ++__initcall(dmap_cache_init); ++ ++ ++static inline unsigned int __hashval(dev_t dev, int bits) ++{ ++ return hash_long((unsigned long)dev, bits); ++} ++ ++ ++/* __hash_mapping() ++ * add the mapping to the hash table ++ */ ++static inline void __hash_mapping(struct vx_info *vxi, struct vs_mapping *vdm) ++{ ++ spinlock_t *hash_lock = &dmap_main_hash_lock; ++ struct hlist_head *head, *hash = dmap_main_hash; ++ int device = vdm->device; ++ ++ spin_lock(hash_lock); ++ vxdprintk(VXD_CBIT(misc, 8), "__hash_mapping: %p[#%d] %08x:%08x", ++ vxi, vxi ? vxi->vx_id : 0, device, vdm->target.target); ++ ++ head = &hash[__hashval(device, DMAP_HASH_BITS)]; ++ hlist_add_head(&vdm->dm_hlist, head); ++ spin_unlock(hash_lock); ++} ++ ++ ++static inline int __mode_to_default(umode_t mode) ++{ ++ switch (mode) { ++ case S_IFBLK: ++ return 0; ++ case S_IFCHR: ++ return 1; ++ default: ++ BUG(); ++ } ++} ++ ++ ++/* __set_default() ++ * set a default ++ */ ++static inline void __set_default(struct vx_info *vxi, umode_t mode, ++ struct vx_dmap_target *vdmt) ++{ ++ spinlock_t *hash_lock = &dmap_main_hash_lock; ++ spin_lock(hash_lock); ++ ++ if (vxi) ++ vxi->dmap.targets[__mode_to_default(mode)] = *vdmt; ++ else ++ dmap_defaults[__mode_to_default(mode)] = *vdmt; ++ ++ ++ spin_unlock(hash_lock); ++ ++ vxdprintk(VXD_CBIT(misc, 8), "__set_default: %p[#%u] %08x %04x", ++ vxi, vxi ? vxi->vx_id : 0, vdmt->target, vdmt->flags); ++} ++ ++ ++/* __remove_default() ++ * remove a default ++ */ ++static inline int __remove_default(struct vx_info *vxi, umode_t mode) ++{ ++ spinlock_t *hash_lock = &dmap_main_hash_lock; ++ spin_lock(hash_lock); ++ ++ if (vxi) ++ vxi->dmap.targets[__mode_to_default(mode)].flags = 0; ++ else /* remove == reset */ ++ dmap_defaults[__mode_to_default(mode)].flags = DATTR_OPEN | mode; ++ ++ spin_unlock(hash_lock); ++ return 0; ++} ++ ++ ++/* __find_mapping() ++ * find a mapping in the hash table ++ * ++ * caller must hold hash_lock ++ */ ++static inline int __find_mapping(xid_t xid, dev_t device, umode_t mode, ++ struct vs_mapping **local, struct vs_mapping **global) ++{ ++ struct hlist_head *hash = dmap_main_hash; ++ struct hlist_head *head = &hash[__hashval(device, DMAP_HASH_BITS)]; ++ struct hlist_node *pos; ++ struct vs_mapping *vdm; ++ ++ *local = NULL; ++ if (global) ++ *global = NULL; ++ ++ hlist_for_each(pos, head) { ++ vdm = hlist_entry(pos, struct vs_mapping, dm_hlist); ++ ++ if ((vdm->device == device) && ++ !((vdm->target.flags ^ mode) & S_IFMT)) { ++ if (vdm->xid == xid) { ++ *local = vdm; ++ return 1; ++ } else if (global && vdm->xid == 0) ++ *global = vdm; ++ } ++ } ++ ++ if (global && *global) ++ return 0; ++ else ++ return -ENOENT; ++} ++ ++ ++/* __lookup_mapping() ++ * find a mapping and store the result in target and flags ++ */ ++static inline int __lookup_mapping(struct vx_info *vxi, ++ dev_t device, dev_t *target, int *flags, umode_t mode) ++{ ++ spinlock_t *hash_lock = &dmap_main_hash_lock; ++ struct vs_mapping *vdm, *global; ++ struct vx_dmap_target *vdmt; ++ int ret = 0; ++ xid_t xid = vxi->vx_id; ++ int index; ++ ++ spin_lock(hash_lock); ++ if (__find_mapping(xid, device, mode, &vdm, &global) > 0) { ++ ret = 1; ++ vdmt = &vdm->target; ++ goto found; ++ } ++ ++ index = __mode_to_default(mode); ++ if (vxi && vxi->dmap.targets[index].flags) { ++ ret = 2; ++ vdmt = &vxi->dmap.targets[index]; ++ } else if (global) { ++ ret = 3; ++ vdmt = &global->target; ++ goto found; ++ } else { ++ ret = 4; ++ vdmt = &dmap_defaults[index]; ++ } ++ ++found: ++ if (target && (vdmt->flags & DATTR_REMAP)) ++ *target = vdmt->target; ++ else if (target) ++ *target = device; ++ if (flags) ++ *flags = vdmt->flags; ++ ++ spin_unlock(hash_lock); ++ ++ return ret; ++} ++ ++ ++/* __remove_mapping() ++ * remove a mapping from the hash table ++ */ ++static inline int __remove_mapping(struct vx_info *vxi, dev_t device, ++ umode_t mode) ++{ ++ spinlock_t *hash_lock = &dmap_main_hash_lock; ++ struct vs_mapping *vdm = NULL; ++ int ret = 0; ++ ++ spin_lock(hash_lock); ++ ++ ret = __find_mapping((vxi ? vxi->vx_id : 0), device, mode, &vdm, ++ NULL); ++ vxdprintk(VXD_CBIT(misc, 8), "__remove_mapping: %p[#%d] %08x %04x", ++ vxi, vxi ? vxi->vx_id : 0, device, mode); ++ if (ret < 0) ++ goto out; ++ hlist_del(&vdm->dm_hlist); ++ ++out: ++ spin_unlock(hash_lock); ++ if (vdm) ++ kmem_cache_free(dmap_cachep, vdm); ++ return ret; ++} ++ ++ ++ ++int vs_map_device(struct vx_info *vxi, ++ dev_t device, dev_t *target, umode_t mode) ++{ ++ int ret, flags = DATTR_MASK; ++ ++ if (!vxi) { ++ if (target) ++ *target = device; ++ goto out; ++ } ++ ret = __lookup_mapping(vxi, device, target, &flags, mode); ++ vxdprintk(VXD_CBIT(misc, 8), "vs_map_device: %08x target: %08x flags: %04x mode: %04x mapped=%d", ++ device, target ? *target : 0, flags, mode, ret); ++out: ++ return (flags & DATTR_MASK); ++} ++ ++ ++ ++static int do_set_mapping(struct vx_info *vxi, ++ dev_t device, dev_t target, int flags, umode_t mode) ++{ ++ if (device) { ++ struct vs_mapping *new; ++ ++ new = kmem_cache_alloc(dmap_cachep, GFP_KERNEL); ++ if (!new) ++ return -ENOMEM; ++ ++ INIT_HLIST_NODE(&new->dm_hlist); ++ new->device = device; ++ new->target.target = target; ++ new->target.flags = flags | mode; ++ new->xid = (vxi ? vxi->vx_id : 0); ++ ++ vxdprintk(VXD_CBIT(misc, 8), "do_set_mapping: %08x target: %08x flags: %04x", device, target, flags); ++ __hash_mapping(vxi, new); ++ } else { ++ struct vx_dmap_target new = { ++ .target = target, ++ .flags = flags | mode, ++ }; ++ __set_default(vxi, mode, &new); ++ } ++ return 0; ++} ++ ++ ++static int do_unset_mapping(struct vx_info *vxi, ++ dev_t device, dev_t target, int flags, umode_t mode) ++{ ++ int ret = -EINVAL; ++ ++ if (device) { ++ ret = __remove_mapping(vxi, device, mode); ++ if (ret < 0) ++ goto out; ++ } else { ++ ret = __remove_default(vxi, mode); ++ if (ret < 0) ++ goto out; ++ } ++ ++out: ++ return ret; ++} ++ ++ ++static inline int __user_device(const char __user *name, dev_t *dev, ++ umode_t *mode) ++{ ++ struct nameidata nd; ++ int ret; ++ ++ if (!name) { ++ *dev = 0; ++ return 0; ++ } ++ ret = user_lpath(name, &nd.path); ++ if (ret) ++ return ret; ++ if (nd.path.dentry->d_inode) { ++ *dev = nd.path.dentry->d_inode->i_rdev; ++ *mode = nd.path.dentry->d_inode->i_mode; ++ } ++ path_put(&nd.path); ++ return 0; ++} ++ ++static inline int __mapping_mode(dev_t device, dev_t target, ++ umode_t device_mode, umode_t target_mode, umode_t *mode) ++{ ++ if (device) ++ *mode = device_mode & S_IFMT; ++ else if (target) ++ *mode = target_mode & S_IFMT; ++ else ++ return -EINVAL; ++ ++ /* if both given, device and target mode have to match */ ++ if (device && target && ++ ((device_mode ^ target_mode) & S_IFMT)) ++ return -EINVAL; ++ return 0; ++} ++ ++ ++static inline int do_mapping(struct vx_info *vxi, const char __user *device_path, ++ const char __user *target_path, int flags, int set) ++{ ++ dev_t device = ~0, target = ~0; ++ umode_t device_mode = 0, target_mode = 0, mode; ++ int ret; ++ ++ ret = __user_device(device_path, &device, &device_mode); ++ if (ret) ++ return ret; ++ ret = __user_device(target_path, &target, &target_mode); ++ if (ret) ++ return ret; ++ ++ ret = __mapping_mode(device, target, ++ device_mode, target_mode, &mode); ++ if (ret) ++ return ret; ++ ++ if (set) ++ return do_set_mapping(vxi, device, target, ++ flags, mode); ++ else ++ return do_unset_mapping(vxi, device, target, ++ flags, mode); ++} ++ ++ ++int vc_set_mapping(struct vx_info *vxi, void __user *data) ++{ ++ struct vcmd_set_mapping_v0 vc_data; ++ ++ if (copy_from_user(&vc_data, data, sizeof(vc_data))) ++ return -EFAULT; ++ ++ return do_mapping(vxi, vc_data.device, vc_data.target, ++ vc_data.flags, 1); ++} ++ ++int vc_unset_mapping(struct vx_info *vxi, void __user *data) ++{ ++ struct vcmd_set_mapping_v0 vc_data; ++ ++ if (copy_from_user(&vc_data, data, sizeof(vc_data))) ++ return -EFAULT; ++ ++ return do_mapping(vxi, vc_data.device, vc_data.target, ++ vc_data.flags, 0); ++} ++ ++ ++#ifdef CONFIG_COMPAT ++ ++int vc_set_mapping_x32(struct vx_info *vxi, void __user *data) ++{ ++ struct vcmd_set_mapping_v0_x32 vc_data; ++ ++ if (copy_from_user(&vc_data, data, sizeof(vc_data))) ++ return -EFAULT; ++ ++ return do_mapping(vxi, compat_ptr(vc_data.device_ptr), ++ compat_ptr(vc_data.target_ptr), vc_data.flags, 1); ++} ++ ++int vc_unset_mapping_x32(struct vx_info *vxi, void __user *data) ++{ ++ struct vcmd_set_mapping_v0_x32 vc_data; ++ ++ if (copy_from_user(&vc_data, data, sizeof(vc_data))) ++ return -EFAULT; ++ ++ return do_mapping(vxi, compat_ptr(vc_data.device_ptr), ++ compat_ptr(vc_data.target_ptr), vc_data.flags, 0); ++} ++ ++#endif /* CONFIG_COMPAT */ ++ ++ +diff -NurpP --minimal linux-3.0.9/kernel/vserver/dlimit.c linux-3.0.9-vs2.3.2.1/kernel/vserver/dlimit.c +--- linux-3.0.9/kernel/vserver/dlimit.c 1970-01-01 01:00:00.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/kernel/vserver/dlimit.c 2011-06-10 23:20:56.000000000 +0200 +@@ -0,0 +1,531 @@ ++/* ++ * linux/kernel/vserver/dlimit.c ++ * ++ * Virtual Server: Context Disk Limits ++ * ++ * Copyright (C) 2004-2009 Herbert Pötzl ++ * ++ * V0.01 initial version ++ * V0.02 compat32 splitup ++ * V0.03 extended interface ++ * ++ */ ++ ++#include <linux/statfs.h> ++#include <linux/sched.h> ++#include <linux/namei.h> ++#include <linux/vs_tag.h> ++#include <linux/vs_dlimit.h> ++#include <linux/vserver/dlimit_cmd.h> ++#include <linux/slab.h> ++// #include <linux/gfp.h> ++ ++#include <asm/uaccess.h> ++ ++/* __alloc_dl_info() ++ ++ * allocate an initialized dl_info struct ++ * doesn't make it visible (hash) */ ++ ++static struct dl_info *__alloc_dl_info(struct super_block *sb, tag_t tag) ++{ ++ struct dl_info *new = NULL; ++ ++ vxdprintk(VXD_CBIT(dlim, 5), ++ "alloc_dl_info(%p,%d)*", sb, tag); ++ ++ /* would this benefit from a slab cache? */ ++ new = kmalloc(sizeof(struct dl_info), GFP_KERNEL); ++ if (!new) ++ return 0; ++ ++ memset(new, 0, sizeof(struct dl_info)); ++ new->dl_tag = tag; ++ new->dl_sb = sb; ++ // INIT_RCU_HEAD(&new->dl_rcu); ++ INIT_HLIST_NODE(&new->dl_hlist); ++ spin_lock_init(&new->dl_lock); ++ atomic_set(&new->dl_refcnt, 0); ++ atomic_set(&new->dl_usecnt, 0); ++ ++ /* rest of init goes here */ ++ ++ vxdprintk(VXD_CBIT(dlim, 4), ++ "alloc_dl_info(%p,%d) = %p", sb, tag, new); ++ return new; ++} ++ ++/* __dealloc_dl_info() ++ ++ * final disposal of dl_info */ ++ ++static void __dealloc_dl_info(struct dl_info *dli) ++{ ++ vxdprintk(VXD_CBIT(dlim, 4), ++ "dealloc_dl_info(%p)", dli); ++ ++ dli->dl_hlist.next = LIST_POISON1; ++ dli->dl_tag = -1; ++ dli->dl_sb = 0; ++ ++ BUG_ON(atomic_read(&dli->dl_usecnt)); ++ BUG_ON(atomic_read(&dli->dl_refcnt)); ++ ++ kfree(dli); ++} ++ ++ ++/* hash table for dl_info hash */ ++ ++#define DL_HASH_SIZE 13 ++ ++struct hlist_head dl_info_hash[DL_HASH_SIZE]; ++ ++static DEFINE_SPINLOCK(dl_info_hash_lock); ++ ++ ++static inline unsigned int __hashval(struct super_block *sb, tag_t tag) ++{ ++ return ((tag ^ (unsigned long)sb) % DL_HASH_SIZE); ++} ++ ++ ++ ++/* __hash_dl_info() ++ ++ * add the dli to the global hash table ++ * requires the hash_lock to be held */ ++ ++static inline void __hash_dl_info(struct dl_info *dli) ++{ ++ struct hlist_head *head; ++ ++ vxdprintk(VXD_CBIT(dlim, 6), ++ "__hash_dl_info: %p[#%d]", dli, dli->dl_tag); ++ get_dl_info(dli); ++ head = &dl_info_hash[__hashval(dli->dl_sb, dli->dl_tag)]; ++ hlist_add_head_rcu(&dli->dl_hlist, head); ++} ++ ++/* __unhash_dl_info() ++ ++ * remove the dli from the global hash table ++ * requires the hash_lock to be held */ ++ ++static inline void __unhash_dl_info(struct dl_info *dli) ++{ ++ vxdprintk(VXD_CBIT(dlim, 6), ++ "__unhash_dl_info: %p[#%d]", dli, dli->dl_tag); ++ hlist_del_rcu(&dli->dl_hlist); ++ put_dl_info(dli); ++} ++ ++ ++/* __lookup_dl_info() ++ ++ * requires the rcu_read_lock() ++ * doesn't increment the dl_refcnt */ ++ ++static inline struct dl_info *__lookup_dl_info(struct super_block *sb, tag_t tag) ++{ ++ struct hlist_head *head = &dl_info_hash[__hashval(sb, tag)]; ++ struct hlist_node *pos; ++ struct dl_info *dli; ++ ++ hlist_for_each_entry_rcu(dli, pos, head, dl_hlist) { ++ ++ if (dli->dl_tag == tag && dli->dl_sb == sb) { ++ return dli; ++ } ++ } ++ return NULL; ++} ++ ++ ++struct dl_info *locate_dl_info(struct super_block *sb, tag_t tag) ++{ ++ struct dl_info *dli; ++ ++ rcu_read_lock(); ++ dli = get_dl_info(__lookup_dl_info(sb, tag)); ++ vxdprintk(VXD_CBIT(dlim, 7), ++ "locate_dl_info(%p,#%d) = %p", sb, tag, dli); ++ rcu_read_unlock(); ++ return dli; ++} ++ ++void rcu_free_dl_info(struct rcu_head *head) ++{ ++ struct dl_info *dli = container_of(head, struct dl_info, dl_rcu); ++ int usecnt, refcnt; ++ ++ BUG_ON(!dli || !head); ++ ++ usecnt = atomic_read(&dli->dl_usecnt); ++ BUG_ON(usecnt < 0); ++ ++ refcnt = atomic_read(&dli->dl_refcnt); ++ BUG_ON(refcnt < 0); ++ ++ vxdprintk(VXD_CBIT(dlim, 3), ++ "rcu_free_dl_info(%p)", dli); ++ if (!usecnt) ++ __dealloc_dl_info(dli); ++ else ++ printk("!!! rcu didn't free\n"); ++} ++ ++ ++ ++ ++static int do_addrem_dlimit(uint32_t id, const char __user *name, ++ uint32_t flags, int add) ++{ ++ struct path path; ++ int ret; ++ ++ ret = user_lpath(name, &path); ++ if (!ret) { ++ struct super_block *sb; ++ struct dl_info *dli; ++ ++ ret = -EINVAL; ++ if (!path.dentry->d_inode) ++ goto out_release; ++ if (!(sb = path.dentry->d_inode->i_sb)) ++ goto out_release; ++ ++ if (add) { ++ dli = __alloc_dl_info(sb, id); ++ spin_lock(&dl_info_hash_lock); ++ ++ ret = -EEXIST; ++ if (__lookup_dl_info(sb, id)) ++ goto out_unlock; ++ __hash_dl_info(dli); ++ dli = NULL; ++ } else { ++ spin_lock(&dl_info_hash_lock); ++ dli = __lookup_dl_info(sb, id); ++ ++ ret = -ESRCH; ++ if (!dli) ++ goto out_unlock; ++ __unhash_dl_info(dli); ++ } ++ ret = 0; ++ out_unlock: ++ spin_unlock(&dl_info_hash_lock); ++ if (add && dli) ++ __dealloc_dl_info(dli); ++ out_release: ++ path_put(&path); ++ } ++ return ret; ++} ++ ++int vc_add_dlimit(uint32_t id, void __user *data) ++{ ++ struct vcmd_ctx_dlimit_base_v0 vc_data; ++ ++ if (copy_from_user(&vc_data, data, sizeof(vc_data))) ++ return -EFAULT; ++ ++ return do_addrem_dlimit(id, vc_data.name, vc_data.flags, 1); ++} ++ ++int vc_rem_dlimit(uint32_t id, void __user *data) ++{ ++ struct vcmd_ctx_dlimit_base_v0 vc_data; ++ ++ if (copy_from_user(&vc_data, data, sizeof(vc_data))) ++ return -EFAULT; ++ ++ return do_addrem_dlimit(id, vc_data.name, vc_data.flags, 0); ++} ++ ++#ifdef CONFIG_COMPAT ++ ++int vc_add_dlimit_x32(uint32_t id, void __user *data) ++{ ++ struct vcmd_ctx_dlimit_base_v0_x32 vc_data; ++ ++ if (copy_from_user(&vc_data, data, sizeof(vc_data))) ++ return -EFAULT; ++ ++ return do_addrem_dlimit(id, ++ compat_ptr(vc_data.name_ptr), vc_data.flags, 1); ++} ++ ++int vc_rem_dlimit_x32(uint32_t id, void __user *data) ++{ ++ struct vcmd_ctx_dlimit_base_v0_x32 vc_data; ++ ++ if (copy_from_user(&vc_data, data, sizeof(vc_data))) ++ return -EFAULT; ++ ++ return do_addrem_dlimit(id, ++ compat_ptr(vc_data.name_ptr), vc_data.flags, 0); ++} ++ ++#endif /* CONFIG_COMPAT */ ++ ++ ++static inline ++int do_set_dlimit(uint32_t id, const char __user *name, ++ uint32_t space_used, uint32_t space_total, ++ uint32_t inodes_used, uint32_t inodes_total, ++ uint32_t reserved, uint32_t flags) ++{ ++ struct path path; ++ int ret; ++ ++ ret = user_lpath(name, &path); ++ if (!ret) { ++ struct super_block *sb; ++ struct dl_info *dli; ++ ++ ret = -EINVAL; ++ if (!path.dentry->d_inode) ++ goto out_release; ++ if (!(sb = path.dentry->d_inode->i_sb)) ++ goto out_release; ++ ++ /* sanity checks */ ++ if ((reserved != CDLIM_KEEP && ++ reserved > 100) || ++ (inodes_used != CDLIM_KEEP && ++ inodes_used > inodes_total) || ++ (space_used != CDLIM_KEEP && ++ space_used > space_total)) ++ goto out_release; ++ ++ ret = -ESRCH; ++ dli = locate_dl_info(sb, id); ++ if (!dli) ++ goto out_release; ++ ++ spin_lock(&dli->dl_lock); ++ ++ if (inodes_used != CDLIM_KEEP) ++ dli->dl_inodes_used = inodes_used; ++ if (inodes_total != CDLIM_KEEP) ++ dli->dl_inodes_total = inodes_total; ++ if (space_used != CDLIM_KEEP) ++ dli->dl_space_used = dlimit_space_32to64( ++ space_used, flags, DLIMS_USED); ++ ++ if (space_total == CDLIM_INFINITY) ++ dli->dl_space_total = DLIM_INFINITY; ++ else if (space_total != CDLIM_KEEP) ++ dli->dl_space_total = dlimit_space_32to64( ++ space_total, flags, DLIMS_TOTAL); ++ ++ if (reserved != CDLIM_KEEP) ++ dli->dl_nrlmult = (1 << 10) * (100 - reserved) / 100; ++ ++ spin_unlock(&dli->dl_lock); ++ ++ put_dl_info(dli); ++ ret = 0; ++ ++ out_release: ++ path_put(&path); ++ } ++ return ret; ++} ++ ++int vc_set_dlimit(uint32_t id, void __user *data) ++{ ++ struct vcmd_ctx_dlimit_v0 vc_data; ++ ++ if (copy_from_user(&vc_data, data, sizeof(vc_data))) ++ return -EFAULT; ++ ++ return do_set_dlimit(id, vc_data.name, ++ vc_data.space_used, vc_data.space_total, ++ vc_data.inodes_used, vc_data.inodes_total, ++ vc_data.reserved, vc_data.flags); ++} ++ ++#ifdef CONFIG_COMPAT ++ ++int vc_set_dlimit_x32(uint32_t id, void __user *data) ++{ ++ struct vcmd_ctx_dlimit_v0_x32 vc_data; ++ ++ if (copy_from_user(&vc_data, data, sizeof(vc_data))) ++ return -EFAULT; ++ ++ return do_set_dlimit(id, compat_ptr(vc_data.name_ptr), ++ vc_data.space_used, vc_data.space_total, ++ vc_data.inodes_used, vc_data.inodes_total, ++ vc_data.reserved, vc_data.flags); ++} ++ ++#endif /* CONFIG_COMPAT */ ++ ++ ++static inline ++int do_get_dlimit(uint32_t id, const char __user *name, ++ uint32_t *space_used, uint32_t *space_total, ++ uint32_t *inodes_used, uint32_t *inodes_total, ++ uint32_t *reserved, uint32_t *flags) ++{ ++ struct path path; ++ int ret; ++ ++ ret = user_lpath(name, &path); ++ if (!ret) { ++ struct super_block *sb; ++ struct dl_info *dli; ++ ++ ret = -EINVAL; ++ if (!path.dentry->d_inode) ++ goto out_release; ++ if (!(sb = path.dentry->d_inode->i_sb)) ++ goto out_release; ++ ++ ret = -ESRCH; ++ dli = locate_dl_info(sb, id); ++ if (!dli) ++ goto out_release; ++ ++ spin_lock(&dli->dl_lock); ++ *inodes_used = dli->dl_inodes_used; ++ *inodes_total = dli->dl_inodes_total; ++ ++ *space_used = dlimit_space_64to32( ++ dli->dl_space_used, flags, DLIMS_USED); ++ ++ if (dli->dl_space_total == DLIM_INFINITY) ++ *space_total = CDLIM_INFINITY; ++ else ++ *space_total = dlimit_space_64to32( ++ dli->dl_space_total, flags, DLIMS_TOTAL); ++ ++ *reserved = 100 - ((dli->dl_nrlmult * 100 + 512) >> 10); ++ spin_unlock(&dli->dl_lock); ++ ++ put_dl_info(dli); ++ ret = -EFAULT; ++ ++ ret = 0; ++ out_release: ++ path_put(&path); ++ } ++ return ret; ++} ++ ++ ++int vc_get_dlimit(uint32_t id, void __user *data) ++{ ++ struct vcmd_ctx_dlimit_v0 vc_data; ++ int ret; ++ ++ if (copy_from_user(&vc_data, data, sizeof(vc_data))) ++ return -EFAULT; ++ ++ ret = do_get_dlimit(id, vc_data.name, ++ &vc_data.space_used, &vc_data.space_total, ++ &vc_data.inodes_used, &vc_data.inodes_total, ++ &vc_data.reserved, &vc_data.flags); ++ if (ret) ++ return ret; ++ ++ if (copy_to_user(data, &vc_data, sizeof(vc_data))) ++ return -EFAULT; ++ return 0; ++} ++ ++#ifdef CONFIG_COMPAT ++ ++int vc_get_dlimit_x32(uint32_t id, void __user *data) ++{ ++ struct vcmd_ctx_dlimit_v0_x32 vc_data; ++ int ret; ++ ++ if (copy_from_user(&vc_data, data, sizeof(vc_data))) ++ return -EFAULT; ++ ++ ret = do_get_dlimit(id, compat_ptr(vc_data.name_ptr), ++ &vc_data.space_used, &vc_data.space_total, ++ &vc_data.inodes_used, &vc_data.inodes_total, ++ &vc_data.reserved, &vc_data.flags); ++ if (ret) ++ return ret; ++ ++ if (copy_to_user(data, &vc_data, sizeof(vc_data))) ++ return -EFAULT; ++ return 0; ++} ++ ++#endif /* CONFIG_COMPAT */ ++ ++ ++void vx_vsi_statfs(struct super_block *sb, struct kstatfs *buf) ++{ ++ struct dl_info *dli; ++ __u64 blimit, bfree, bavail; ++ __u32 ifree; ++ ++ dli = locate_dl_info(sb, dx_current_tag()); ++ if (!dli) ++ return; ++ ++ spin_lock(&dli->dl_lock); ++ if (dli->dl_inodes_total == (unsigned long)DLIM_INFINITY) ++ goto no_ilim; ++ ++ /* reduce max inodes available to limit */ ++ if (buf->f_files > dli->dl_inodes_total) ++ buf->f_files = dli->dl_inodes_total; ++ ++ ifree = dli->dl_inodes_total - dli->dl_inodes_used; ++ /* reduce free inodes to min */ ++ if (ifree < buf->f_ffree) ++ buf->f_ffree = ifree; ++ ++no_ilim: ++ if (dli->dl_space_total == DLIM_INFINITY) ++ goto no_blim; ++ ++ blimit = dli->dl_space_total >> sb->s_blocksize_bits; ++ ++ if (dli->dl_space_total < dli->dl_space_used) ++ bfree = 0; ++ else ++ bfree = (dli->dl_space_total - dli->dl_space_used) ++ >> sb->s_blocksize_bits; ++ ++ bavail = ((dli->dl_space_total >> 10) * dli->dl_nrlmult); ++ if (bavail < dli->dl_space_used) ++ bavail = 0; ++ else ++ bavail = (bavail - dli->dl_space_used) ++ >> sb->s_blocksize_bits; ++ ++ /* reduce max space available to limit */ ++ if (buf->f_blocks > blimit) ++ buf->f_blocks = blimit; ++ ++ /* reduce free space to min */ ++ if (bfree < buf->f_bfree) ++ buf->f_bfree = bfree; ++ ++ /* reduce avail space to min */ ++ if (bavail < buf->f_bavail) ++ buf->f_bavail = bavail; ++ ++no_blim: ++ spin_unlock(&dli->dl_lock); ++ put_dl_info(dli); ++ ++ return; ++} ++ ++#include <linux/module.h> ++ ++EXPORT_SYMBOL_GPL(locate_dl_info); ++EXPORT_SYMBOL_GPL(rcu_free_dl_info); ++ +diff -NurpP --minimal linux-3.0.9/kernel/vserver/helper.c linux-3.0.9-vs2.3.2.1/kernel/vserver/helper.c +--- linux-3.0.9/kernel/vserver/helper.c 1970-01-01 01:00:00.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/kernel/vserver/helper.c 2011-06-10 22:11:24.000000000 +0200 +@@ -0,0 +1,223 @@ ++/* ++ * linux/kernel/vserver/helper.c ++ * ++ * Virtual Context Support ++ * ++ * Copyright (C) 2004-2007 Herbert Pötzl ++ * ++ * V0.01 basic helper ++ * ++ */ ++ ++#include <linux/kmod.h> ++#include <linux/reboot.h> ++#include <linux/vs_context.h> ++#include <linux/vs_network.h> ++#include <linux/vserver/signal.h> ++ ++ ++char vshelper_path[255] = "/sbin/vshelper"; ++ ++ ++static int do_vshelper(char *name, char *argv[], char *envp[], int sync) ++{ ++ int ret; ++ ++ if ((ret = call_usermodehelper(name, argv, envp, sync))) { ++ printk( KERN_WARNING ++ "%s: (%s %s) returned %s with %d\n", ++ name, argv[1], argv[2], ++ sync ? "sync" : "async", ret); ++ } ++ vxdprintk(VXD_CBIT(switch, 4), ++ "%s: (%s %s) returned %s with %d", ++ name, argv[1], argv[2], sync ? "sync" : "async", ret); ++ return ret; ++} ++ ++/* ++ * vshelper path is set via /proc/sys ++ * invoked by vserver sys_reboot(), with ++ * the following arguments ++ * ++ * argv [0] = vshelper_path; ++ * argv [1] = action: "restart", "halt", "poweroff", ... ++ * argv [2] = context identifier ++ * ++ * envp [*] = type-specific parameters ++ */ ++ ++long vs_reboot_helper(struct vx_info *vxi, int cmd, void __user *arg) ++{ ++ char id_buf[8], cmd_buf[16]; ++ char uid_buf[16], pid_buf[16]; ++ int ret; ++ ++ char *argv[] = {vshelper_path, NULL, id_buf, 0}; ++ char *envp[] = {"HOME=/", "TERM=linux", ++ "PATH=/sbin:/usr/sbin:/bin:/usr/bin", ++ uid_buf, pid_buf, cmd_buf, 0}; ++ ++ if (vx_info_state(vxi, VXS_HELPER)) ++ return -EAGAIN; ++ vxi->vx_state |= VXS_HELPER; ++ ++ snprintf(id_buf, sizeof(id_buf)-1, "%d", vxi->vx_id); ++ ++ snprintf(cmd_buf, sizeof(cmd_buf)-1, "VS_CMD=%08x", cmd); ++ snprintf(uid_buf, sizeof(uid_buf)-1, "VS_UID=%d", current_uid()); ++ snprintf(pid_buf, sizeof(pid_buf)-1, "VS_PID=%d", current->pid); ++ ++ switch (cmd) { ++ case LINUX_REBOOT_CMD_RESTART: ++ argv[1] = "restart"; ++ break; ++ ++ case LINUX_REBOOT_CMD_HALT: ++ argv[1] = "halt"; ++ break; ++ ++ case LINUX_REBOOT_CMD_POWER_OFF: ++ argv[1] = "poweroff"; ++ break; ++ ++ case LINUX_REBOOT_CMD_SW_SUSPEND: ++ argv[1] = "swsusp"; ++ break; ++ ++ case LINUX_REBOOT_CMD_OOM: ++ argv[1] = "oom"; ++ break; ++ ++ default: ++ vxi->vx_state &= ~VXS_HELPER; ++ return 0; ++ } ++ ++ ret = do_vshelper(vshelper_path, argv, envp, 0); ++ vxi->vx_state &= ~VXS_HELPER; ++ __wakeup_vx_info(vxi); ++ return (ret) ? -EPERM : 0; ++} ++ ++ ++long vs_reboot(unsigned int cmd, void __user *arg) ++{ ++ struct vx_info *vxi = current_vx_info(); ++ long ret = 0; ++ ++ vxdprintk(VXD_CBIT(misc, 5), ++ "vs_reboot(%p[#%d],%u)", ++ vxi, vxi ? vxi->vx_id : 0, cmd); ++ ++ ret = vs_reboot_helper(vxi, cmd, arg); ++ if (ret) ++ return ret; ++ ++ vxi->reboot_cmd = cmd; ++ if (vx_info_flags(vxi, VXF_REBOOT_KILL, 0)) { ++ switch (cmd) { ++ case LINUX_REBOOT_CMD_RESTART: ++ case LINUX_REBOOT_CMD_HALT: ++ case LINUX_REBOOT_CMD_POWER_OFF: ++ vx_info_kill(vxi, 0, SIGKILL); ++ vx_info_kill(vxi, 1, SIGKILL); ++ default: ++ break; ++ } ++ } ++ return 0; ++} ++ ++long vs_oom_action(unsigned int cmd) ++{ ++ struct vx_info *vxi = current_vx_info(); ++ long ret = 0; ++ ++ vxdprintk(VXD_CBIT(misc, 5), ++ "vs_oom_action(%p[#%d],%u)", ++ vxi, vxi ? vxi->vx_id : 0, cmd); ++ ++ ret = vs_reboot_helper(vxi, cmd, NULL); ++ if (ret) ++ return ret; ++ ++ vxi->reboot_cmd = cmd; ++ if (vx_info_flags(vxi, VXF_REBOOT_KILL, 0)) { ++ vx_info_kill(vxi, 0, SIGKILL); ++ vx_info_kill(vxi, 1, SIGKILL); ++ } ++ return 0; ++} ++ ++/* ++ * argv [0] = vshelper_path; ++ * argv [1] = action: "startup", "shutdown" ++ * argv [2] = context identifier ++ * ++ * envp [*] = type-specific parameters ++ */ ++ ++long vs_state_change(struct vx_info *vxi, unsigned int cmd) ++{ ++ char id_buf[8], cmd_buf[16]; ++ char *argv[] = {vshelper_path, NULL, id_buf, 0}; ++ char *envp[] = {"HOME=/", "TERM=linux", ++ "PATH=/sbin:/usr/sbin:/bin:/usr/bin", cmd_buf, 0}; ++ ++ if (!vx_info_flags(vxi, VXF_SC_HELPER, 0)) ++ return 0; ++ ++ snprintf(id_buf, sizeof(id_buf)-1, "%d", vxi->vx_id); ++ snprintf(cmd_buf, sizeof(cmd_buf)-1, "VS_CMD=%08x", cmd); ++ ++ switch (cmd) { ++ case VSC_STARTUP: ++ argv[1] = "startup"; ++ break; ++ case VSC_SHUTDOWN: ++ argv[1] = "shutdown"; ++ break; ++ default: ++ return 0; ++ } ++ ++ return do_vshelper(vshelper_path, argv, envp, 1); ++} ++ ++ ++/* ++ * argv [0] = vshelper_path; ++ * argv [1] = action: "netup", "netdown" ++ * argv [2] = context identifier ++ * ++ * envp [*] = type-specific parameters ++ */ ++ ++long vs_net_change(struct nx_info *nxi, unsigned int cmd) ++{ ++ char id_buf[8], cmd_buf[16]; ++ char *argv[] = {vshelper_path, NULL, id_buf, 0}; ++ char *envp[] = {"HOME=/", "TERM=linux", ++ "PATH=/sbin:/usr/sbin:/bin:/usr/bin", cmd_buf, 0}; ++ ++ if (!nx_info_flags(nxi, NXF_SC_HELPER, 0)) ++ return 0; ++ ++ snprintf(id_buf, sizeof(id_buf)-1, "%d", nxi->nx_id); ++ snprintf(cmd_buf, sizeof(cmd_buf)-1, "VS_CMD=%08x", cmd); ++ ++ switch (cmd) { ++ case VSC_NETUP: ++ argv[1] = "netup"; ++ break; ++ case VSC_NETDOWN: ++ argv[1] = "netdown"; ++ break; ++ default: ++ return 0; ++ } ++ ++ return do_vshelper(vshelper_path, argv, envp, 1); ++} ++ +diff -NurpP --minimal linux-3.0.9/kernel/vserver/history.c linux-3.0.9-vs2.3.2.1/kernel/vserver/history.c +--- linux-3.0.9/kernel/vserver/history.c 1970-01-01 01:00:00.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/kernel/vserver/history.c 2011-06-10 22:11:24.000000000 +0200 +@@ -0,0 +1,258 @@ ++/* ++ * kernel/vserver/history.c ++ * ++ * Virtual Context History Backtrace ++ * ++ * Copyright (C) 2004-2007 Herbert Pötzl ++ * ++ * V0.01 basic structure ++ * V0.02 hash/unhash and trace ++ * V0.03 preemption fixes ++ * ++ */ ++ ++#include <linux/module.h> ++#include <asm/uaccess.h> ++ ++#include <linux/vserver/context.h> ++#include <linux/vserver/debug.h> ++#include <linux/vserver/debug_cmd.h> ++#include <linux/vserver/history.h> ++ ++ ++#ifdef CONFIG_VSERVER_HISTORY ++#define VXH_SIZE CONFIG_VSERVER_HISTORY_SIZE ++#else ++#define VXH_SIZE 64 ++#endif ++ ++struct _vx_history { ++ unsigned int counter; ++ ++ struct _vx_hist_entry entry[VXH_SIZE + 1]; ++}; ++ ++ ++DEFINE_PER_CPU(struct _vx_history, vx_history_buffer); ++ ++unsigned volatile int vxh_active = 1; ++ ++static atomic_t sequence = ATOMIC_INIT(0); ++ ++ ++/* vxh_advance() ++ ++ * requires disabled preemption */ ++ ++struct _vx_hist_entry *vxh_advance(void *loc) ++{ ++ unsigned int cpu = smp_processor_id(); ++ struct _vx_history *hist = &per_cpu(vx_history_buffer, cpu); ++ struct _vx_hist_entry *entry; ++ unsigned int index; ++ ++ index = vxh_active ? (hist->counter++ % VXH_SIZE) : VXH_SIZE; ++ entry = &hist->entry[index]; ++ ++ entry->seq = atomic_inc_return(&sequence); ++ entry->loc = loc; ++ return entry; ++} ++ ++EXPORT_SYMBOL_GPL(vxh_advance); ++ ++ ++#define VXH_LOC_FMTS "(#%04x,*%d):%p" ++ ++#define VXH_LOC_ARGS(e) (e)->seq, cpu, (e)->loc ++ ++ ++#define VXH_VXI_FMTS "%p[#%d,%d.%d]" ++ ++#define VXH_VXI_ARGS(e) (e)->vxi.ptr, \ ++ (e)->vxi.ptr ? (e)->vxi.xid : 0, \ ++ (e)->vxi.ptr ? (e)->vxi.usecnt : 0, \ ++ (e)->vxi.ptr ? (e)->vxi.tasks : 0 ++ ++void vxh_dump_entry(struct _vx_hist_entry *e, unsigned cpu) ++{ ++ switch (e->type) { ++ case VXH_THROW_OOPS: ++ printk( VXH_LOC_FMTS " oops \n", VXH_LOC_ARGS(e)); ++ break; ++ ++ case VXH_GET_VX_INFO: ++ case VXH_PUT_VX_INFO: ++ printk( VXH_LOC_FMTS " %s_vx_info " VXH_VXI_FMTS "\n", ++ VXH_LOC_ARGS(e), ++ (e->type == VXH_GET_VX_INFO) ? "get" : "put", ++ VXH_VXI_ARGS(e)); ++ break; ++ ++ case VXH_INIT_VX_INFO: ++ case VXH_SET_VX_INFO: ++ case VXH_CLR_VX_INFO: ++ printk( VXH_LOC_FMTS " %s_vx_info " VXH_VXI_FMTS " @%p\n", ++ VXH_LOC_ARGS(e), ++ (e->type == VXH_INIT_VX_INFO) ? "init" : ++ ((e->type == VXH_SET_VX_INFO) ? "set" : "clr"), ++ VXH_VXI_ARGS(e), e->sc.data); ++ break; ++ ++ case VXH_CLAIM_VX_INFO: ++ case VXH_RELEASE_VX_INFO: ++ printk( VXH_LOC_FMTS " %s_vx_info " VXH_VXI_FMTS " @%p\n", ++ VXH_LOC_ARGS(e), ++ (e->type == VXH_CLAIM_VX_INFO) ? "claim" : "release", ++ VXH_VXI_ARGS(e), e->sc.data); ++ break; ++ ++ case VXH_ALLOC_VX_INFO: ++ case VXH_DEALLOC_VX_INFO: ++ printk( VXH_LOC_FMTS " %s_vx_info " VXH_VXI_FMTS "\n", ++ VXH_LOC_ARGS(e), ++ (e->type == VXH_ALLOC_VX_INFO) ? "alloc" : "dealloc", ++ VXH_VXI_ARGS(e)); ++ break; ++ ++ case VXH_HASH_VX_INFO: ++ case VXH_UNHASH_VX_INFO: ++ printk( VXH_LOC_FMTS " __%s_vx_info " VXH_VXI_FMTS "\n", ++ VXH_LOC_ARGS(e), ++ (e->type == VXH_HASH_VX_INFO) ? "hash" : "unhash", ++ VXH_VXI_ARGS(e)); ++ break; ++ ++ case VXH_LOC_VX_INFO: ++ case VXH_LOOKUP_VX_INFO: ++ case VXH_CREATE_VX_INFO: ++ printk( VXH_LOC_FMTS " __%s_vx_info [#%d] -> " VXH_VXI_FMTS "\n", ++ VXH_LOC_ARGS(e), ++ (e->type == VXH_CREATE_VX_INFO) ? "create" : ++ ((e->type == VXH_LOC_VX_INFO) ? "loc" : "lookup"), ++ e->ll.arg, VXH_VXI_ARGS(e)); ++ break; ++ } ++} ++ ++static void __vxh_dump_history(void) ++{ ++ unsigned int i, cpu; ++ ++ printk("History:\tSEQ: %8x\tNR_CPUS: %d\n", ++ atomic_read(&sequence), NR_CPUS); ++ ++ for (i = 0; i < VXH_SIZE; i++) { ++ for_each_online_cpu(cpu) { ++ struct _vx_history *hist = ++ &per_cpu(vx_history_buffer, cpu); ++ unsigned int index = (hist->counter - i) % VXH_SIZE; ++ struct _vx_hist_entry *entry = &hist->entry[index]; ++ ++ vxh_dump_entry(entry, cpu); ++ } ++ } ++} ++ ++void vxh_dump_history(void) ++{ ++ vxh_active = 0; ++#ifdef CONFIG_SMP ++ local_irq_enable(); ++ smp_send_stop(); ++ local_irq_disable(); ++#endif ++ __vxh_dump_history(); ++} ++ ++ ++/* vserver syscall commands below here */ ++ ++ ++int vc_dump_history(uint32_t id) ++{ ++ vxh_active = 0; ++ __vxh_dump_history(); ++ vxh_active = 1; ++ ++ return 0; ++} ++ ++ ++int do_read_history(struct __user _vx_hist_entry *data, ++ int cpu, uint32_t *index, uint32_t *count) ++{ ++ int pos, ret = 0; ++ struct _vx_history *hist = &per_cpu(vx_history_buffer, cpu); ++ int end = hist->counter; ++ int start = end - VXH_SIZE + 2; ++ int idx = *index; ++ ++ /* special case: get current pos */ ++ if (!*count) { ++ *index = end; ++ return 0; ++ } ++ ++ /* have we lost some data? */ ++ if (idx < start) ++ idx = start; ++ ++ for (pos = 0; (pos < *count) && (idx < end); pos++, idx++) { ++ struct _vx_hist_entry *entry = ++ &hist->entry[idx % VXH_SIZE]; ++ ++ /* send entry to userspace */ ++ ret = copy_to_user(&data[pos], entry, sizeof(*entry)); ++ if (ret) ++ break; ++ } ++ /* save new index and count */ ++ *index = idx; ++ *count = pos; ++ return ret ? ret : (*index < end); ++} ++ ++int vc_read_history(uint32_t id, void __user *data) ++{ ++ struct vcmd_read_history_v0 vc_data; ++ int ret; ++ ++ if (id >= NR_CPUS) ++ return -EINVAL; ++ ++ if (copy_from_user(&vc_data, data, sizeof(vc_data))) ++ return -EFAULT; ++ ++ ret = do_read_history((struct __user _vx_hist_entry *)vc_data.data, ++ id, &vc_data.index, &vc_data.count); ++ ++ if (copy_to_user(data, &vc_data, sizeof(vc_data))) ++ return -EFAULT; ++ return ret; ++} ++ ++#ifdef CONFIG_COMPAT ++ ++int vc_read_history_x32(uint32_t id, void __user *data) ++{ ++ struct vcmd_read_history_v0_x32 vc_data; ++ int ret; ++ ++ if (id >= NR_CPUS) ++ return -EINVAL; ++ ++ if (copy_from_user(&vc_data, data, sizeof(vc_data))) ++ return -EFAULT; ++ ++ ret = do_read_history((struct __user _vx_hist_entry *) ++ compat_ptr(vc_data.data_ptr), ++ id, &vc_data.index, &vc_data.count); ++ ++ if (copy_to_user(data, &vc_data, sizeof(vc_data))) ++ return -EFAULT; ++ return ret; ++} ++ ++#endif /* CONFIG_COMPAT */ ++ +diff -NurpP --minimal linux-3.0.9/kernel/vserver/inet.c linux-3.0.9-vs2.3.2.1/kernel/vserver/inet.c +--- linux-3.0.9/kernel/vserver/inet.c 1970-01-01 01:00:00.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/kernel/vserver/inet.c 2011-08-30 20:14:33.000000000 +0200 +@@ -0,0 +1,225 @@ ++ ++#include <linux/in.h> ++#include <linux/inetdevice.h> ++#include <linux/vs_inet.h> ++#include <linux/vs_inet6.h> ++#include <linux/vserver/debug.h> ++#include <net/route.h> ++#include <net/addrconf.h> ++ ++ ++int nx_v4_addr_conflict(struct nx_info *nxi1, struct nx_info *nxi2) ++{ ++ int ret = 0; ++ ++ if (!nxi1 || !nxi2 || nxi1 == nxi2) ++ ret = 1; ++ else { ++ struct nx_addr_v4 *ptr; ++ ++ for (ptr = &nxi1->v4; ptr; ptr = ptr->next) { ++ if (v4_nx_addr_in_nx_info(nxi2, ptr, -1)) { ++ ret = 1; ++ break; ++ } ++ } ++ } ++ ++ vxdprintk(VXD_CBIT(net, 2), ++ "nx_v4_addr_conflict(%p,%p): %d", ++ nxi1, nxi2, ret); ++ ++ return ret; ++} ++ ++ ++#ifdef CONFIG_IPV6 ++ ++int nx_v6_addr_conflict(struct nx_info *nxi1, struct nx_info *nxi2) ++{ ++ int ret = 0; ++ ++ if (!nxi1 || !nxi2 || nxi1 == nxi2) ++ ret = 1; ++ else { ++ struct nx_addr_v6 *ptr; ++ ++ for (ptr = &nxi1->v6; ptr; ptr = ptr->next) { ++ if (v6_nx_addr_in_nx_info(nxi2, ptr, -1)) { ++ ret = 1; ++ break; ++ } ++ } ++ } ++ ++ vxdprintk(VXD_CBIT(net, 2), ++ "nx_v6_addr_conflict(%p,%p): %d", ++ nxi1, nxi2, ret); ++ ++ return ret; ++} ++ ++#endif ++ ++int v4_dev_in_nx_info(struct net_device *dev, struct nx_info *nxi) ++{ ++ struct in_device *in_dev; ++ struct in_ifaddr **ifap; ++ struct in_ifaddr *ifa; ++ int ret = 0; ++ ++ if (!dev) ++ goto out; ++ in_dev = in_dev_get(dev); ++ if (!in_dev) ++ goto out; ++ ++ for (ifap = &in_dev->ifa_list; (ifa = *ifap) != NULL; ++ ifap = &ifa->ifa_next) { ++ if (v4_addr_in_nx_info(nxi, ifa->ifa_local, NXA_MASK_SHOW)) { ++ ret = 1; ++ break; ++ } ++ } ++ in_dev_put(in_dev); ++out: ++ return ret; ++} ++ ++ ++#ifdef CONFIG_IPV6 ++ ++int v6_dev_in_nx_info(struct net_device *dev, struct nx_info *nxi) ++{ ++ struct inet6_dev *in_dev; ++ struct inet6_ifaddr *ifa; ++ int ret = 0; ++ ++ if (!dev) ++ goto out; ++ in_dev = in6_dev_get(dev); ++ if (!in_dev) ++ goto out; ++ ++ // for (ifap = &in_dev->addr_list; (ifa = *ifap) != NULL; ++ list_for_each_entry(ifa, &in_dev->addr_list, if_list) { ++ if (v6_addr_in_nx_info(nxi, &ifa->addr, -1)) { ++ ret = 1; ++ break; ++ } ++ } ++ in6_dev_put(in_dev); ++out: ++ return ret; ++} ++ ++#endif ++ ++int dev_in_nx_info(struct net_device *dev, struct nx_info *nxi) ++{ ++ int ret = 1; ++ ++ if (!nxi) ++ goto out; ++ if (nxi->v4.type && v4_dev_in_nx_info(dev, nxi)) ++ goto out; ++#ifdef CONFIG_IPV6 ++ ret = 2; ++ if (nxi->v6.type && v6_dev_in_nx_info(dev, nxi)) ++ goto out; ++#endif ++ ret = 0; ++out: ++ vxdprintk(VXD_CBIT(net, 3), ++ "dev_in_nx_info(%p,%p[#%d]) = %d", ++ dev, nxi, nxi ? nxi->nx_id : 0, ret); ++ return ret; ++} ++ ++struct rtable *ip_v4_find_src(struct net *net, struct nx_info *nxi, ++ struct flowi4 *fl4) ++{ ++ struct rtable *rt; ++ ++ if (!nxi) ++ return NULL; ++ ++ /* FIXME: handle lback only case */ ++ if (!NX_IPV4(nxi)) ++ return ERR_PTR(-EPERM); ++ ++ vxdprintk(VXD_CBIT(net, 4), ++ "ip_v4_find_src(%p[#%u]) " NIPQUAD_FMT " -> " NIPQUAD_FMT, ++ nxi, nxi ? nxi->nx_id : 0, ++ NIPQUAD(fl4->saddr), NIPQUAD(fl4->daddr)); ++ ++ /* single IP is unconditional */ ++ if (nx_info_flags(nxi, NXF_SINGLE_IP, 0) && ++ (fl4->saddr == INADDR_ANY)) ++ fl4->saddr = nxi->v4.ip[0].s_addr; ++ ++ if (fl4->saddr == INADDR_ANY) { ++ struct nx_addr_v4 *ptr; ++ __be32 found = 0; ++ ++ rt = __ip_route_output_key(net, fl4); ++ if (!IS_ERR(rt)) { ++ found = fl4->saddr; ++ ip_rt_put(rt); ++ vxdprintk(VXD_CBIT(net, 4), ++ "ip_v4_find_src(%p[#%u]) rok[%u]: " NIPQUAD_FMT, ++ nxi, nxi ? nxi->nx_id : 0, fl4->flowi4_oif, NIPQUAD(found)); ++ if (v4_addr_in_nx_info(nxi, found, NXA_MASK_BIND)) ++ goto found; ++ } ++ ++ for (ptr = &nxi->v4; ptr; ptr = ptr->next) { ++ __be32 primary = ptr->ip[0].s_addr; ++ __be32 mask = ptr->mask.s_addr; ++ __be32 neta = primary & mask; ++ ++ vxdprintk(VXD_CBIT(net, 4), "ip_v4_find_src(%p[#%u]) chk: " ++ NIPQUAD_FMT "/" NIPQUAD_FMT "/" NIPQUAD_FMT, ++ nxi, nxi ? nxi->nx_id : 0, NIPQUAD(primary), ++ NIPQUAD(mask), NIPQUAD(neta)); ++ if ((found & mask) != neta) ++ continue; ++ ++ fl4->saddr = primary; ++ rt = __ip_route_output_key(net, fl4); ++ vxdprintk(VXD_CBIT(net, 4), ++ "ip_v4_find_src(%p[#%u]) rok[%u]: " NIPQUAD_FMT, ++ nxi, nxi ? nxi->nx_id : 0, fl4->flowi4_oif, NIPQUAD(primary)); ++ if (!IS_ERR(rt)) { ++ found = fl4->saddr; ++ ip_rt_put(rt); ++ if (found == primary) ++ goto found; ++ } ++ } ++ /* still no source ip? */ ++ found = ipv4_is_loopback(fl4->daddr) ++ ? IPI_LOOPBACK : nxi->v4.ip[0].s_addr; ++ found: ++ /* assign src ip to flow */ ++ fl4->saddr = found; ++ ++ } else { ++ if (!v4_addr_in_nx_info(nxi, fl4->saddr, NXA_MASK_BIND)) ++ return ERR_PTR(-EPERM); ++ } ++ ++ if (nx_info_flags(nxi, NXF_LBACK_REMAP, 0)) { ++ if (ipv4_is_loopback(fl4->daddr)) ++ fl4->daddr = nxi->v4_lback.s_addr; ++ if (ipv4_is_loopback(fl4->saddr)) ++ fl4->saddr = nxi->v4_lback.s_addr; ++ } else if (ipv4_is_loopback(fl4->daddr) && ++ !nx_info_flags(nxi, NXF_LBACK_ALLOW, 0)) ++ return ERR_PTR(-EPERM); ++ ++ return NULL; ++} ++ ++EXPORT_SYMBOL_GPL(ip_v4_find_src); ++ +diff -NurpP --minimal linux-3.0.9/kernel/vserver/init.c linux-3.0.9-vs2.3.2.1/kernel/vserver/init.c +--- linux-3.0.9/kernel/vserver/init.c 1970-01-01 01:00:00.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/kernel/vserver/init.c 2011-06-10 22:11:24.000000000 +0200 +@@ -0,0 +1,45 @@ ++/* ++ * linux/kernel/init.c ++ * ++ * Virtual Server Init ++ * ++ * Copyright (C) 2004-2007 Herbert Pötzl ++ * ++ * V0.01 basic structure ++ * ++ */ ++ ++#include <linux/init.h> ++ ++int vserver_register_sysctl(void); ++void vserver_unregister_sysctl(void); ++ ++ ++static int __init init_vserver(void) ++{ ++ int ret = 0; ++ ++#ifdef CONFIG_VSERVER_DEBUG ++ vserver_register_sysctl(); ++#endif ++ return ret; ++} ++ ++ ++static void __exit exit_vserver(void) ++{ ++ ++#ifdef CONFIG_VSERVER_DEBUG ++ vserver_unregister_sysctl(); ++#endif ++ return; ++} ++ ++/* FIXME: GFP_ZONETYPES gone ++long vx_slab[GFP_ZONETYPES]; */ ++long vx_area; ++ ++ ++module_init(init_vserver); ++module_exit(exit_vserver); ++ +diff -NurpP --minimal linux-3.0.9/kernel/vserver/inode.c linux-3.0.9-vs2.3.2.1/kernel/vserver/inode.c +--- linux-3.0.9/kernel/vserver/inode.c 1970-01-01 01:00:00.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/kernel/vserver/inode.c 2011-06-10 22:11:24.000000000 +0200 +@@ -0,0 +1,437 @@ ++/* ++ * linux/kernel/vserver/inode.c ++ * ++ * Virtual Server: File System Support ++ * ++ * Copyright (C) 2004-2007 Herbert Pötzl ++ * ++ * V0.01 separated from vcontext V0.05 ++ * V0.02 moved to tag (instead of xid) ++ * ++ */ ++ ++#include <linux/tty.h> ++#include <linux/proc_fs.h> ++#include <linux/devpts_fs.h> ++#include <linux/fs.h> ++#include <linux/file.h> ++#include <linux/mount.h> ++#include <linux/parser.h> ++#include <linux/namei.h> ++#include <linux/vserver/inode.h> ++#include <linux/vserver/inode_cmd.h> ++#include <linux/vs_base.h> ++#include <linux/vs_tag.h> ++ ++#include <asm/uaccess.h> ++ ++ ++static int __vc_get_iattr(struct inode *in, uint32_t *tag, uint32_t *flags, uint32_t *mask) ++{ ++ struct proc_dir_entry *entry; ++ ++ if (!in || !in->i_sb) ++ return -ESRCH; ++ ++ *flags = IATTR_TAG ++ | (IS_IMMUTABLE(in) ? IATTR_IMMUTABLE : 0) ++ | (IS_IXUNLINK(in) ? IATTR_IXUNLINK : 0) ++ | (IS_BARRIER(in) ? IATTR_BARRIER : 0) ++ | (IS_COW(in) ? IATTR_COW : 0); ++ *mask = IATTR_IXUNLINK | IATTR_IMMUTABLE | IATTR_COW; ++ ++ if (S_ISDIR(in->i_mode)) ++ *mask |= IATTR_BARRIER; ++ ++ if (IS_TAGGED(in)) { ++ *tag = in->i_tag; ++ *mask |= IATTR_TAG; ++ } ++ ++ switch (in->i_sb->s_magic) { ++ case PROC_SUPER_MAGIC: ++ entry = PROC_I(in)->pde; ++ ++ /* check for specific inodes? */ ++ if (entry) ++ *mask |= IATTR_FLAGS; ++ if (entry) ++ *flags |= (entry->vx_flags & IATTR_FLAGS); ++ else ++ *flags |= (PROC_I(in)->vx_flags & IATTR_FLAGS); ++ break; ++ ++ case DEVPTS_SUPER_MAGIC: ++ *tag = in->i_tag; ++ *mask |= IATTR_TAG; ++ break; ++ ++ default: ++ break; ++ } ++ return 0; ++} ++ ++int vc_get_iattr(void __user *data) ++{ ++ struct path path; ++ struct vcmd_ctx_iattr_v1 vc_data = { .tag = -1 }; ++ int ret; ++ ++ if (copy_from_user(&vc_data, data, sizeof(vc_data))) ++ return -EFAULT; ++ ++ ret = user_lpath(vc_data.name, &path); ++ if (!ret) { ++ ret = __vc_get_iattr(path.dentry->d_inode, ++ &vc_data.tag, &vc_data.flags, &vc_data.mask); ++ path_put(&path); ++ } ++ if (ret) ++ return ret; ++ ++ if (copy_to_user(data, &vc_data, sizeof(vc_data))) ++ ret = -EFAULT; ++ return ret; ++} ++ ++#ifdef CONFIG_COMPAT ++ ++int vc_get_iattr_x32(void __user *data) ++{ ++ struct path path; ++ struct vcmd_ctx_iattr_v1_x32 vc_data = { .tag = -1 }; ++ int ret; ++ ++ if (copy_from_user(&vc_data, data, sizeof(vc_data))) ++ return -EFAULT; ++ ++ ret = user_lpath(compat_ptr(vc_data.name_ptr), &path); ++ if (!ret) { ++ ret = __vc_get_iattr(path.dentry->d_inode, ++ &vc_data.tag, &vc_data.flags, &vc_data.mask); ++ path_put(&path); ++ } ++ if (ret) ++ return ret; ++ ++ if (copy_to_user(data, &vc_data, sizeof(vc_data))) ++ ret = -EFAULT; ++ return ret; ++} ++ ++#endif /* CONFIG_COMPAT */ ++ ++ ++int vc_fget_iattr(uint32_t fd, void __user *data) ++{ ++ struct file *filp; ++ struct vcmd_ctx_fiattr_v0 vc_data = { .tag = -1 }; ++ int ret; ++ ++ if (copy_from_user(&vc_data, data, sizeof(vc_data))) ++ return -EFAULT; ++ ++ filp = fget(fd); ++ if (!filp || !filp->f_dentry || !filp->f_dentry->d_inode) ++ return -EBADF; ++ ++ ret = __vc_get_iattr(filp->f_dentry->d_inode, ++ &vc_data.tag, &vc_data.flags, &vc_data.mask); ++ ++ fput(filp); ++ ++ if (copy_to_user(data, &vc_data, sizeof(vc_data))) ++ ret = -EFAULT; ++ return ret; ++} ++ ++ ++static int __vc_set_iattr(struct dentry *de, uint32_t *tag, uint32_t *flags, uint32_t *mask) ++{ ++ struct inode *in = de->d_inode; ++ int error = 0, is_proc = 0, has_tag = 0; ++ struct iattr attr = { 0 }; ++ ++ if (!in || !in->i_sb) ++ return -ESRCH; ++ ++ is_proc = (in->i_sb->s_magic == PROC_SUPER_MAGIC); ++ if ((*mask & IATTR_FLAGS) && !is_proc) ++ return -EINVAL; ++ ++ has_tag = IS_TAGGED(in) || ++ (in->i_sb->s_magic == DEVPTS_SUPER_MAGIC); ++ if ((*mask & IATTR_TAG) && !has_tag) ++ return -EINVAL; ++ ++ mutex_lock(&in->i_mutex); ++ if (*mask & IATTR_TAG) { ++ attr.ia_tag = *tag; ++ attr.ia_valid |= ATTR_TAG; ++ } ++ ++ if (*mask & IATTR_FLAGS) { ++ struct proc_dir_entry *entry = PROC_I(in)->pde; ++ unsigned int iflags = PROC_I(in)->vx_flags; ++ ++ iflags = (iflags & ~(*mask & IATTR_FLAGS)) ++ | (*flags & IATTR_FLAGS); ++ PROC_I(in)->vx_flags = iflags; ++ if (entry) ++ entry->vx_flags = iflags; ++ } ++ ++ if (*mask & (IATTR_IMMUTABLE | IATTR_IXUNLINK | ++ IATTR_BARRIER | IATTR_COW)) { ++ int iflags = in->i_flags; ++ int vflags = in->i_vflags; ++ ++ if (*mask & IATTR_IMMUTABLE) { ++ if (*flags & IATTR_IMMUTABLE) ++ iflags |= S_IMMUTABLE; ++ else ++ iflags &= ~S_IMMUTABLE; ++ } ++ if (*mask & IATTR_IXUNLINK) { ++ if (*flags & IATTR_IXUNLINK) ++ iflags |= S_IXUNLINK; ++ else ++ iflags &= ~S_IXUNLINK; ++ } ++ if (S_ISDIR(in->i_mode) && (*mask & IATTR_BARRIER)) { ++ if (*flags & IATTR_BARRIER) ++ vflags |= V_BARRIER; ++ else ++ vflags &= ~V_BARRIER; ++ } ++ if (S_ISREG(in->i_mode) && (*mask & IATTR_COW)) { ++ if (*flags & IATTR_COW) ++ vflags |= V_COW; ++ else ++ vflags &= ~V_COW; ++ } ++ if (in->i_op && in->i_op->sync_flags) { ++ error = in->i_op->sync_flags(in, iflags, vflags); ++ if (error) ++ goto out; ++ } ++ } ++ ++ if (attr.ia_valid) { ++ if (in->i_op && in->i_op->setattr) ++ error = in->i_op->setattr(de, &attr); ++ else { ++ error = inode_change_ok(in, &attr); ++ if (!error) { ++ setattr_copy(in, &attr); ++ mark_inode_dirty(in); ++ } ++ } ++ } ++ ++out: ++ mutex_unlock(&in->i_mutex); ++ return error; ++} ++ ++int vc_set_iattr(void __user *data) ++{ ++ struct path path; ++ struct vcmd_ctx_iattr_v1 vc_data; ++ int ret; ++ ++ if (!capable(CAP_LINUX_IMMUTABLE)) ++ return -EPERM; ++ if (copy_from_user(&vc_data, data, sizeof(vc_data))) ++ return -EFAULT; ++ ++ ret = user_lpath(vc_data.name, &path); ++ if (!ret) { ++ ret = __vc_set_iattr(path.dentry, ++ &vc_data.tag, &vc_data.flags, &vc_data.mask); ++ path_put(&path); ++ } ++ ++ if (copy_to_user(data, &vc_data, sizeof(vc_data))) ++ ret = -EFAULT; ++ return ret; ++} ++ ++#ifdef CONFIG_COMPAT ++ ++int vc_set_iattr_x32(void __user *data) ++{ ++ struct path path; ++ struct vcmd_ctx_iattr_v1_x32 vc_data; ++ int ret; ++ ++ if (!capable(CAP_LINUX_IMMUTABLE)) ++ return -EPERM; ++ if (copy_from_user(&vc_data, data, sizeof(vc_data))) ++ return -EFAULT; ++ ++ ret = user_lpath(compat_ptr(vc_data.name_ptr), &path); ++ if (!ret) { ++ ret = __vc_set_iattr(path.dentry, ++ &vc_data.tag, &vc_data.flags, &vc_data.mask); ++ path_put(&path); ++ } ++ ++ if (copy_to_user(data, &vc_data, sizeof(vc_data))) ++ ret = -EFAULT; ++ return ret; ++} ++ ++#endif /* CONFIG_COMPAT */ ++ ++int vc_fset_iattr(uint32_t fd, void __user *data) ++{ ++ struct file *filp; ++ struct vcmd_ctx_fiattr_v0 vc_data; ++ int ret; ++ ++ if (!capable(CAP_LINUX_IMMUTABLE)) ++ return -EPERM; ++ if (copy_from_user(&vc_data, data, sizeof(vc_data))) ++ return -EFAULT; ++ ++ filp = fget(fd); ++ if (!filp || !filp->f_dentry || !filp->f_dentry->d_inode) ++ return -EBADF; ++ ++ ret = __vc_set_iattr(filp->f_dentry, &vc_data.tag, ++ &vc_data.flags, &vc_data.mask); ++ ++ fput(filp); ++ ++ if (copy_to_user(data, &vc_data, sizeof(vc_data))) ++ return -EFAULT; ++ return ret; ++} ++ ++ ++enum { Opt_notagcheck, Opt_tag, Opt_notag, Opt_tagid, Opt_err }; ++ ++static match_table_t tokens = { ++ {Opt_notagcheck, "notagcheck"}, ++#ifdef CONFIG_PROPAGATE ++ {Opt_notag, "notag"}, ++ {Opt_tag, "tag"}, ++ {Opt_tagid, "tagid=%u"}, ++#endif ++ {Opt_err, NULL} ++}; ++ ++ ++static void __dx_parse_remove(char *string, char *opt) ++{ ++ char *p = strstr(string, opt); ++ char *q = p; ++ ++ if (p) { ++ while (*q != '\0' && *q != ',') ++ q++; ++ while (*q) ++ *p++ = *q++; ++ while (*p) ++ *p++ = '\0'; ++ } ++} ++ ++int dx_parse_tag(char *string, tag_t *tag, int remove, int *mnt_flags, ++ unsigned long *flags) ++{ ++ int set = 0; ++ substring_t args[MAX_OPT_ARGS]; ++ int token; ++ char *s, *p, *opts; ++#if defined(CONFIG_PROPAGATE) || defined(CONFIG_VSERVER_WARN) ++ int option = 0; ++#endif ++ ++ if (!string) ++ return 0; ++ s = kstrdup(string, GFP_KERNEL | GFP_ATOMIC); ++ if (!s) ++ return 0; ++ ++ opts = s; ++ while ((p = strsep(&opts, ",")) != NULL) { ++ token = match_token(p, tokens, args); ++ ++ switch (token) { ++#ifdef CONFIG_PROPAGATE ++ case Opt_tag: ++ if (tag) ++ *tag = 0; ++ if (remove) ++ __dx_parse_remove(s, "tag"); ++ *mnt_flags |= MNT_TAGID; ++ set |= MNT_TAGID; ++ break; ++ case Opt_notag: ++ if (remove) ++ __dx_parse_remove(s, "notag"); ++ *mnt_flags |= MNT_NOTAG; ++ set |= MNT_NOTAG; ++ break; ++ case Opt_tagid: ++ if (tag && !match_int(args, &option)) ++ *tag = option; ++ if (remove) ++ __dx_parse_remove(s, "tagid"); ++ *mnt_flags |= MNT_TAGID; ++ set |= MNT_TAGID; ++ break; ++#endif ++ case Opt_notagcheck: ++ if (remove) ++ __dx_parse_remove(s, "notagcheck"); ++ *flags |= MS_NOTAGCHECK; ++ set |= MS_NOTAGCHECK; ++ break; ++ } ++ vxdprintk(VXD_CBIT(tag, 7), ++ "dx_parse_tag(" VS_Q("%s") "): %d:#%d", ++ p, token, option); ++ } ++ if (set) ++ strcpy(string, s); ++ kfree(s); ++ return set; ++} ++ ++#ifdef CONFIG_PROPAGATE ++ ++void __dx_propagate_tag(struct nameidata *nd, struct inode *inode) ++{ ++ tag_t new_tag = 0; ++ struct vfsmount *mnt; ++ int propagate; ++ ++ if (!nd) ++ return; ++ mnt = nd->path.mnt; ++ if (!mnt) ++ return; ++ ++ propagate = (mnt->mnt_flags & MNT_TAGID); ++ if (propagate) ++ new_tag = mnt->mnt_tag; ++ ++ vxdprintk(VXD_CBIT(tag, 7), ++ "dx_propagate_tag(%p[#%lu.%d]): %d,%d", ++ inode, inode->i_ino, inode->i_tag, ++ new_tag, (propagate) ? 1 : 0); ++ ++ if (propagate) ++ inode->i_tag = new_tag; ++} ++ ++#include <linux/module.h> ++ ++EXPORT_SYMBOL_GPL(__dx_propagate_tag); ++ ++#endif /* CONFIG_PROPAGATE */ ++ +diff -NurpP --minimal linux-3.0.9/kernel/vserver/limit.c linux-3.0.9-vs2.3.2.1/kernel/vserver/limit.c +--- linux-3.0.9/kernel/vserver/limit.c 1970-01-01 01:00:00.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/kernel/vserver/limit.c 2011-10-11 20:48:08.000000000 +0200 +@@ -0,0 +1,360 @@ ++/* ++ * linux/kernel/vserver/limit.c ++ * ++ * Virtual Server: Context Limits ++ * ++ * Copyright (C) 2004-2010 Herbert Pötzl ++ * ++ * V0.01 broken out from vcontext V0.05 ++ * V0.02 changed vcmds to vxi arg ++ * V0.03 added memory cgroup support ++ * ++ */ ++ ++#include <linux/sched.h> ++#include <linux/module.h> ++#include <linux/memcontrol.h> ++#include <linux/res_counter.h> ++#include <linux/vs_limit.h> ++#include <linux/vserver/limit.h> ++#include <linux/vserver/limit_cmd.h> ++ ++#include <asm/uaccess.h> ++ ++ ++const char *vlimit_name[NUM_LIMITS] = { ++#ifdef CONFIG_VSERVER_LEGACY_MEM ++ [RLIMIT_RSS] = "RSS", ++ [RLIMIT_AS] = "VM", ++#endif /* CONFIG_VSERVER_LEGACY_MEM */ ++ [RLIMIT_CPU] = "CPU", ++ [RLIMIT_NPROC] = "NPROC", ++ [RLIMIT_NOFILE] = "NOFILE", ++ [RLIMIT_LOCKS] = "LOCKS", ++ [RLIMIT_SIGPENDING] = "SIGP", ++ [RLIMIT_MSGQUEUE] = "MSGQ", ++ ++ [VLIMIT_NSOCK] = "NSOCK", ++ [VLIMIT_OPENFD] = "OPENFD", ++ [VLIMIT_SHMEM] = "SHMEM", ++ [VLIMIT_DENTRY] = "DENTRY", ++}; ++ ++EXPORT_SYMBOL_GPL(vlimit_name); ++ ++#define MASK_ENTRY(x) (1 << (x)) ++ ++const struct vcmd_ctx_rlimit_mask_v0 vlimit_mask = { ++ /* minimum */ ++ 0 ++ , /* softlimit */ ++#ifdef CONFIG_VSERVER_LEGACY_MEM ++ MASK_ENTRY( RLIMIT_RSS ) | ++#endif /* CONFIG_VSERVER_LEGACY_MEM */ ++ 0 ++ , /* maximum */ ++#ifdef CONFIG_VSERVER_LEGACY_MEM ++ MASK_ENTRY( RLIMIT_RSS ) | ++ MASK_ENTRY( RLIMIT_AS ) | ++#endif /* CONFIG_VSERVER_LEGACY_MEM */ ++ MASK_ENTRY( RLIMIT_NPROC ) | ++ MASK_ENTRY( RLIMIT_NOFILE ) | ++ MASK_ENTRY( RLIMIT_LOCKS ) | ++ MASK_ENTRY( RLIMIT_MSGQUEUE ) | ++ ++ MASK_ENTRY( VLIMIT_NSOCK ) | ++ MASK_ENTRY( VLIMIT_OPENFD ) | ++ MASK_ENTRY( VLIMIT_SHMEM ) | ++ MASK_ENTRY( VLIMIT_DENTRY ) | ++ 0 ++}; ++ /* accounting only */ ++uint32_t account_mask = ++ MASK_ENTRY( VLIMIT_SEMARY ) | ++ MASK_ENTRY( VLIMIT_NSEMS ) | ++ MASK_ENTRY( VLIMIT_MAPPED ) | ++ 0; ++ ++ ++static int is_valid_vlimit(int id) ++{ ++ uint32_t mask = vlimit_mask.minimum | ++ vlimit_mask.softlimit | vlimit_mask.maximum; ++ return mask & (1 << id); ++} ++ ++static int is_accounted_vlimit(int id) ++{ ++ if (is_valid_vlimit(id)) ++ return 1; ++ return account_mask & (1 << id); ++} ++ ++ ++static inline uint64_t vc_get_soft(struct vx_info *vxi, int id) ++{ ++ rlim_t limit = __rlim_soft(&vxi->limit, id); ++ return VX_VLIM(limit); ++} ++ ++static inline uint64_t vc_get_hard(struct vx_info *vxi, int id) ++{ ++ rlim_t limit = __rlim_hard(&vxi->limit, id); ++ return VX_VLIM(limit); ++} ++ ++static int do_get_rlimit(struct vx_info *vxi, uint32_t id, ++ uint64_t *minimum, uint64_t *softlimit, uint64_t *maximum) ++{ ++ if (!is_valid_vlimit(id)) ++ return -EINVAL; ++ ++ if (minimum) ++ *minimum = CRLIM_UNSET; ++ if (softlimit) ++ *softlimit = vc_get_soft(vxi, id); ++ if (maximum) ++ *maximum = vc_get_hard(vxi, id); ++ return 0; ++} ++ ++int vc_get_rlimit(struct vx_info *vxi, void __user *data) ++{ ++ struct vcmd_ctx_rlimit_v0 vc_data; ++ int ret; ++ ++ if (copy_from_user(&vc_data, data, sizeof(vc_data))) ++ return -EFAULT; ++ ++ ret = do_get_rlimit(vxi, vc_data.id, ++ &vc_data.minimum, &vc_data.softlimit, &vc_data.maximum); ++ if (ret) ++ return ret; ++ ++ if (copy_to_user(data, &vc_data, sizeof(vc_data))) ++ return -EFAULT; ++ return 0; ++} ++ ++static int do_set_rlimit(struct vx_info *vxi, uint32_t id, ++ uint64_t minimum, uint64_t softlimit, uint64_t maximum) ++{ ++ if (!is_valid_vlimit(id)) ++ return -EINVAL; ++ ++ if (maximum != CRLIM_KEEP) ++ __rlim_hard(&vxi->limit, id) = VX_RLIM(maximum); ++ if (softlimit != CRLIM_KEEP) ++ __rlim_soft(&vxi->limit, id) = VX_RLIM(softlimit); ++ ++ /* clamp soft limit */ ++ if (__rlim_soft(&vxi->limit, id) > __rlim_hard(&vxi->limit, id)) ++ __rlim_soft(&vxi->limit, id) = __rlim_hard(&vxi->limit, id); ++ ++ return 0; ++} ++ ++int vc_set_rlimit(struct vx_info *vxi, void __user *data) ++{ ++ struct vcmd_ctx_rlimit_v0 vc_data; ++ ++ if (copy_from_user(&vc_data, data, sizeof(vc_data))) ++ return -EFAULT; ++ ++ return do_set_rlimit(vxi, vc_data.id, ++ vc_data.minimum, vc_data.softlimit, vc_data.maximum); ++} ++ ++#ifdef CONFIG_IA32_EMULATION ++ ++int vc_set_rlimit_x32(struct vx_info *vxi, void __user *data) ++{ ++ struct vcmd_ctx_rlimit_v0_x32 vc_data; ++ ++ if (copy_from_user(&vc_data, data, sizeof(vc_data))) ++ return -EFAULT; ++ ++ return do_set_rlimit(vxi, vc_data.id, ++ vc_data.minimum, vc_data.softlimit, vc_data.maximum); ++} ++ ++int vc_get_rlimit_x32(struct vx_info *vxi, void __user *data) ++{ ++ struct vcmd_ctx_rlimit_v0_x32 vc_data; ++ int ret; ++ ++ if (copy_from_user(&vc_data, data, sizeof(vc_data))) ++ return -EFAULT; ++ ++ ret = do_get_rlimit(vxi, vc_data.id, ++ &vc_data.minimum, &vc_data.softlimit, &vc_data.maximum); ++ if (ret) ++ return ret; ++ ++ if (copy_to_user(data, &vc_data, sizeof(vc_data))) ++ return -EFAULT; ++ return 0; ++} ++ ++#endif /* CONFIG_IA32_EMULATION */ ++ ++ ++int vc_get_rlimit_mask(uint32_t id, void __user *data) ++{ ++ if (copy_to_user(data, &vlimit_mask, sizeof(vlimit_mask))) ++ return -EFAULT; ++ return 0; ++} ++ ++ ++static inline void vx_reset_hits(struct _vx_limit *limit) ++{ ++ int lim; ++ ++ for (lim = 0; lim < NUM_LIMITS; lim++) { ++ atomic_set(&__rlim_lhit(limit, lim), 0); ++ } ++} ++ ++int vc_reset_hits(struct vx_info *vxi, void __user *data) ++{ ++ vx_reset_hits(&vxi->limit); ++ return 0; ++} ++ ++static inline void vx_reset_minmax(struct _vx_limit *limit) ++{ ++ rlim_t value; ++ int lim; ++ ++ for (lim = 0; lim < NUM_LIMITS; lim++) { ++ value = __rlim_get(limit, lim); ++ __rlim_rmax(limit, lim) = value; ++ __rlim_rmin(limit, lim) = value; ++ } ++} ++ ++int vc_reset_minmax(struct vx_info *vxi, void __user *data) ++{ ++ vx_reset_minmax(&vxi->limit); ++ return 0; ++} ++ ++ ++int vc_rlimit_stat(struct vx_info *vxi, void __user *data) ++{ ++ struct vcmd_rlimit_stat_v0 vc_data; ++ struct _vx_limit *limit = &vxi->limit; ++ int id; ++ ++ if (copy_from_user(&vc_data, data, sizeof(vc_data))) ++ return -EFAULT; ++ ++ id = vc_data.id; ++ if (!is_accounted_vlimit(id)) ++ return -EINVAL; ++ ++ vx_limit_fixup(limit, id); ++ vc_data.hits = atomic_read(&__rlim_lhit(limit, id)); ++ vc_data.value = __rlim_get(limit, id); ++ vc_data.minimum = __rlim_rmin(limit, id); ++ vc_data.maximum = __rlim_rmax(limit, id); ++ ++ if (copy_to_user(data, &vc_data, sizeof(vc_data))) ++ return -EFAULT; ++ return 0; ++} ++ ++ ++void vx_vsi_meminfo(struct sysinfo *val) ++{ ++#ifdef CONFIG_CGROUP_MEM_RES_CTLR ++ struct mem_cgroup *mcg = mem_cgroup_from_task(current); ++ u64 res_limit, res_usage; ++ ++ if (!mcg) ++ return; ++ ++ res_limit = mem_cgroup_res_read_u64(mcg, RES_LIMIT); ++ res_usage = mem_cgroup_res_read_u64(mcg, RES_USAGE); ++ ++ if (res_limit != RESOURCE_MAX) ++ val->totalram = (res_limit >> PAGE_SHIFT); ++ val->freeram = val->totalram - (res_usage >> PAGE_SHIFT); ++ val->bufferram = 0; ++ val->totalhigh = 0; ++ val->freehigh = 0; ++#endif /* CONFIG_CGROUP_MEM_RES_CTLR */ ++ return; ++} ++ ++void vx_vsi_swapinfo(struct sysinfo *val) ++{ ++#ifdef CONFIG_CGROUP_MEM_RES_CTLR ++#ifdef CONFIG_CGROUP_MEM_RES_CTLR_SWAP ++ struct mem_cgroup *mcg = mem_cgroup_from_task(current); ++ u64 res_limit, res_usage, memsw_limit, memsw_usage; ++ s64 swap_limit, swap_usage; ++ ++ if (!mcg) ++ return; ++ ++ res_limit = mem_cgroup_res_read_u64(mcg, RES_LIMIT); ++ res_usage = mem_cgroup_res_read_u64(mcg, RES_USAGE); ++ memsw_limit = mem_cgroup_memsw_read_u64(mcg, RES_LIMIT); ++ memsw_usage = mem_cgroup_memsw_read_u64(mcg, RES_USAGE); ++ ++ /* memory unlimited */ ++ if (res_limit == RESOURCE_MAX) ++ return; ++ ++ swap_limit = memsw_limit - res_limit; ++ /* we have a swap limit? */ ++ if (memsw_limit != RESOURCE_MAX) ++ val->totalswap = swap_limit >> PAGE_SHIFT; ++ ++ /* calculate swap part */ ++ swap_usage = (memsw_usage > res_usage) ? ++ memsw_usage - res_usage : 0; ++ ++ /* total shown minus usage gives free swap */ ++ val->freeswap = (swap_usage < swap_limit) ? ++ val->totalswap - (swap_usage >> PAGE_SHIFT) : 0; ++#else /* !CONFIG_CGROUP_MEM_RES_CTLR_SWAP */ ++ val->totalswap = 0; ++ val->freeswap = 0; ++#endif /* !CONFIG_CGROUP_MEM_RES_CTLR_SWAP */ ++#endif /* CONFIG_CGROUP_MEM_RES_CTLR */ ++ return; ++} ++ ++long vx_vsi_cached(struct sysinfo *val) ++{ ++#ifdef CONFIG_CGROUP_MEM_RES_CTLR ++ struct mem_cgroup *mcg = mem_cgroup_from_task(current); ++ ++ return mem_cgroup_stat_read_cache(mcg); ++#else ++ return 0; ++#endif ++} ++ ++ ++unsigned long vx_badness(struct task_struct *task, struct mm_struct *mm) ++{ ++ struct vx_info *vxi = mm->mm_vx_info; ++ unsigned long points; ++ rlim_t v, w; ++ ++ if (!vxi) ++ return 0; ++ ++ points = vxi->vx_badness_bias; ++ ++ v = __vx_cres_array_fixup(&vxi->limit, VLA_RSS); ++ w = __rlim_soft(&vxi->limit, RLIMIT_RSS); ++ points += (v > w) ? (v - w) : 0; ++ ++ return points; ++} ++ +diff -NurpP --minimal linux-3.0.9/kernel/vserver/limit_init.h linux-3.0.9-vs2.3.2.1/kernel/vserver/limit_init.h +--- linux-3.0.9/kernel/vserver/limit_init.h 1970-01-01 01:00:00.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/kernel/vserver/limit_init.h 2011-06-10 22:11:24.000000000 +0200 +@@ -0,0 +1,31 @@ ++ ++ ++static inline void vx_info_init_limit(struct _vx_limit *limit) ++{ ++ int lim; ++ ++ for (lim = 0; lim < NUM_LIMITS; lim++) { ++ __rlim_soft(limit, lim) = RLIM_INFINITY; ++ __rlim_hard(limit, lim) = RLIM_INFINITY; ++ __rlim_set(limit, lim, 0); ++ atomic_set(&__rlim_lhit(limit, lim), 0); ++ __rlim_rmin(limit, lim) = 0; ++ __rlim_rmax(limit, lim) = 0; ++ } ++} ++ ++static inline void vx_info_exit_limit(struct _vx_limit *limit) ++{ ++ rlim_t value; ++ int lim; ++ ++ for (lim = 0; lim < NUM_LIMITS; lim++) { ++ if ((1 << lim) & VLIM_NOCHECK) ++ continue; ++ value = __rlim_get(limit, lim); ++ vxwprintk_xid(value, ++ "!!! limit: %p[%s,%d] = %ld on exit.", ++ limit, vlimit_name[lim], lim, (long)value); ++ } ++} ++ +diff -NurpP --minimal linux-3.0.9/kernel/vserver/limit_proc.h linux-3.0.9-vs2.3.2.1/kernel/vserver/limit_proc.h +--- linux-3.0.9/kernel/vserver/limit_proc.h 1970-01-01 01:00:00.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/kernel/vserver/limit_proc.h 2011-06-10 22:11:24.000000000 +0200 +@@ -0,0 +1,57 @@ ++#ifndef _VX_LIMIT_PROC_H ++#define _VX_LIMIT_PROC_H ++ ++#include <linux/vserver/limit_int.h> ++ ++ ++#define VX_LIMIT_FMT ":\t%8ld\t%8ld/%8ld\t%8lld/%8lld\t%6d\n" ++#define VX_LIMIT_TOP \ ++ "Limit\t current\t min/max\t\t soft/hard\t\thits\n" ++ ++#define VX_LIMIT_ARG(r) \ ++ (unsigned long)__rlim_get(limit, r), \ ++ (unsigned long)__rlim_rmin(limit, r), \ ++ (unsigned long)__rlim_rmax(limit, r), \ ++ VX_VLIM(__rlim_soft(limit, r)), \ ++ VX_VLIM(__rlim_hard(limit, r)), \ ++ atomic_read(&__rlim_lhit(limit, r)) ++ ++static inline int vx_info_proc_limit(struct _vx_limit *limit, char *buffer) ++{ ++ vx_limit_fixup(limit, -1); ++ return sprintf(buffer, VX_LIMIT_TOP ++ "PROC" VX_LIMIT_FMT ++ "VM" VX_LIMIT_FMT ++ "VML" VX_LIMIT_FMT ++ "RSS" VX_LIMIT_FMT ++ "ANON" VX_LIMIT_FMT ++ "RMAP" VX_LIMIT_FMT ++ "FILES" VX_LIMIT_FMT ++ "OFD" VX_LIMIT_FMT ++ "LOCKS" VX_LIMIT_FMT ++ "SOCK" VX_LIMIT_FMT ++ "MSGQ" VX_LIMIT_FMT ++ "SHM" VX_LIMIT_FMT ++ "SEMA" VX_LIMIT_FMT ++ "SEMS" VX_LIMIT_FMT ++ "DENT" VX_LIMIT_FMT, ++ VX_LIMIT_ARG(RLIMIT_NPROC), ++ VX_LIMIT_ARG(RLIMIT_AS), ++ VX_LIMIT_ARG(RLIMIT_MEMLOCK), ++ VX_LIMIT_ARG(RLIMIT_RSS), ++ VX_LIMIT_ARG(VLIMIT_ANON), ++ VX_LIMIT_ARG(VLIMIT_MAPPED), ++ VX_LIMIT_ARG(RLIMIT_NOFILE), ++ VX_LIMIT_ARG(VLIMIT_OPENFD), ++ VX_LIMIT_ARG(RLIMIT_LOCKS), ++ VX_LIMIT_ARG(VLIMIT_NSOCK), ++ VX_LIMIT_ARG(RLIMIT_MSGQUEUE), ++ VX_LIMIT_ARG(VLIMIT_SHMEM), ++ VX_LIMIT_ARG(VLIMIT_SEMARY), ++ VX_LIMIT_ARG(VLIMIT_NSEMS), ++ VX_LIMIT_ARG(VLIMIT_DENTRY)); ++} ++ ++#endif /* _VX_LIMIT_PROC_H */ ++ ++ +diff -NurpP --minimal linux-3.0.9/kernel/vserver/network.c linux-3.0.9-vs2.3.2.1/kernel/vserver/network.c +--- linux-3.0.9/kernel/vserver/network.c 1970-01-01 01:00:00.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/kernel/vserver/network.c 2011-06-10 23:20:56.000000000 +0200 +@@ -0,0 +1,912 @@ ++/* ++ * linux/kernel/vserver/network.c ++ * ++ * Virtual Server: Network Support ++ * ++ * Copyright (C) 2003-2007 Herbert Pötzl ++ * ++ * V0.01 broken out from vcontext V0.05 ++ * V0.02 cleaned up implementation ++ * V0.03 added equiv nx commands ++ * V0.04 switch to RCU based hash ++ * V0.05 and back to locking again ++ * V0.06 changed vcmds to nxi arg ++ * V0.07 have __create claim() the nxi ++ * ++ */ ++ ++#include <linux/err.h> ++#include <linux/slab.h> ++#include <linux/rcupdate.h> ++ ++#include <linux/vs_network.h> ++#include <linux/vs_pid.h> ++#include <linux/vserver/network_cmd.h> ++ ++ ++atomic_t nx_global_ctotal = ATOMIC_INIT(0); ++atomic_t nx_global_cactive = ATOMIC_INIT(0); ++ ++static struct kmem_cache *nx_addr_v4_cachep = NULL; ++static struct kmem_cache *nx_addr_v6_cachep = NULL; ++ ++ ++static int __init init_network(void) ++{ ++ nx_addr_v4_cachep = kmem_cache_create("nx_v4_addr_cache", ++ sizeof(struct nx_addr_v4), 0, ++ SLAB_HWCACHE_ALIGN|SLAB_PANIC, NULL); ++ nx_addr_v6_cachep = kmem_cache_create("nx_v6_addr_cache", ++ sizeof(struct nx_addr_v6), 0, ++ SLAB_HWCACHE_ALIGN|SLAB_PANIC, NULL); ++ return 0; ++} ++ ++ ++/* __alloc_nx_addr_v4() */ ++ ++static inline struct nx_addr_v4 *__alloc_nx_addr_v4(void) ++{ ++ struct nx_addr_v4 *nxa = kmem_cache_alloc( ++ nx_addr_v4_cachep, GFP_KERNEL); ++ ++ if (!IS_ERR(nxa)) ++ memset(nxa, 0, sizeof(*nxa)); ++ return nxa; ++} ++ ++/* __dealloc_nx_addr_v4() */ ++ ++static inline void __dealloc_nx_addr_v4(struct nx_addr_v4 *nxa) ++{ ++ kmem_cache_free(nx_addr_v4_cachep, nxa); ++} ++ ++/* __dealloc_nx_addr_v4_all() */ ++ ++static inline void __dealloc_nx_addr_v4_all(struct nx_addr_v4 *nxa) ++{ ++ while (nxa) { ++ struct nx_addr_v4 *next = nxa->next; ++ ++ __dealloc_nx_addr_v4(nxa); ++ nxa = next; ++ } ++} ++ ++ ++#ifdef CONFIG_IPV6 ++ ++/* __alloc_nx_addr_v6() */ ++ ++static inline struct nx_addr_v6 *__alloc_nx_addr_v6(void) ++{ ++ struct nx_addr_v6 *nxa = kmem_cache_alloc( ++ nx_addr_v6_cachep, GFP_KERNEL); ++ ++ if (!IS_ERR(nxa)) ++ memset(nxa, 0, sizeof(*nxa)); ++ return nxa; ++} ++ ++/* __dealloc_nx_addr_v6() */ ++ ++static inline void __dealloc_nx_addr_v6(struct nx_addr_v6 *nxa) ++{ ++ kmem_cache_free(nx_addr_v6_cachep, nxa); ++} ++ ++/* __dealloc_nx_addr_v6_all() */ ++ ++static inline void __dealloc_nx_addr_v6_all(struct nx_addr_v6 *nxa) ++{ ++ while (nxa) { ++ struct nx_addr_v6 *next = nxa->next; ++ ++ __dealloc_nx_addr_v6(nxa); ++ nxa = next; ++ } ++} ++ ++#endif /* CONFIG_IPV6 */ ++ ++/* __alloc_nx_info() ++ ++ * allocate an initialized nx_info struct ++ * doesn't make it visible (hash) */ ++ ++static struct nx_info *__alloc_nx_info(nid_t nid) ++{ ++ struct nx_info *new = NULL; ++ ++ vxdprintk(VXD_CBIT(nid, 1), "alloc_nx_info(%d)*", nid); ++ ++ /* would this benefit from a slab cache? */ ++ new = kmalloc(sizeof(struct nx_info), GFP_KERNEL); ++ if (!new) ++ return 0; ++ ++ memset(new, 0, sizeof(struct nx_info)); ++ new->nx_id = nid; ++ INIT_HLIST_NODE(&new->nx_hlist); ++ atomic_set(&new->nx_usecnt, 0); ++ atomic_set(&new->nx_tasks, 0); ++ new->nx_state = 0; ++ ++ new->nx_flags = NXF_INIT_SET; ++ ++ /* rest of init goes here */ ++ ++ new->v4_lback.s_addr = htonl(INADDR_LOOPBACK); ++ new->v4_bcast.s_addr = htonl(INADDR_BROADCAST); ++ ++ vxdprintk(VXD_CBIT(nid, 0), ++ "alloc_nx_info(%d) = %p", nid, new); ++ atomic_inc(&nx_global_ctotal); ++ return new; ++} ++ ++/* __dealloc_nx_info() ++ ++ * final disposal of nx_info */ ++ ++static void __dealloc_nx_info(struct nx_info *nxi) ++{ ++ vxdprintk(VXD_CBIT(nid, 0), ++ "dealloc_nx_info(%p)", nxi); ++ ++ nxi->nx_hlist.next = LIST_POISON1; ++ nxi->nx_id = -1; ++ ++ BUG_ON(atomic_read(&nxi->nx_usecnt)); ++ BUG_ON(atomic_read(&nxi->nx_tasks)); ++ ++ __dealloc_nx_addr_v4_all(nxi->v4.next); ++ ++ nxi->nx_state |= NXS_RELEASED; ++ kfree(nxi); ++ atomic_dec(&nx_global_ctotal); ++} ++ ++static void __shutdown_nx_info(struct nx_info *nxi) ++{ ++ nxi->nx_state |= NXS_SHUTDOWN; ++ vs_net_change(nxi, VSC_NETDOWN); ++} ++ ++/* exported stuff */ ++ ++void free_nx_info(struct nx_info *nxi) ++{ ++ /* context shutdown is mandatory */ ++ BUG_ON(nxi->nx_state != NXS_SHUTDOWN); ++ ++ /* context must not be hashed */ ++ BUG_ON(nxi->nx_state & NXS_HASHED); ++ ++ BUG_ON(atomic_read(&nxi->nx_usecnt)); ++ BUG_ON(atomic_read(&nxi->nx_tasks)); ++ ++ __dealloc_nx_info(nxi); ++} ++ ++ ++void __nx_set_lback(struct nx_info *nxi) ++{ ++ int nid = nxi->nx_id; ++ __be32 lback = htonl(INADDR_LOOPBACK ^ ((nid & 0xFFFF) << 8)); ++ ++ nxi->v4_lback.s_addr = lback; ++} ++ ++extern int __nx_inet_add_lback(__be32 addr); ++extern int __nx_inet_del_lback(__be32 addr); ++ ++ ++/* hash table for nx_info hash */ ++ ++#define NX_HASH_SIZE 13 ++ ++struct hlist_head nx_info_hash[NX_HASH_SIZE]; ++ ++static DEFINE_SPINLOCK(nx_info_hash_lock); ++ ++ ++static inline unsigned int __hashval(nid_t nid) ++{ ++ return (nid % NX_HASH_SIZE); ++} ++ ++ ++ ++/* __hash_nx_info() ++ ++ * add the nxi to the global hash table ++ * requires the hash_lock to be held */ ++ ++static inline void __hash_nx_info(struct nx_info *nxi) ++{ ++ struct hlist_head *head; ++ ++ vxd_assert_lock(&nx_info_hash_lock); ++ vxdprintk(VXD_CBIT(nid, 4), ++ "__hash_nx_info: %p[#%d]", nxi, nxi->nx_id); ++ ++ /* context must not be hashed */ ++ BUG_ON(nx_info_state(nxi, NXS_HASHED)); ++ ++ nxi->nx_state |= NXS_HASHED; ++ head = &nx_info_hash[__hashval(nxi->nx_id)]; ++ hlist_add_head(&nxi->nx_hlist, head); ++ atomic_inc(&nx_global_cactive); ++} ++ ++/* __unhash_nx_info() ++ ++ * remove the nxi from the global hash table ++ * requires the hash_lock to be held */ ++ ++static inline void __unhash_nx_info(struct nx_info *nxi) ++{ ++ vxd_assert_lock(&nx_info_hash_lock); ++ vxdprintk(VXD_CBIT(nid, 4), ++ "__unhash_nx_info: %p[#%d.%d.%d]", nxi, nxi->nx_id, ++ atomic_read(&nxi->nx_usecnt), atomic_read(&nxi->nx_tasks)); ++ ++ /* context must be hashed */ ++ BUG_ON(!nx_info_state(nxi, NXS_HASHED)); ++ /* but without tasks */ ++ BUG_ON(atomic_read(&nxi->nx_tasks)); ++ ++ nxi->nx_state &= ~NXS_HASHED; ++ hlist_del(&nxi->nx_hlist); ++ atomic_dec(&nx_global_cactive); ++} ++ ++ ++/* __lookup_nx_info() ++ ++ * requires the hash_lock to be held ++ * doesn't increment the nx_refcnt */ ++ ++static inline struct nx_info *__lookup_nx_info(nid_t nid) ++{ ++ struct hlist_head *head = &nx_info_hash[__hashval(nid)]; ++ struct hlist_node *pos; ++ struct nx_info *nxi; ++ ++ vxd_assert_lock(&nx_info_hash_lock); ++ hlist_for_each(pos, head) { ++ nxi = hlist_entry(pos, struct nx_info, nx_hlist); ++ ++ if (nxi->nx_id == nid) ++ goto found; ++ } ++ nxi = NULL; ++found: ++ vxdprintk(VXD_CBIT(nid, 0), ++ "__lookup_nx_info(#%u): %p[#%u]", ++ nid, nxi, nxi ? nxi->nx_id : 0); ++ return nxi; ++} ++ ++ ++/* __create_nx_info() ++ ++ * create the requested context ++ * get(), claim() and hash it */ ++ ++static struct nx_info *__create_nx_info(int id) ++{ ++ struct nx_info *new, *nxi = NULL; ++ ++ vxdprintk(VXD_CBIT(nid, 1), "create_nx_info(%d)*", id); ++ ++ if (!(new = __alloc_nx_info(id))) ++ return ERR_PTR(-ENOMEM); ++ ++ /* required to make dynamic xids unique */ ++ spin_lock(&nx_info_hash_lock); ++ ++ /* static context requested */ ++ if ((nxi = __lookup_nx_info(id))) { ++ vxdprintk(VXD_CBIT(nid, 0), ++ "create_nx_info(%d) = %p (already there)", id, nxi); ++ if (nx_info_flags(nxi, NXF_STATE_SETUP, 0)) ++ nxi = ERR_PTR(-EBUSY); ++ else ++ nxi = ERR_PTR(-EEXIST); ++ goto out_unlock; ++ } ++ /* new context */ ++ vxdprintk(VXD_CBIT(nid, 0), ++ "create_nx_info(%d) = %p (new)", id, new); ++ claim_nx_info(new, NULL); ++ __nx_set_lback(new); ++ __hash_nx_info(get_nx_info(new)); ++ nxi = new, new = NULL; ++ ++out_unlock: ++ spin_unlock(&nx_info_hash_lock); ++ if (new) ++ __dealloc_nx_info(new); ++ return nxi; ++} ++ ++ ++ ++/* exported stuff */ ++ ++ ++void unhash_nx_info(struct nx_info *nxi) ++{ ++ __shutdown_nx_info(nxi); ++ spin_lock(&nx_info_hash_lock); ++ __unhash_nx_info(nxi); ++ spin_unlock(&nx_info_hash_lock); ++} ++ ++/* lookup_nx_info() ++ ++ * search for a nx_info and get() it ++ * negative id means current */ ++ ++struct nx_info *lookup_nx_info(int id) ++{ ++ struct nx_info *nxi = NULL; ++ ++ if (id < 0) { ++ nxi = get_nx_info(current_nx_info()); ++ } else if (id > 1) { ++ spin_lock(&nx_info_hash_lock); ++ nxi = get_nx_info(__lookup_nx_info(id)); ++ spin_unlock(&nx_info_hash_lock); ++ } ++ return nxi; ++} ++ ++/* nid_is_hashed() ++ ++ * verify that nid is still hashed */ ++ ++int nid_is_hashed(nid_t nid) ++{ ++ int hashed; ++ ++ spin_lock(&nx_info_hash_lock); ++ hashed = (__lookup_nx_info(nid) != NULL); ++ spin_unlock(&nx_info_hash_lock); ++ return hashed; ++} ++ ++ ++#ifdef CONFIG_PROC_FS ++ ++/* get_nid_list() ++ ++ * get a subset of hashed nids for proc ++ * assumes size is at least one */ ++ ++int get_nid_list(int index, unsigned int *nids, int size) ++{ ++ int hindex, nr_nids = 0; ++ ++ /* only show current and children */ ++ if (!nx_check(0, VS_ADMIN | VS_WATCH)) { ++ if (index > 0) ++ return 0; ++ nids[nr_nids] = nx_current_nid(); ++ return 1; ++ } ++ ++ for (hindex = 0; hindex < NX_HASH_SIZE; hindex++) { ++ struct hlist_head *head = &nx_info_hash[hindex]; ++ struct hlist_node *pos; ++ ++ spin_lock(&nx_info_hash_lock); ++ hlist_for_each(pos, head) { ++ struct nx_info *nxi; ++ ++ if (--index > 0) ++ continue; ++ ++ nxi = hlist_entry(pos, struct nx_info, nx_hlist); ++ nids[nr_nids] = nxi->nx_id; ++ if (++nr_nids >= size) { ++ spin_unlock(&nx_info_hash_lock); ++ goto out; ++ } ++ } ++ /* keep the lock time short */ ++ spin_unlock(&nx_info_hash_lock); ++ } ++out: ++ return nr_nids; ++} ++#endif ++ ++ ++/* ++ * migrate task to new network ++ * gets nxi, puts old_nxi on change ++ */ ++ ++int nx_migrate_task(struct task_struct *p, struct nx_info *nxi) ++{ ++ struct nx_info *old_nxi; ++ int ret = 0; ++ ++ if (!p || !nxi) ++ BUG(); ++ ++ vxdprintk(VXD_CBIT(nid, 5), ++ "nx_migrate_task(%p,%p[#%d.%d.%d])", ++ p, nxi, nxi->nx_id, ++ atomic_read(&nxi->nx_usecnt), ++ atomic_read(&nxi->nx_tasks)); ++ ++ if (nx_info_flags(nxi, NXF_INFO_PRIVATE, 0) && ++ !nx_info_flags(nxi, NXF_STATE_SETUP, 0)) ++ return -EACCES; ++ ++ if (nx_info_state(nxi, NXS_SHUTDOWN)) ++ return -EFAULT; ++ ++ /* maybe disallow this completely? */ ++ old_nxi = task_get_nx_info(p); ++ if (old_nxi == nxi) ++ goto out; ++ ++ task_lock(p); ++ if (old_nxi) ++ clr_nx_info(&p->nx_info); ++ claim_nx_info(nxi, p); ++ set_nx_info(&p->nx_info, nxi); ++ p->nid = nxi->nx_id; ++ task_unlock(p); ++ ++ vxdprintk(VXD_CBIT(nid, 5), ++ "moved task %p into nxi:%p[#%d]", ++ p, nxi, nxi->nx_id); ++ ++ if (old_nxi) ++ release_nx_info(old_nxi, p); ++ ret = 0; ++out: ++ put_nx_info(old_nxi); ++ return ret; ++} ++ ++ ++void nx_set_persistent(struct nx_info *nxi) ++{ ++ vxdprintk(VXD_CBIT(nid, 6), ++ "nx_set_persistent(%p[#%d])", nxi, nxi->nx_id); ++ ++ get_nx_info(nxi); ++ claim_nx_info(nxi, NULL); ++} ++ ++void nx_clear_persistent(struct nx_info *nxi) ++{ ++ vxdprintk(VXD_CBIT(nid, 6), ++ "nx_clear_persistent(%p[#%d])", nxi, nxi->nx_id); ++ ++ release_nx_info(nxi, NULL); ++ put_nx_info(nxi); ++} ++ ++void nx_update_persistent(struct nx_info *nxi) ++{ ++ if (nx_info_flags(nxi, NXF_PERSISTENT, 0)) ++ nx_set_persistent(nxi); ++ else ++ nx_clear_persistent(nxi); ++} ++ ++/* vserver syscall commands below here */ ++ ++/* taks nid and nx_info functions */ ++ ++#include <asm/uaccess.h> ++ ++ ++int vc_task_nid(uint32_t id) ++{ ++ nid_t nid; ++ ++ if (id) { ++ struct task_struct *tsk; ++ ++ rcu_read_lock(); ++ tsk = find_task_by_real_pid(id); ++ nid = (tsk) ? tsk->nid : -ESRCH; ++ rcu_read_unlock(); ++ } else ++ nid = nx_current_nid(); ++ return nid; ++} ++ ++ ++int vc_nx_info(struct nx_info *nxi, void __user *data) ++{ ++ struct vcmd_nx_info_v0 vc_data; ++ ++ vc_data.nid = nxi->nx_id; ++ ++ if (copy_to_user(data, &vc_data, sizeof(vc_data))) ++ return -EFAULT; ++ return 0; ++} ++ ++ ++/* network functions */ ++ ++int vc_net_create(uint32_t nid, void __user *data) ++{ ++ struct vcmd_net_create vc_data = { .flagword = NXF_INIT_SET }; ++ struct nx_info *new_nxi; ++ int ret; ++ ++ if (data && copy_from_user(&vc_data, data, sizeof(vc_data))) ++ return -EFAULT; ++ ++ if ((nid > MAX_S_CONTEXT) || (nid < 2)) ++ return -EINVAL; ++ ++ new_nxi = __create_nx_info(nid); ++ if (IS_ERR(new_nxi)) ++ return PTR_ERR(new_nxi); ++ ++ /* initial flags */ ++ new_nxi->nx_flags = vc_data.flagword; ++ ++ ret = -ENOEXEC; ++ if (vs_net_change(new_nxi, VSC_NETUP)) ++ goto out; ++ ++ ret = nx_migrate_task(current, new_nxi); ++ if (ret) ++ goto out; ++ ++ /* return context id on success */ ++ ret = new_nxi->nx_id; ++ ++ /* get a reference for persistent contexts */ ++ if ((vc_data.flagword & NXF_PERSISTENT)) ++ nx_set_persistent(new_nxi); ++out: ++ release_nx_info(new_nxi, NULL); ++ put_nx_info(new_nxi); ++ return ret; ++} ++ ++ ++int vc_net_migrate(struct nx_info *nxi, void __user *data) ++{ ++ return nx_migrate_task(current, nxi); ++} ++ ++ ++ ++int do_add_v4_addr(struct nx_info *nxi, __be32 ip, __be32 ip2, __be32 mask, ++ uint16_t type, uint16_t flags) ++{ ++ struct nx_addr_v4 *nxa = &nxi->v4; ++ ++ if (NX_IPV4(nxi)) { ++ /* locate last entry */ ++ for (; nxa->next; nxa = nxa->next); ++ nxa->next = __alloc_nx_addr_v4(); ++ nxa = nxa->next; ++ ++ if (IS_ERR(nxa)) ++ return PTR_ERR(nxa); ++ } ++ ++ if (nxi->v4.next) ++ /* remove single ip for ip list */ ++ nxi->nx_flags &= ~NXF_SINGLE_IP; ++ ++ nxa->ip[0].s_addr = ip; ++ nxa->ip[1].s_addr = ip2; ++ nxa->mask.s_addr = mask; ++ nxa->type = type; ++ nxa->flags = flags; ++ return 0; ++} ++ ++int do_remove_v4_addr(struct nx_info *nxi, __be32 ip, __be32 ip2, __be32 mask, ++ uint16_t type, uint16_t flags) ++{ ++ struct nx_addr_v4 *nxa = &nxi->v4; ++ ++ switch (type) { ++/* case NXA_TYPE_ADDR: ++ break; */ ++ ++ case NXA_TYPE_ANY: ++ __dealloc_nx_addr_v4_all(xchg(&nxa->next, NULL)); ++ memset(nxa, 0, sizeof(*nxa)); ++ break; ++ ++ default: ++ return -EINVAL; ++ } ++ return 0; ++} ++ ++ ++int vc_net_add(struct nx_info *nxi, void __user *data) ++{ ++ struct vcmd_net_addr_v0 vc_data; ++ int index, ret = 0; ++ ++ if (data && copy_from_user(&vc_data, data, sizeof(vc_data))) ++ return -EFAULT; ++ ++ switch (vc_data.type) { ++ case NXA_TYPE_IPV4: ++ if ((vc_data.count < 1) || (vc_data.count > 4)) ++ return -EINVAL; ++ ++ index = 0; ++ while (index < vc_data.count) { ++ ret = do_add_v4_addr(nxi, vc_data.ip[index].s_addr, 0, ++ vc_data.mask[index].s_addr, NXA_TYPE_ADDR, 0); ++ if (ret) ++ return ret; ++ index++; ++ } ++ ret = index; ++ break; ++ ++ case NXA_TYPE_IPV4|NXA_MOD_BCAST: ++ nxi->v4_bcast = vc_data.ip[0]; ++ ret = 1; ++ break; ++ ++ case NXA_TYPE_IPV4|NXA_MOD_LBACK: ++ nxi->v4_lback = vc_data.ip[0]; ++ ret = 1; ++ break; ++ ++ default: ++ ret = -EINVAL; ++ break; ++ } ++ return ret; ++} ++ ++int vc_net_remove(struct nx_info *nxi, void __user *data) ++{ ++ struct vcmd_net_addr_v0 vc_data; ++ ++ if (data && copy_from_user(&vc_data, data, sizeof(vc_data))) ++ return -EFAULT; ++ ++ switch (vc_data.type) { ++ case NXA_TYPE_ANY: ++ __dealloc_nx_addr_v4_all(xchg(&nxi->v4.next, NULL)); ++ memset(&nxi->v4, 0, sizeof(nxi->v4)); ++ break; ++ ++ default: ++ return -EINVAL; ++ } ++ return 0; ++} ++ ++ ++int vc_net_add_ipv4_v1(struct nx_info *nxi, void __user *data) ++{ ++ struct vcmd_net_addr_ipv4_v1 vc_data; ++ ++ if (data && copy_from_user(&vc_data, data, sizeof(vc_data))) ++ return -EFAULT; ++ ++ switch (vc_data.type) { ++ case NXA_TYPE_ADDR: ++ case NXA_TYPE_MASK: ++ return do_add_v4_addr(nxi, vc_data.ip.s_addr, 0, ++ vc_data.mask.s_addr, vc_data.type, vc_data.flags); ++ ++ case NXA_TYPE_ADDR | NXA_MOD_BCAST: ++ nxi->v4_bcast = vc_data.ip; ++ break; ++ ++ case NXA_TYPE_ADDR | NXA_MOD_LBACK: ++ nxi->v4_lback = vc_data.ip; ++ break; ++ ++ default: ++ return -EINVAL; ++ } ++ return 0; ++} ++ ++int vc_net_add_ipv4(struct nx_info *nxi, void __user *data) ++{ ++ struct vcmd_net_addr_ipv4_v2 vc_data; ++ ++ if (data && copy_from_user(&vc_data, data, sizeof(vc_data))) ++ return -EFAULT; ++ ++ switch (vc_data.type) { ++ case NXA_TYPE_ADDR: ++ case NXA_TYPE_MASK: ++ case NXA_TYPE_RANGE: ++ return do_add_v4_addr(nxi, vc_data.ip.s_addr, vc_data.ip2.s_addr, ++ vc_data.mask.s_addr, vc_data.type, vc_data.flags); ++ ++ case NXA_TYPE_ADDR | NXA_MOD_BCAST: ++ nxi->v4_bcast = vc_data.ip; ++ break; ++ ++ case NXA_TYPE_ADDR | NXA_MOD_LBACK: ++ nxi->v4_lback = vc_data.ip; ++ break; ++ ++ default: ++ return -EINVAL; ++ } ++ return 0; ++} ++ ++int vc_net_rem_ipv4_v1(struct nx_info *nxi, void __user *data) ++{ ++ struct vcmd_net_addr_ipv4_v1 vc_data; ++ ++ if (data && copy_from_user(&vc_data, data, sizeof(vc_data))) ++ return -EFAULT; ++ ++ return do_remove_v4_addr(nxi, vc_data.ip.s_addr, 0, ++ vc_data.mask.s_addr, vc_data.type, vc_data.flags); ++} ++ ++int vc_net_rem_ipv4(struct nx_info *nxi, void __user *data) ++{ ++ struct vcmd_net_addr_ipv4_v2 vc_data; ++ ++ if (data && copy_from_user(&vc_data, data, sizeof(vc_data))) ++ return -EFAULT; ++ ++ return do_remove_v4_addr(nxi, vc_data.ip.s_addr, vc_data.ip2.s_addr, ++ vc_data.mask.s_addr, vc_data.type, vc_data.flags); ++} ++ ++#ifdef CONFIG_IPV6 ++ ++int do_add_v6_addr(struct nx_info *nxi, ++ struct in6_addr *ip, struct in6_addr *mask, ++ uint32_t prefix, uint16_t type, uint16_t flags) ++{ ++ struct nx_addr_v6 *nxa = &nxi->v6; ++ ++ if (NX_IPV6(nxi)) { ++ /* locate last entry */ ++ for (; nxa->next; nxa = nxa->next); ++ nxa->next = __alloc_nx_addr_v6(); ++ nxa = nxa->next; ++ ++ if (IS_ERR(nxa)) ++ return PTR_ERR(nxa); ++ } ++ ++ nxa->ip = *ip; ++ nxa->mask = *mask; ++ nxa->prefix = prefix; ++ nxa->type = type; ++ nxa->flags = flags; ++ return 0; ++} ++ ++ ++int vc_net_add_ipv6(struct nx_info *nxi, void __user *data) ++{ ++ struct vcmd_net_addr_ipv6_v1 vc_data; ++ ++ if (data && copy_from_user(&vc_data, data, sizeof(vc_data))) ++ return -EFAULT; ++ ++ switch (vc_data.type) { ++ case NXA_TYPE_ADDR: ++ memset(&vc_data.mask, ~0, sizeof(vc_data.mask)); ++ /* fallthrough */ ++ case NXA_TYPE_MASK: ++ return do_add_v6_addr(nxi, &vc_data.ip, &vc_data.mask, ++ vc_data.prefix, vc_data.type, vc_data.flags); ++ default: ++ return -EINVAL; ++ } ++ return 0; ++} ++ ++int vc_net_remove_ipv6(struct nx_info *nxi, void __user *data) ++{ ++ struct vcmd_net_addr_ipv6_v1 vc_data; ++ ++ if (data && copy_from_user(&vc_data, data, sizeof(vc_data))) ++ return -EFAULT; ++ ++ switch (vc_data.type) { ++ case NXA_TYPE_ANY: ++ __dealloc_nx_addr_v6_all(xchg(&nxi->v6.next, NULL)); ++ memset(&nxi->v6, 0, sizeof(nxi->v6)); ++ break; ++ ++ default: ++ return -EINVAL; ++ } ++ return 0; ++} ++ ++#endif /* CONFIG_IPV6 */ ++ ++ ++int vc_get_nflags(struct nx_info *nxi, void __user *data) ++{ ++ struct vcmd_net_flags_v0 vc_data; ++ ++ vc_data.flagword = nxi->nx_flags; ++ ++ /* special STATE flag handling */ ++ vc_data.mask = vs_mask_flags(~0ULL, nxi->nx_flags, NXF_ONE_TIME); ++ ++ if (copy_to_user(data, &vc_data, sizeof(vc_data))) ++ return -EFAULT; ++ return 0; ++} ++ ++int vc_set_nflags(struct nx_info *nxi, void __user *data) ++{ ++ struct vcmd_net_flags_v0 vc_data; ++ uint64_t mask, trigger; ++ ++ if (copy_from_user(&vc_data, data, sizeof(vc_data))) ++ return -EFAULT; ++ ++ /* special STATE flag handling */ ++ mask = vs_mask_mask(vc_data.mask, nxi->nx_flags, NXF_ONE_TIME); ++ trigger = (mask & nxi->nx_flags) ^ (mask & vc_data.flagword); ++ ++ nxi->nx_flags = vs_mask_flags(nxi->nx_flags, ++ vc_data.flagword, mask); ++ if (trigger & NXF_PERSISTENT) ++ nx_update_persistent(nxi); ++ ++ return 0; ++} ++ ++int vc_get_ncaps(struct nx_info *nxi, void __user *data) ++{ ++ struct vcmd_net_caps_v0 vc_data; ++ ++ vc_data.ncaps = nxi->nx_ncaps; ++ vc_data.cmask = ~0ULL; ++ ++ if (copy_to_user(data, &vc_data, sizeof(vc_data))) ++ return -EFAULT; ++ return 0; ++} ++ ++int vc_set_ncaps(struct nx_info *nxi, void __user *data) ++{ ++ struct vcmd_net_caps_v0 vc_data; ++ ++ if (copy_from_user(&vc_data, data, sizeof(vc_data))) ++ return -EFAULT; ++ ++ nxi->nx_ncaps = vs_mask_flags(nxi->nx_ncaps, ++ vc_data.ncaps, vc_data.cmask); ++ return 0; ++} ++ ++ ++#include <linux/module.h> ++ ++module_init(init_network); ++ ++EXPORT_SYMBOL_GPL(free_nx_info); ++EXPORT_SYMBOL_GPL(unhash_nx_info); ++ +diff -NurpP --minimal linux-3.0.9/kernel/vserver/proc.c linux-3.0.9-vs2.3.2.1/kernel/vserver/proc.c +--- linux-3.0.9/kernel/vserver/proc.c 1970-01-01 01:00:00.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/kernel/vserver/proc.c 2011-08-01 19:57:21.000000000 +0200 +@@ -0,0 +1,1103 @@ ++/* ++ * linux/kernel/vserver/proc.c ++ * ++ * Virtual Context Support ++ * ++ * Copyright (C) 2003-2011 Herbert Pötzl ++ * ++ * V0.01 basic structure ++ * V0.02 adaptation vs1.3.0 ++ * V0.03 proc permissions ++ * V0.04 locking/generic ++ * V0.05 next generation procfs ++ * V0.06 inode validation ++ * V0.07 generic rewrite vid ++ * V0.08 remove inode type ++ * V0.09 added u/wmask info ++ * ++ */ ++ ++#include <linux/proc_fs.h> ++#include <linux/fs_struct.h> ++#include <linux/mount.h> ++#include <asm/unistd.h> ++ ++#include <linux/vs_context.h> ++#include <linux/vs_network.h> ++#include <linux/vs_cvirt.h> ++ ++#include <linux/in.h> ++#include <linux/inetdevice.h> ++#include <linux/vs_inet.h> ++#include <linux/vs_inet6.h> ++ ++#include <linux/vserver/global.h> ++ ++#include "cvirt_proc.h" ++#include "cacct_proc.h" ++#include "limit_proc.h" ++#include "sched_proc.h" ++#include "vci_config.h" ++ ++ ++static inline char *print_cap_t(char *buffer, kernel_cap_t *c) ++{ ++ unsigned __capi; ++ ++ CAP_FOR_EACH_U32(__capi) { ++ buffer += sprintf(buffer, "%08x", ++ c->cap[(_KERNEL_CAPABILITY_U32S-1) - __capi]); ++ } ++ return buffer; ++} ++ ++ ++static struct proc_dir_entry *proc_virtual; ++ ++static struct proc_dir_entry *proc_virtnet; ++ ++ ++/* first the actual feeds */ ++ ++ ++static int proc_vci(char *buffer) ++{ ++ return sprintf(buffer, ++ "VCIVersion:\t%04x:%04x\n" ++ "VCISyscall:\t%d\n" ++ "VCIKernel:\t%08x\n", ++ VCI_VERSION >> 16, ++ VCI_VERSION & 0xFFFF, ++ __NR_vserver, ++ vci_kernel_config()); ++} ++ ++static int proc_virtual_info(char *buffer) ++{ ++ return proc_vci(buffer); ++} ++ ++static int proc_virtual_status(char *buffer) ++{ ++ return sprintf(buffer, ++ "#CTotal:\t%d\n" ++ "#CActive:\t%d\n" ++ "#NSProxy:\t%d\t%d %d %d %d %d %d\n" ++ "#InitTask:\t%d\t%d %d\n", ++ atomic_read(&vx_global_ctotal), ++ atomic_read(&vx_global_cactive), ++ atomic_read(&vs_global_nsproxy), ++ atomic_read(&vs_global_fs), ++ atomic_read(&vs_global_mnt_ns), ++ atomic_read(&vs_global_uts_ns), ++ atomic_read(&nr_ipc_ns), ++ atomic_read(&vs_global_user_ns), ++ atomic_read(&vs_global_pid_ns), ++ atomic_read(&init_task.usage), ++ atomic_read(&init_task.nsproxy->count), ++ init_task.fs->users); ++} ++ ++ ++int proc_vxi_info(struct vx_info *vxi, char *buffer) ++{ ++ int length; ++ ++ length = sprintf(buffer, ++ "ID:\t%d\n" ++ "Info:\t%p\n" ++ "Init:\t%d\n" ++ "OOM:\t%lld\n", ++ vxi->vx_id, ++ vxi, ++ vxi->vx_initpid, ++ vxi->vx_badness_bias); ++ return length; ++} ++ ++int proc_vxi_status(struct vx_info *vxi, char *buffer) ++{ ++ char *orig = buffer; ++ ++ buffer += sprintf(buffer, ++ "UseCnt:\t%d\n" ++ "Tasks:\t%d\n" ++ "Flags:\t%016llx\n", ++ atomic_read(&vxi->vx_usecnt), ++ atomic_read(&vxi->vx_tasks), ++ (unsigned long long)vxi->vx_flags); ++ ++ buffer += sprintf(buffer, "BCaps:\t"); ++ buffer = print_cap_t(buffer, &vxi->vx_bcaps); ++ buffer += sprintf(buffer, "\n"); ++ ++ buffer += sprintf(buffer, ++ "CCaps:\t%016llx\n" ++ "Umask:\t%16llx\n" ++ "Wmask:\t%16llx\n" ++ "Spaces:\t%08lx %08lx\n", ++ (unsigned long long)vxi->vx_ccaps, ++ (unsigned long long)vxi->vx_umask, ++ (unsigned long long)vxi->vx_wmask, ++ vxi->space[0].vx_nsmask, vxi->space[1].vx_nsmask); ++ return buffer - orig; ++} ++ ++int proc_vxi_limit(struct vx_info *vxi, char *buffer) ++{ ++ return vx_info_proc_limit(&vxi->limit, buffer); ++} ++ ++int proc_vxi_sched(struct vx_info *vxi, char *buffer) ++{ ++ int cpu, length; ++ ++ length = vx_info_proc_sched(&vxi->sched, buffer); ++ for_each_online_cpu(cpu) { ++ length += vx_info_proc_sched_pc( ++ &vx_per_cpu(vxi, sched_pc, cpu), ++ buffer + length, cpu); ++ } ++ return length; ++} ++ ++int proc_vxi_nsproxy0(struct vx_info *vxi, char *buffer) ++{ ++ return vx_info_proc_nsproxy(vxi->space[0].vx_nsproxy, buffer); ++} ++ ++int proc_vxi_nsproxy1(struct vx_info *vxi, char *buffer) ++{ ++ return vx_info_proc_nsproxy(vxi->space[1].vx_nsproxy, buffer); ++} ++ ++int proc_vxi_cvirt(struct vx_info *vxi, char *buffer) ++{ ++ int cpu, length; ++ ++ vx_update_load(vxi); ++ length = vx_info_proc_cvirt(&vxi->cvirt, buffer); ++ for_each_online_cpu(cpu) { ++ length += vx_info_proc_cvirt_pc( ++ &vx_per_cpu(vxi, cvirt_pc, cpu), ++ buffer + length, cpu); ++ } ++ return length; ++} ++ ++int proc_vxi_cacct(struct vx_info *vxi, char *buffer) ++{ ++ return vx_info_proc_cacct(&vxi->cacct, buffer); ++} ++ ++ ++static int proc_virtnet_info(char *buffer) ++{ ++ return proc_vci(buffer); ++} ++ ++static int proc_virtnet_status(char *buffer) ++{ ++ return sprintf(buffer, ++ "#CTotal:\t%d\n" ++ "#CActive:\t%d\n", ++ atomic_read(&nx_global_ctotal), ++ atomic_read(&nx_global_cactive)); ++} ++ ++int proc_nxi_info(struct nx_info *nxi, char *buffer) ++{ ++ struct nx_addr_v4 *v4a; ++#ifdef CONFIG_IPV6 ++ struct nx_addr_v6 *v6a; ++#endif ++ int length, i; ++ ++ length = sprintf(buffer, ++ "ID:\t%d\n" ++ "Info:\t%p\n" ++ "Bcast:\t" NIPQUAD_FMT "\n" ++ "Lback:\t" NIPQUAD_FMT "\n", ++ nxi->nx_id, ++ nxi, ++ NIPQUAD(nxi->v4_bcast.s_addr), ++ NIPQUAD(nxi->v4_lback.s_addr)); ++ ++ if (!NX_IPV4(nxi)) ++ goto skip_v4; ++ for (i = 0, v4a = &nxi->v4; v4a; i++, v4a = v4a->next) ++ length += sprintf(buffer + length, "%d:\t" NXAV4_FMT "\n", ++ i, NXAV4(v4a)); ++skip_v4: ++#ifdef CONFIG_IPV6 ++ if (!NX_IPV6(nxi)) ++ goto skip_v6; ++ for (i = 0, v6a = &nxi->v6; v6a; i++, v6a = v6a->next) ++ length += sprintf(buffer + length, "%d:\t" NXAV6_FMT "\n", ++ i, NXAV6(v6a)); ++skip_v6: ++#endif ++ return length; ++} ++ ++int proc_nxi_status(struct nx_info *nxi, char *buffer) ++{ ++ int length; ++ ++ length = sprintf(buffer, ++ "UseCnt:\t%d\n" ++ "Tasks:\t%d\n" ++ "Flags:\t%016llx\n" ++ "NCaps:\t%016llx\n", ++ atomic_read(&nxi->nx_usecnt), ++ atomic_read(&nxi->nx_tasks), ++ (unsigned long long)nxi->nx_flags, ++ (unsigned long long)nxi->nx_ncaps); ++ return length; ++} ++ ++ ++ ++/* here the inode helpers */ ++ ++struct vs_entry { ++ int len; ++ char *name; ++ mode_t mode; ++ struct inode_operations *iop; ++ struct file_operations *fop; ++ union proc_op op; ++}; ++ ++static struct inode *vs_proc_make_inode(struct super_block *sb, struct vs_entry *p) ++{ ++ struct inode *inode = new_inode(sb); ++ ++ if (!inode) ++ goto out; ++ ++ inode->i_mode = p->mode; ++ if (p->iop) ++ inode->i_op = p->iop; ++ if (p->fop) ++ inode->i_fop = p->fop; ++ ++ inode->i_nlink = (p->mode & S_IFDIR) ? 2 : 1; ++ inode->i_flags |= S_IMMUTABLE; ++ ++ inode->i_mtime = inode->i_atime = inode->i_ctime = CURRENT_TIME; ++ ++ inode->i_uid = 0; ++ inode->i_gid = 0; ++ inode->i_tag = 0; ++out: ++ return inode; ++} ++ ++static struct dentry *vs_proc_instantiate(struct inode *dir, ++ struct dentry *dentry, int id, void *ptr) ++{ ++ struct vs_entry *p = ptr; ++ struct inode *inode = vs_proc_make_inode(dir->i_sb, p); ++ struct dentry *error = ERR_PTR(-EINVAL); ++ ++ if (!inode) ++ goto out; ++ ++ PROC_I(inode)->op = p->op; ++ PROC_I(inode)->fd = id; ++ d_add(dentry, inode); ++ error = NULL; ++out: ++ return error; ++} ++ ++/* Lookups */ ++ ++typedef struct dentry *instantiate_t(struct inode *, struct dentry *, int, void *); ++ ++/* ++ * Fill a directory entry. ++ * ++ * If possible create the dcache entry and derive our inode number and ++ * file type from dcache entry. ++ * ++ * Since all of the proc inode numbers are dynamically generated, the inode ++ * numbers do not exist until the inode is cache. This means creating the ++ * the dcache entry in readdir is necessary to keep the inode numbers ++ * reported by readdir in sync with the inode numbers reported ++ * by stat. ++ */ ++static int proc_fill_cache(struct file *filp, void *dirent, filldir_t filldir, ++ char *name, int len, instantiate_t instantiate, int id, void *ptr) ++{ ++ struct dentry *child, *dir = filp->f_dentry; ++ struct inode *inode; ++ struct qstr qname; ++ ino_t ino = 0; ++ unsigned type = DT_UNKNOWN; ++ ++ qname.name = name; ++ qname.len = len; ++ qname.hash = full_name_hash(name, len); ++ ++ child = d_lookup(dir, &qname); ++ if (!child) { ++ struct dentry *new; ++ new = d_alloc(dir, &qname); ++ if (new) { ++ child = instantiate(dir->d_inode, new, id, ptr); ++ if (child) ++ dput(new); ++ else ++ child = new; ++ } ++ } ++ if (!child || IS_ERR(child) || !child->d_inode) ++ goto end_instantiate; ++ inode = child->d_inode; ++ if (inode) { ++ ino = inode->i_ino; ++ type = inode->i_mode >> 12; ++ } ++ dput(child); ++end_instantiate: ++ if (!ino) ++ ino = find_inode_number(dir, &qname); ++ if (!ino) ++ ino = 1; ++ return filldir(dirent, name, len, filp->f_pos, ino, type); ++} ++ ++ ++ ++/* get and revalidate vx_info/xid */ ++ ++static inline ++struct vx_info *get_proc_vx_info(struct inode *inode) ++{ ++ return lookup_vx_info(PROC_I(inode)->fd); ++} ++ ++static int proc_xid_revalidate(struct dentry *dentry, struct nameidata *nd) ++{ ++ struct inode *inode = dentry->d_inode; ++ xid_t xid = PROC_I(inode)->fd; ++ ++ if (!xid || xid_is_hashed(xid)) ++ return 1; ++ d_drop(dentry); ++ return 0; ++} ++ ++ ++/* get and revalidate nx_info/nid */ ++ ++static int proc_nid_revalidate(struct dentry *dentry, struct nameidata *nd) ++{ ++ struct inode *inode = dentry->d_inode; ++ nid_t nid = PROC_I(inode)->fd; ++ ++ if (!nid || nid_is_hashed(nid)) ++ return 1; ++ d_drop(dentry); ++ return 0; ++} ++ ++ ++ ++#define PROC_BLOCK_SIZE (PAGE_SIZE - 1024) ++ ++static ssize_t proc_vs_info_read(struct file *file, char __user *buf, ++ size_t count, loff_t *ppos) ++{ ++ struct inode *inode = file->f_dentry->d_inode; ++ unsigned long page; ++ ssize_t length = 0; ++ ++ if (count > PROC_BLOCK_SIZE) ++ count = PROC_BLOCK_SIZE; ++ ++ /* fade that out as soon as stable */ ++ WARN_ON(PROC_I(inode)->fd); ++ ++ if (!(page = __get_free_page(GFP_KERNEL))) ++ return -ENOMEM; ++ ++ BUG_ON(!PROC_I(inode)->op.proc_vs_read); ++ length = PROC_I(inode)->op.proc_vs_read((char *)page); ++ ++ if (length >= 0) ++ length = simple_read_from_buffer(buf, count, ppos, ++ (char *)page, length); ++ ++ free_page(page); ++ return length; ++} ++ ++static ssize_t proc_vx_info_read(struct file *file, char __user *buf, ++ size_t count, loff_t *ppos) ++{ ++ struct inode *inode = file->f_dentry->d_inode; ++ struct vx_info *vxi = NULL; ++ xid_t xid = PROC_I(inode)->fd; ++ unsigned long page; ++ ssize_t length = 0; ++ ++ if (count > PROC_BLOCK_SIZE) ++ count = PROC_BLOCK_SIZE; ++ ++ /* fade that out as soon as stable */ ++ WARN_ON(!xid); ++ vxi = lookup_vx_info(xid); ++ if (!vxi) ++ goto out; ++ ++ length = -ENOMEM; ++ if (!(page = __get_free_page(GFP_KERNEL))) ++ goto out_put; ++ ++ BUG_ON(!PROC_I(inode)->op.proc_vxi_read); ++ length = PROC_I(inode)->op.proc_vxi_read(vxi, (char *)page); ++ ++ if (length >= 0) ++ length = simple_read_from_buffer(buf, count, ppos, ++ (char *)page, length); ++ ++ free_page(page); ++out_put: ++ put_vx_info(vxi); ++out: ++ return length; ++} ++ ++static ssize_t proc_nx_info_read(struct file *file, char __user *buf, ++ size_t count, loff_t *ppos) ++{ ++ struct inode *inode = file->f_dentry->d_inode; ++ struct nx_info *nxi = NULL; ++ nid_t nid = PROC_I(inode)->fd; ++ unsigned long page; ++ ssize_t length = 0; ++ ++ if (count > PROC_BLOCK_SIZE) ++ count = PROC_BLOCK_SIZE; ++ ++ /* fade that out as soon as stable */ ++ WARN_ON(!nid); ++ nxi = lookup_nx_info(nid); ++ if (!nxi) ++ goto out; ++ ++ length = -ENOMEM; ++ if (!(page = __get_free_page(GFP_KERNEL))) ++ goto out_put; ++ ++ BUG_ON(!PROC_I(inode)->op.proc_nxi_read); ++ length = PROC_I(inode)->op.proc_nxi_read(nxi, (char *)page); ++ ++ if (length >= 0) ++ length = simple_read_from_buffer(buf, count, ppos, ++ (char *)page, length); ++ ++ free_page(page); ++out_put: ++ put_nx_info(nxi); ++out: ++ return length; ++} ++ ++ ++ ++/* here comes the lower level */ ++ ++ ++#define NOD(NAME, MODE, IOP, FOP, OP) { \ ++ .len = sizeof(NAME) - 1, \ ++ .name = (NAME), \ ++ .mode = MODE, \ ++ .iop = IOP, \ ++ .fop = FOP, \ ++ .op = OP, \ ++} ++ ++ ++#define DIR(NAME, MODE, OTYPE) \ ++ NOD(NAME, (S_IFDIR | (MODE)), \ ++ &proc_ ## OTYPE ## _inode_operations, \ ++ &proc_ ## OTYPE ## _file_operations, { } ) ++ ++#define INF(NAME, MODE, OTYPE) \ ++ NOD(NAME, (S_IFREG | (MODE)), NULL, \ ++ &proc_vs_info_file_operations, \ ++ { .proc_vs_read = &proc_##OTYPE } ) ++ ++#define VINF(NAME, MODE, OTYPE) \ ++ NOD(NAME, (S_IFREG | (MODE)), NULL, \ ++ &proc_vx_info_file_operations, \ ++ { .proc_vxi_read = &proc_##OTYPE } ) ++ ++#define NINF(NAME, MODE, OTYPE) \ ++ NOD(NAME, (S_IFREG | (MODE)), NULL, \ ++ &proc_nx_info_file_operations, \ ++ { .proc_nxi_read = &proc_##OTYPE } ) ++ ++ ++static struct file_operations proc_vs_info_file_operations = { ++ .read = proc_vs_info_read, ++}; ++ ++static struct file_operations proc_vx_info_file_operations = { ++ .read = proc_vx_info_read, ++}; ++ ++static struct dentry_operations proc_xid_dentry_operations = { ++ .d_revalidate = proc_xid_revalidate, ++}; ++ ++static struct vs_entry vx_base_stuff[] = { ++ VINF("info", S_IRUGO, vxi_info), ++ VINF("status", S_IRUGO, vxi_status), ++ VINF("limit", S_IRUGO, vxi_limit), ++ VINF("sched", S_IRUGO, vxi_sched), ++ VINF("nsproxy", S_IRUGO, vxi_nsproxy0), ++ VINF("nsproxy1",S_IRUGO, vxi_nsproxy1), ++ VINF("cvirt", S_IRUGO, vxi_cvirt), ++ VINF("cacct", S_IRUGO, vxi_cacct), ++ {} ++}; ++ ++ ++ ++ ++static struct dentry *proc_xid_instantiate(struct inode *dir, ++ struct dentry *dentry, int id, void *ptr) ++{ ++ dentry->d_op = &proc_xid_dentry_operations; ++ return vs_proc_instantiate(dir, dentry, id, ptr); ++} ++ ++static struct dentry *proc_xid_lookup(struct inode *dir, ++ struct dentry *dentry, struct nameidata *nd) ++{ ++ struct vs_entry *p = vx_base_stuff; ++ struct dentry *error = ERR_PTR(-ENOENT); ++ ++ for (; p->name; p++) { ++ if (p->len != dentry->d_name.len) ++ continue; ++ if (!memcmp(dentry->d_name.name, p->name, p->len)) ++ break; ++ } ++ if (!p->name) ++ goto out; ++ ++ error = proc_xid_instantiate(dir, dentry, PROC_I(dir)->fd, p); ++out: ++ return error; ++} ++ ++static int proc_xid_readdir(struct file *filp, ++ void *dirent, filldir_t filldir) ++{ ++ struct dentry *dentry = filp->f_dentry; ++ struct inode *inode = dentry->d_inode; ++ struct vs_entry *p = vx_base_stuff; ++ int size = sizeof(vx_base_stuff) / sizeof(struct vs_entry); ++ int pos, index; ++ u64 ino; ++ ++ pos = filp->f_pos; ++ switch (pos) { ++ case 0: ++ ino = inode->i_ino; ++ if (filldir(dirent, ".", 1, pos, ino, DT_DIR) < 0) ++ goto out; ++ pos++; ++ /* fall through */ ++ case 1: ++ ino = parent_ino(dentry); ++ if (filldir(dirent, "..", 2, pos, ino, DT_DIR) < 0) ++ goto out; ++ pos++; ++ /* fall through */ ++ default: ++ index = pos - 2; ++ if (index >= size) ++ goto out; ++ for (p += index; p->name; p++) { ++ if (proc_fill_cache(filp, dirent, filldir, p->name, p->len, ++ vs_proc_instantiate, PROC_I(inode)->fd, p)) ++ goto out; ++ pos++; ++ } ++ } ++out: ++ filp->f_pos = pos; ++ return 1; ++} ++ ++ ++ ++static struct file_operations proc_nx_info_file_operations = { ++ .read = proc_nx_info_read, ++}; ++ ++static struct dentry_operations proc_nid_dentry_operations = { ++ .d_revalidate = proc_nid_revalidate, ++}; ++ ++static struct vs_entry nx_base_stuff[] = { ++ NINF("info", S_IRUGO, nxi_info), ++ NINF("status", S_IRUGO, nxi_status), ++ {} ++}; ++ ++ ++static struct dentry *proc_nid_instantiate(struct inode *dir, ++ struct dentry *dentry, int id, void *ptr) ++{ ++ dentry->d_op = &proc_nid_dentry_operations; ++ return vs_proc_instantiate(dir, dentry, id, ptr); ++} ++ ++static struct dentry *proc_nid_lookup(struct inode *dir, ++ struct dentry *dentry, struct nameidata *nd) ++{ ++ struct vs_entry *p = nx_base_stuff; ++ struct dentry *error = ERR_PTR(-ENOENT); ++ ++ for (; p->name; p++) { ++ if (p->len != dentry->d_name.len) ++ continue; ++ if (!memcmp(dentry->d_name.name, p->name, p->len)) ++ break; ++ } ++ if (!p->name) ++ goto out; ++ ++ error = proc_nid_instantiate(dir, dentry, PROC_I(dir)->fd, p); ++out: ++ return error; ++} ++ ++static int proc_nid_readdir(struct file *filp, ++ void *dirent, filldir_t filldir) ++{ ++ struct dentry *dentry = filp->f_dentry; ++ struct inode *inode = dentry->d_inode; ++ struct vs_entry *p = nx_base_stuff; ++ int size = sizeof(nx_base_stuff) / sizeof(struct vs_entry); ++ int pos, index; ++ u64 ino; ++ ++ pos = filp->f_pos; ++ switch (pos) { ++ case 0: ++ ino = inode->i_ino; ++ if (filldir(dirent, ".", 1, pos, ino, DT_DIR) < 0) ++ goto out; ++ pos++; ++ /* fall through */ ++ case 1: ++ ino = parent_ino(dentry); ++ if (filldir(dirent, "..", 2, pos, ino, DT_DIR) < 0) ++ goto out; ++ pos++; ++ /* fall through */ ++ default: ++ index = pos - 2; ++ if (index >= size) ++ goto out; ++ for (p += index; p->name; p++) { ++ if (proc_fill_cache(filp, dirent, filldir, p->name, p->len, ++ vs_proc_instantiate, PROC_I(inode)->fd, p)) ++ goto out; ++ pos++; ++ } ++ } ++out: ++ filp->f_pos = pos; ++ return 1; ++} ++ ++ ++#define MAX_MULBY10 ((~0U - 9) / 10) ++ ++static inline int atovid(const char *str, int len) ++{ ++ int vid, c; ++ ++ vid = 0; ++ while (len-- > 0) { ++ c = *str - '0'; ++ str++; ++ if (c > 9) ++ return -1; ++ if (vid >= MAX_MULBY10) ++ return -1; ++ vid *= 10; ++ vid += c; ++ if (!vid) ++ return -1; ++ } ++ return vid; ++} ++ ++/* now the upper level (virtual) */ ++ ++ ++static struct file_operations proc_xid_file_operations = { ++ .read = generic_read_dir, ++ .readdir = proc_xid_readdir, ++}; ++ ++static struct inode_operations proc_xid_inode_operations = { ++ .lookup = proc_xid_lookup, ++}; ++ ++static struct vs_entry vx_virtual_stuff[] = { ++ INF("info", S_IRUGO, virtual_info), ++ INF("status", S_IRUGO, virtual_status), ++ DIR(NULL, S_IRUGO | S_IXUGO, xid), ++}; ++ ++ ++static struct dentry *proc_virtual_lookup(struct inode *dir, ++ struct dentry *dentry, struct nameidata *nd) ++{ ++ struct vs_entry *p = vx_virtual_stuff; ++ struct dentry *error = ERR_PTR(-ENOENT); ++ int id = 0; ++ ++ for (; p->name; p++) { ++ if (p->len != dentry->d_name.len) ++ continue; ++ if (!memcmp(dentry->d_name.name, p->name, p->len)) ++ break; ++ } ++ if (p->name) ++ goto instantiate; ++ ++ id = atovid(dentry->d_name.name, dentry->d_name.len); ++ if ((id < 0) || !xid_is_hashed(id)) ++ goto out; ++ ++instantiate: ++ error = proc_xid_instantiate(dir, dentry, id, p); ++out: ++ return error; ++} ++ ++static struct file_operations proc_nid_file_operations = { ++ .read = generic_read_dir, ++ .readdir = proc_nid_readdir, ++}; ++ ++static struct inode_operations proc_nid_inode_operations = { ++ .lookup = proc_nid_lookup, ++}; ++ ++static struct vs_entry nx_virtnet_stuff[] = { ++ INF("info", S_IRUGO, virtnet_info), ++ INF("status", S_IRUGO, virtnet_status), ++ DIR(NULL, S_IRUGO | S_IXUGO, nid), ++}; ++ ++ ++static struct dentry *proc_virtnet_lookup(struct inode *dir, ++ struct dentry *dentry, struct nameidata *nd) ++{ ++ struct vs_entry *p = nx_virtnet_stuff; ++ struct dentry *error = ERR_PTR(-ENOENT); ++ int id = 0; ++ ++ for (; p->name; p++) { ++ if (p->len != dentry->d_name.len) ++ continue; ++ if (!memcmp(dentry->d_name.name, p->name, p->len)) ++ break; ++ } ++ if (p->name) ++ goto instantiate; ++ ++ id = atovid(dentry->d_name.name, dentry->d_name.len); ++ if ((id < 0) || !nid_is_hashed(id)) ++ goto out; ++ ++instantiate: ++ error = proc_nid_instantiate(dir, dentry, id, p); ++out: ++ return error; ++} ++ ++ ++#define PROC_MAXVIDS 32 ++ ++int proc_virtual_readdir(struct file *filp, ++ void *dirent, filldir_t filldir) ++{ ++ struct dentry *dentry = filp->f_dentry; ++ struct inode *inode = dentry->d_inode; ++ struct vs_entry *p = vx_virtual_stuff; ++ int size = sizeof(vx_virtual_stuff) / sizeof(struct vs_entry); ++ int pos, index; ++ unsigned int xid_array[PROC_MAXVIDS]; ++ char buf[PROC_NUMBUF]; ++ unsigned int nr_xids, i; ++ u64 ino; ++ ++ pos = filp->f_pos; ++ switch (pos) { ++ case 0: ++ ino = inode->i_ino; ++ if (filldir(dirent, ".", 1, pos, ino, DT_DIR) < 0) ++ goto out; ++ pos++; ++ /* fall through */ ++ case 1: ++ ino = parent_ino(dentry); ++ if (filldir(dirent, "..", 2, pos, ino, DT_DIR) < 0) ++ goto out; ++ pos++; ++ /* fall through */ ++ default: ++ index = pos - 2; ++ if (index >= size) ++ goto entries; ++ for (p += index; p->name; p++) { ++ if (proc_fill_cache(filp, dirent, filldir, p->name, p->len, ++ vs_proc_instantiate, 0, p)) ++ goto out; ++ pos++; ++ } ++ entries: ++ index = pos - size; ++ p = &vx_virtual_stuff[size - 1]; ++ nr_xids = get_xid_list(index, xid_array, PROC_MAXVIDS); ++ for (i = 0; i < nr_xids; i++) { ++ int n, xid = xid_array[i]; ++ unsigned int j = PROC_NUMBUF; ++ ++ n = xid; ++ do ++ buf[--j] = '0' + (n % 10); ++ while (n /= 10); ++ ++ if (proc_fill_cache(filp, dirent, filldir, ++ buf + j, PROC_NUMBUF - j, ++ vs_proc_instantiate, xid, p)) ++ goto out; ++ pos++; ++ } ++ } ++out: ++ filp->f_pos = pos; ++ return 0; ++} ++ ++static int proc_virtual_getattr(struct vfsmount *mnt, ++ struct dentry *dentry, struct kstat *stat) ++{ ++ struct inode *inode = dentry->d_inode; ++ ++ generic_fillattr(inode, stat); ++ stat->nlink = 2 + atomic_read(&vx_global_cactive); ++ return 0; ++} ++ ++static struct file_operations proc_virtual_dir_operations = { ++ .read = generic_read_dir, ++ .readdir = proc_virtual_readdir, ++}; ++ ++static struct inode_operations proc_virtual_dir_inode_operations = { ++ .getattr = proc_virtual_getattr, ++ .lookup = proc_virtual_lookup, ++}; ++ ++ ++ ++ ++ ++int proc_virtnet_readdir(struct file *filp, ++ void *dirent, filldir_t filldir) ++{ ++ struct dentry *dentry = filp->f_dentry; ++ struct inode *inode = dentry->d_inode; ++ struct vs_entry *p = nx_virtnet_stuff; ++ int size = sizeof(nx_virtnet_stuff) / sizeof(struct vs_entry); ++ int pos, index; ++ unsigned int nid_array[PROC_MAXVIDS]; ++ char buf[PROC_NUMBUF]; ++ unsigned int nr_nids, i; ++ u64 ino; ++ ++ pos = filp->f_pos; ++ switch (pos) { ++ case 0: ++ ino = inode->i_ino; ++ if (filldir(dirent, ".", 1, pos, ino, DT_DIR) < 0) ++ goto out; ++ pos++; ++ /* fall through */ ++ case 1: ++ ino = parent_ino(dentry); ++ if (filldir(dirent, "..", 2, pos, ino, DT_DIR) < 0) ++ goto out; ++ pos++; ++ /* fall through */ ++ default: ++ index = pos - 2; ++ if (index >= size) ++ goto entries; ++ for (p += index; p->name; p++) { ++ if (proc_fill_cache(filp, dirent, filldir, p->name, p->len, ++ vs_proc_instantiate, 0, p)) ++ goto out; ++ pos++; ++ } ++ entries: ++ index = pos - size; ++ p = &nx_virtnet_stuff[size - 1]; ++ nr_nids = get_nid_list(index, nid_array, PROC_MAXVIDS); ++ for (i = 0; i < nr_nids; i++) { ++ int n, nid = nid_array[i]; ++ unsigned int j = PROC_NUMBUF; ++ ++ n = nid; ++ do ++ buf[--j] = '0' + (n % 10); ++ while (n /= 10); ++ ++ if (proc_fill_cache(filp, dirent, filldir, ++ buf + j, PROC_NUMBUF - j, ++ vs_proc_instantiate, nid, p)) ++ goto out; ++ pos++; ++ } ++ } ++out: ++ filp->f_pos = pos; ++ return 0; ++} ++ ++static int proc_virtnet_getattr(struct vfsmount *mnt, ++ struct dentry *dentry, struct kstat *stat) ++{ ++ struct inode *inode = dentry->d_inode; ++ ++ generic_fillattr(inode, stat); ++ stat->nlink = 2 + atomic_read(&nx_global_cactive); ++ return 0; ++} ++ ++static struct file_operations proc_virtnet_dir_operations = { ++ .read = generic_read_dir, ++ .readdir = proc_virtnet_readdir, ++}; ++ ++static struct inode_operations proc_virtnet_dir_inode_operations = { ++ .getattr = proc_virtnet_getattr, ++ .lookup = proc_virtnet_lookup, ++}; ++ ++ ++ ++void proc_vx_init(void) ++{ ++ struct proc_dir_entry *ent; ++ ++ ent = proc_mkdir("virtual", 0); ++ if (ent) { ++ ent->proc_fops = &proc_virtual_dir_operations; ++ ent->proc_iops = &proc_virtual_dir_inode_operations; ++ } ++ proc_virtual = ent; ++ ++ ent = proc_mkdir("virtnet", 0); ++ if (ent) { ++ ent->proc_fops = &proc_virtnet_dir_operations; ++ ent->proc_iops = &proc_virtnet_dir_inode_operations; ++ } ++ proc_virtnet = ent; ++} ++ ++ ++ ++ ++/* per pid info */ ++ ++ ++int proc_pid_vx_info(struct task_struct *p, char *buffer) ++{ ++ struct vx_info *vxi; ++ char *orig = buffer; ++ ++ buffer += sprintf(buffer, "XID:\t%d\n", vx_task_xid(p)); ++ ++ vxi = task_get_vx_info(p); ++ if (!vxi) ++ goto out; ++ ++ buffer += sprintf(buffer, "BCaps:\t"); ++ buffer = print_cap_t(buffer, &vxi->vx_bcaps); ++ buffer += sprintf(buffer, "\n"); ++ buffer += sprintf(buffer, "CCaps:\t%016llx\n", ++ (unsigned long long)vxi->vx_ccaps); ++ buffer += sprintf(buffer, "CFlags:\t%016llx\n", ++ (unsigned long long)vxi->vx_flags); ++ buffer += sprintf(buffer, "CIPid:\t%d\n", vxi->vx_initpid); ++ ++ put_vx_info(vxi); ++out: ++ return buffer - orig; ++} ++ ++ ++int proc_pid_nx_info(struct task_struct *p, char *buffer) ++{ ++ struct nx_info *nxi; ++ struct nx_addr_v4 *v4a; ++#ifdef CONFIG_IPV6 ++ struct nx_addr_v6 *v6a; ++#endif ++ char *orig = buffer; ++ int i; ++ ++ buffer += sprintf(buffer, "NID:\t%d\n", nx_task_nid(p)); ++ ++ nxi = task_get_nx_info(p); ++ if (!nxi) ++ goto out; ++ ++ buffer += sprintf(buffer, "NCaps:\t%016llx\n", ++ (unsigned long long)nxi->nx_ncaps); ++ buffer += sprintf(buffer, "NFlags:\t%016llx\n", ++ (unsigned long long)nxi->nx_flags); ++ ++ buffer += sprintf(buffer, ++ "V4Root[bcast]:\t" NIPQUAD_FMT "\n", ++ NIPQUAD(nxi->v4_bcast.s_addr)); ++ buffer += sprintf (buffer, ++ "V4Root[lback]:\t" NIPQUAD_FMT "\n", ++ NIPQUAD(nxi->v4_lback.s_addr)); ++ if (!NX_IPV4(nxi)) ++ goto skip_v4; ++ for (i = 0, v4a = &nxi->v4; v4a; i++, v4a = v4a->next) ++ buffer += sprintf(buffer, "V4Root[%d]:\t" NXAV4_FMT "\n", ++ i, NXAV4(v4a)); ++skip_v4: ++#ifdef CONFIG_IPV6 ++ if (!NX_IPV6(nxi)) ++ goto skip_v6; ++ for (i = 0, v6a = &nxi->v6; v6a; i++, v6a = v6a->next) ++ buffer += sprintf(buffer, "V6Root[%d]:\t" NXAV6_FMT "\n", ++ i, NXAV6(v6a)); ++skip_v6: ++#endif ++ put_nx_info(nxi); ++out: ++ return buffer - orig; ++} ++ +diff -NurpP --minimal linux-3.0.9/kernel/vserver/sched.c linux-3.0.9-vs2.3.2.1/kernel/vserver/sched.c +--- linux-3.0.9/kernel/vserver/sched.c 1970-01-01 01:00:00.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/kernel/vserver/sched.c 2011-06-10 22:11:24.000000000 +0200 +@@ -0,0 +1,82 @@ ++/* ++ * linux/kernel/vserver/sched.c ++ * ++ * Virtual Server: Scheduler Support ++ * ++ * Copyright (C) 2004-2010 Herbert Pötzl ++ * ++ * V0.01 adapted Sam Vilains version to 2.6.3 ++ * V0.02 removed legacy interface ++ * V0.03 changed vcmds to vxi arg ++ * V0.04 removed older and legacy interfaces ++ * V0.05 removed scheduler code/commands ++ * ++ */ ++ ++#include <linux/vs_context.h> ++#include <linux/vs_sched.h> ++#include <linux/vserver/sched_cmd.h> ++ ++#include <asm/uaccess.h> ++ ++ ++void vx_update_sched_param(struct _vx_sched *sched, ++ struct _vx_sched_pc *sched_pc) ++{ ++ sched_pc->prio_bias = sched->prio_bias; ++} ++ ++static int do_set_prio_bias(struct vx_info *vxi, struct vcmd_prio_bias *data) ++{ ++ int cpu; ++ ++ if (data->prio_bias > MAX_PRIO_BIAS) ++ data->prio_bias = MAX_PRIO_BIAS; ++ if (data->prio_bias < MIN_PRIO_BIAS) ++ data->prio_bias = MIN_PRIO_BIAS; ++ ++ if (data->cpu_id != ~0) { ++ vxi->sched.update = cpumask_of_cpu(data->cpu_id); ++ cpus_and(vxi->sched.update, cpu_online_map, ++ vxi->sched.update); ++ } else ++ vxi->sched.update = cpu_online_map; ++ ++ for_each_cpu_mask(cpu, vxi->sched.update) ++ vx_update_sched_param(&vxi->sched, ++ &vx_per_cpu(vxi, sched_pc, cpu)); ++ return 0; ++} ++ ++int vc_set_prio_bias(struct vx_info *vxi, void __user *data) ++{ ++ struct vcmd_prio_bias vc_data; ++ ++ if (copy_from_user(&vc_data, data, sizeof(vc_data))) ++ return -EFAULT; ++ ++ return do_set_prio_bias(vxi, &vc_data); ++} ++ ++int vc_get_prio_bias(struct vx_info *vxi, void __user *data) ++{ ++ struct vcmd_prio_bias vc_data; ++ struct _vx_sched_pc *pcd; ++ int cpu; ++ ++ if (copy_from_user(&vc_data, data, sizeof(vc_data))) ++ return -EFAULT; ++ ++ cpu = vc_data.cpu_id; ++ ++ if (!cpu_possible(cpu)) ++ return -EINVAL; ++ ++ pcd = &vx_per_cpu(vxi, sched_pc, cpu); ++ vc_data.prio_bias = pcd->prio_bias; ++ ++ if (copy_to_user(data, &vc_data, sizeof(vc_data))) ++ return -EFAULT; ++ return 0; ++} ++ +diff -NurpP --minimal linux-3.0.9/kernel/vserver/sched_init.h linux-3.0.9-vs2.3.2.1/kernel/vserver/sched_init.h +--- linux-3.0.9/kernel/vserver/sched_init.h 1970-01-01 01:00:00.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/kernel/vserver/sched_init.h 2011-06-10 22:11:24.000000000 +0200 +@@ -0,0 +1,27 @@ ++ ++static inline void vx_info_init_sched(struct _vx_sched *sched) ++{ ++ /* scheduling; hard code starting values as constants */ ++ sched->prio_bias = 0; ++} ++ ++static inline ++void vx_info_init_sched_pc(struct _vx_sched_pc *sched_pc, int cpu) ++{ ++ sched_pc->prio_bias = 0; ++ ++ sched_pc->user_ticks = 0; ++ sched_pc->sys_ticks = 0; ++ sched_pc->hold_ticks = 0; ++} ++ ++static inline void vx_info_exit_sched(struct _vx_sched *sched) ++{ ++ return; ++} ++ ++static inline ++void vx_info_exit_sched_pc(struct _vx_sched_pc *sched_pc, int cpu) ++{ ++ return; ++} +diff -NurpP --minimal linux-3.0.9/kernel/vserver/sched_proc.h linux-3.0.9-vs2.3.2.1/kernel/vserver/sched_proc.h +--- linux-3.0.9/kernel/vserver/sched_proc.h 1970-01-01 01:00:00.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/kernel/vserver/sched_proc.h 2011-06-10 22:11:24.000000000 +0200 +@@ -0,0 +1,32 @@ ++#ifndef _VX_SCHED_PROC_H ++#define _VX_SCHED_PROC_H ++ ++ ++static inline ++int vx_info_proc_sched(struct _vx_sched *sched, char *buffer) ++{ ++ int length = 0; ++ ++ length += sprintf(buffer, ++ "PrioBias:\t%8d\n", ++ sched->prio_bias); ++ return length; ++} ++ ++static inline ++int vx_info_proc_sched_pc(struct _vx_sched_pc *sched_pc, ++ char *buffer, int cpu) ++{ ++ int length = 0; ++ ++ length += sprintf(buffer + length, ++ "cpu %d: %lld %lld %lld", cpu, ++ (unsigned long long)sched_pc->user_ticks, ++ (unsigned long long)sched_pc->sys_ticks, ++ (unsigned long long)sched_pc->hold_ticks); ++ length += sprintf(buffer + length, ++ " %d\n", sched_pc->prio_bias); ++ return length; ++} ++ ++#endif /* _VX_SCHED_PROC_H */ +diff -NurpP --minimal linux-3.0.9/kernel/vserver/signal.c linux-3.0.9-vs2.3.2.1/kernel/vserver/signal.c +--- linux-3.0.9/kernel/vserver/signal.c 1970-01-01 01:00:00.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/kernel/vserver/signal.c 2011-06-10 22:11:24.000000000 +0200 +@@ -0,0 +1,134 @@ ++/* ++ * linux/kernel/vserver/signal.c ++ * ++ * Virtual Server: Signal Support ++ * ++ * Copyright (C) 2003-2007 Herbert Pötzl ++ * ++ * V0.01 broken out from vcontext V0.05 ++ * V0.02 changed vcmds to vxi arg ++ * V0.03 adjusted siginfo for kill ++ * ++ */ ++ ++#include <asm/uaccess.h> ++ ++#include <linux/vs_context.h> ++#include <linux/vs_pid.h> ++#include <linux/vserver/signal_cmd.h> ++ ++ ++int vx_info_kill(struct vx_info *vxi, int pid, int sig) ++{ ++ int retval, count = 0; ++ struct task_struct *p; ++ struct siginfo *sip = SEND_SIG_PRIV; ++ ++ retval = -ESRCH; ++ vxdprintk(VXD_CBIT(misc, 4), ++ "vx_info_kill(%p[#%d],%d,%d)*", ++ vxi, vxi->vx_id, pid, sig); ++ read_lock(&tasklist_lock); ++ switch (pid) { ++ case 0: ++ case -1: ++ for_each_process(p) { ++ int err = 0; ++ ++ if (vx_task_xid(p) != vxi->vx_id || p->pid <= 1 || ++ (pid && vxi->vx_initpid == p->pid)) ++ continue; ++ ++ err = group_send_sig_info(sig, sip, p); ++ ++count; ++ if (err != -EPERM) ++ retval = err; ++ } ++ break; ++ ++ case 1: ++ if (vxi->vx_initpid) { ++ pid = vxi->vx_initpid; ++ /* for now, only SIGINT to private init ... */ ++ if (!vx_info_flags(vxi, VXF_STATE_ADMIN, 0) && ++ /* ... as long as there are tasks left */ ++ (atomic_read(&vxi->vx_tasks) > 1)) ++ sig = SIGINT; ++ } ++ /* fallthrough */ ++ default: ++ rcu_read_lock(); ++ p = find_task_by_real_pid(pid); ++ rcu_read_unlock(); ++ if (p) { ++ if (vx_task_xid(p) == vxi->vx_id) ++ retval = group_send_sig_info(sig, sip, p); ++ } ++ break; ++ } ++ read_unlock(&tasklist_lock); ++ vxdprintk(VXD_CBIT(misc, 4), ++ "vx_info_kill(%p[#%d],%d,%d,%ld) = %d", ++ vxi, vxi->vx_id, pid, sig, (long)sip, retval); ++ return retval; ++} ++ ++int vc_ctx_kill(struct vx_info *vxi, void __user *data) ++{ ++ struct vcmd_ctx_kill_v0 vc_data; ++ ++ if (copy_from_user(&vc_data, data, sizeof(vc_data))) ++ return -EFAULT; ++ ++ /* special check to allow guest shutdown */ ++ if (!vx_info_flags(vxi, VXF_STATE_ADMIN, 0) && ++ /* forbid killall pid=0 when init is present */ ++ (((vc_data.pid < 1) && vxi->vx_initpid) || ++ (vc_data.pid > 1))) ++ return -EACCES; ++ ++ return vx_info_kill(vxi, vc_data.pid, vc_data.sig); ++} ++ ++ ++static int __wait_exit(struct vx_info *vxi) ++{ ++ DECLARE_WAITQUEUE(wait, current); ++ int ret = 0; ++ ++ add_wait_queue(&vxi->vx_wait, &wait); ++ set_current_state(TASK_INTERRUPTIBLE); ++ ++wait: ++ if (vx_info_state(vxi, ++ VXS_SHUTDOWN | VXS_HASHED | VXS_HELPER) == VXS_SHUTDOWN) ++ goto out; ++ if (signal_pending(current)) { ++ ret = -ERESTARTSYS; ++ goto out; ++ } ++ schedule(); ++ goto wait; ++ ++out: ++ set_current_state(TASK_RUNNING); ++ remove_wait_queue(&vxi->vx_wait, &wait); ++ return ret; ++} ++ ++ ++ ++int vc_wait_exit(struct vx_info *vxi, void __user *data) ++{ ++ struct vcmd_wait_exit_v0 vc_data; ++ int ret; ++ ++ ret = __wait_exit(vxi); ++ vc_data.reboot_cmd = vxi->reboot_cmd; ++ vc_data.exit_code = vxi->exit_code; ++ ++ if (copy_to_user(data, &vc_data, sizeof(vc_data))) ++ ret = -EFAULT; ++ return ret; ++} ++ +diff -NurpP --minimal linux-3.0.9/kernel/vserver/space.c linux-3.0.9-vs2.3.2.1/kernel/vserver/space.c +--- linux-3.0.9/kernel/vserver/space.c 1970-01-01 01:00:00.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/kernel/vserver/space.c 2011-07-20 02:11:49.000000000 +0200 +@@ -0,0 +1,435 @@ ++/* ++ * linux/kernel/vserver/space.c ++ * ++ * Virtual Server: Context Space Support ++ * ++ * Copyright (C) 2003-2010 Herbert Pötzl ++ * ++ * V0.01 broken out from context.c 0.07 ++ * V0.02 added task locking for namespace ++ * V0.03 broken out vx_enter_namespace ++ * V0.04 added *space support and commands ++ * V0.05 added credential support ++ * ++ */ ++ ++#include <linux/utsname.h> ++#include <linux/nsproxy.h> ++#include <linux/err.h> ++#include <linux/fs_struct.h> ++#include <linux/cred.h> ++#include <asm/uaccess.h> ++ ++#include <linux/vs_context.h> ++#include <linux/vserver/space.h> ++#include <linux/vserver/space_cmd.h> ++ ++atomic_t vs_global_nsproxy = ATOMIC_INIT(0); ++atomic_t vs_global_fs = ATOMIC_INIT(0); ++atomic_t vs_global_mnt_ns = ATOMIC_INIT(0); ++atomic_t vs_global_uts_ns = ATOMIC_INIT(0); ++atomic_t vs_global_user_ns = ATOMIC_INIT(0); ++atomic_t vs_global_pid_ns = ATOMIC_INIT(0); ++ ++ ++/* namespace functions */ ++ ++#include <linux/mnt_namespace.h> ++#include <linux/user_namespace.h> ++#include <linux/pid_namespace.h> ++#include <linux/ipc_namespace.h> ++#include <net/net_namespace.h> ++ ++ ++static const struct vcmd_space_mask_v1 space_mask_v0 = { ++ .mask = CLONE_FS | ++ CLONE_NEWNS | ++#ifdef CONFIG_UTS_NS ++ CLONE_NEWUTS | ++#endif ++#ifdef CONFIG_IPC_NS ++ CLONE_NEWIPC | ++#endif ++#ifdef CONFIG_USER_NS ++ CLONE_NEWUSER | ++#endif ++ 0 ++}; ++ ++static const struct vcmd_space_mask_v1 space_mask = { ++ .mask = CLONE_FS | ++ CLONE_NEWNS | ++#ifdef CONFIG_UTS_NS ++ CLONE_NEWUTS | ++#endif ++#ifdef CONFIG_IPC_NS ++ CLONE_NEWIPC | ++#endif ++#ifdef CONFIG_USER_NS ++ CLONE_NEWUSER | ++#endif ++#ifdef CONFIG_PID_NS ++ CLONE_NEWPID | ++#endif ++#ifdef CONFIG_NET_NS ++ CLONE_NEWNET | ++#endif ++ 0 ++}; ++ ++static const struct vcmd_space_mask_v1 default_space_mask = { ++ .mask = CLONE_FS | ++ CLONE_NEWNS | ++#ifdef CONFIG_UTS_NS ++ CLONE_NEWUTS | ++#endif ++#ifdef CONFIG_IPC_NS ++ CLONE_NEWIPC | ++#endif ++#ifdef CONFIG_USER_NS ++ CLONE_NEWUSER | ++#endif ++#ifdef CONFIG_PID_NS ++// CLONE_NEWPID | ++#endif ++ 0 ++}; ++ ++/* ++ * build a new nsproxy mix ++ * assumes that both proxies are 'const' ++ * does not touch nsproxy refcounts ++ * will hold a reference on the result. ++ */ ++ ++struct nsproxy *vs_mix_nsproxy(struct nsproxy *old_nsproxy, ++ struct nsproxy *new_nsproxy, unsigned long mask) ++{ ++ struct mnt_namespace *old_ns; ++ struct uts_namespace *old_uts; ++ struct ipc_namespace *old_ipc; ++#ifdef CONFIG_PID_NS ++ struct pid_namespace *old_pid; ++#endif ++#ifdef CONFIG_NET_NS ++ struct net *old_net; ++#endif ++ struct nsproxy *nsproxy; ++ ++ nsproxy = copy_nsproxy(old_nsproxy); ++ if (!nsproxy) ++ goto out; ++ ++ if (mask & CLONE_NEWNS) { ++ old_ns = nsproxy->mnt_ns; ++ nsproxy->mnt_ns = new_nsproxy->mnt_ns; ++ if (nsproxy->mnt_ns) ++ get_mnt_ns(nsproxy->mnt_ns); ++ } else ++ old_ns = NULL; ++ ++ if (mask & CLONE_NEWUTS) { ++ old_uts = nsproxy->uts_ns; ++ nsproxy->uts_ns = new_nsproxy->uts_ns; ++ if (nsproxy->uts_ns) ++ get_uts_ns(nsproxy->uts_ns); ++ } else ++ old_uts = NULL; ++ ++ if (mask & CLONE_NEWIPC) { ++ old_ipc = nsproxy->ipc_ns; ++ nsproxy->ipc_ns = new_nsproxy->ipc_ns; ++ if (nsproxy->ipc_ns) ++ get_ipc_ns(nsproxy->ipc_ns); ++ } else ++ old_ipc = NULL; ++ ++#ifdef CONFIG_PID_NS ++ if (mask & CLONE_NEWPID) { ++ old_pid = nsproxy->pid_ns; ++ nsproxy->pid_ns = new_nsproxy->pid_ns; ++ if (nsproxy->pid_ns) ++ get_pid_ns(nsproxy->pid_ns); ++ } else ++ old_pid = NULL; ++#endif ++#ifdef CONFIG_NET_NS ++ if (mask & CLONE_NEWNET) { ++ old_net = nsproxy->net_ns; ++ nsproxy->net_ns = new_nsproxy->net_ns; ++ if (nsproxy->net_ns) ++ get_net(nsproxy->net_ns); ++ } else ++ old_net = NULL; ++#endif ++ if (old_ns) ++ put_mnt_ns(old_ns); ++ if (old_uts) ++ put_uts_ns(old_uts); ++ if (old_ipc) ++ put_ipc_ns(old_ipc); ++#ifdef CONFIG_PID_NS ++ if (old_pid) ++ put_pid_ns(old_pid); ++#endif ++#ifdef CONFIG_NET_NS ++ if (old_net) ++ put_net(old_net); ++#endif ++out: ++ return nsproxy; ++} ++ ++ ++/* ++ * merge two nsproxy structs into a new one. ++ * will hold a reference on the result. ++ */ ++ ++static inline ++struct nsproxy *__vs_merge_nsproxy(struct nsproxy *old, ++ struct nsproxy *proxy, unsigned long mask) ++{ ++ struct nsproxy null_proxy = { .mnt_ns = NULL }; ++ ++ if (!proxy) ++ return NULL; ++ ++ if (mask) { ++ /* vs_mix_nsproxy returns with reference */ ++ return vs_mix_nsproxy(old ? old : &null_proxy, ++ proxy, mask); ++ } ++ get_nsproxy(proxy); ++ return proxy; ++} ++ ++ ++int vx_enter_space(struct vx_info *vxi, unsigned long mask, unsigned index) ++{ ++ struct nsproxy *proxy, *proxy_cur, *proxy_new; ++ struct fs_struct *fs_cur, *fs = NULL; ++ struct _vx_space *space; ++ int ret, kill = 0; ++ ++ vxdprintk(VXD_CBIT(space, 8), "vx_enter_space(%p[#%u],0x%08lx,%d)", ++ vxi, vxi->vx_id, mask, index); ++ ++ if (vx_info_flags(vxi, VXF_INFO_PRIVATE, 0)) ++ return -EACCES; ++ ++ if (index >= VX_SPACES) ++ return -EINVAL; ++ ++ space = &vxi->space[index]; ++ ++ if (!mask) ++ mask = space->vx_nsmask; ++ ++ if ((mask & space->vx_nsmask) != mask) ++ return -EINVAL; ++ ++ if (mask & CLONE_FS) { ++ fs = copy_fs_struct(space->vx_fs); ++ if (!fs) ++ return -ENOMEM; ++ } ++ proxy = space->vx_nsproxy; ++ ++ vxdprintk(VXD_CBIT(space, 9), ++ "vx_enter_space(%p[#%u],0x%08lx,%d) -> (%p,%p)", ++ vxi, vxi->vx_id, mask, index, proxy, fs); ++ ++ task_lock(current); ++ fs_cur = current->fs; ++ ++ if (mask & CLONE_FS) { ++ spin_lock(&fs_cur->lock); ++ current->fs = fs; ++ kill = !--fs_cur->users; ++ spin_unlock(&fs_cur->lock); ++ } ++ ++ proxy_cur = current->nsproxy; ++ get_nsproxy(proxy_cur); ++ task_unlock(current); ++ ++ if (kill) ++ free_fs_struct(fs_cur); ++ ++ proxy_new = __vs_merge_nsproxy(proxy_cur, proxy, mask); ++ if (IS_ERR(proxy_new)) { ++ ret = PTR_ERR(proxy_new); ++ goto out_put; ++ } ++ ++ proxy_new = xchg(¤t->nsproxy, proxy_new); ++ ++ if (mask & CLONE_NEWUSER) { ++ struct cred *cred; ++ ++ vxdprintk(VXD_CBIT(space, 10), ++ "vx_enter_space(%p[#%u],%p) cred (%p,%p)", ++ vxi, vxi->vx_id, space->vx_cred, ++ current->real_cred, current->cred); ++ ++ if (space->vx_cred) { ++ cred = __prepare_creds(space->vx_cred); ++ if (cred) ++ commit_creds(cred); ++ } ++ } ++ ++ ret = 0; ++ ++ if (proxy_new) ++ put_nsproxy(proxy_new); ++out_put: ++ if (proxy_cur) ++ put_nsproxy(proxy_cur); ++ return ret; ++} ++ ++ ++int vx_set_space(struct vx_info *vxi, unsigned long mask, unsigned index) ++{ ++ struct nsproxy *proxy_vxi, *proxy_cur, *proxy_new; ++ struct fs_struct *fs_vxi, *fs; ++ struct _vx_space *space; ++ int ret, kill = 0; ++ ++ vxdprintk(VXD_CBIT(space, 8), "vx_set_space(%p[#%u],0x%08lx,%d)", ++ vxi, vxi->vx_id, mask, index); ++ ++ if ((mask & space_mask.mask) != mask) ++ return -EINVAL; ++ ++ if (index >= VX_SPACES) ++ return -EINVAL; ++ ++ space = &vxi->space[index]; ++ ++ proxy_vxi = space->vx_nsproxy; ++ fs_vxi = space->vx_fs; ++ ++ if (mask & CLONE_FS) { ++ fs = copy_fs_struct(current->fs); ++ if (!fs) ++ return -ENOMEM; ++ } ++ ++ task_lock(current); ++ ++ if (mask & CLONE_FS) { ++ spin_lock(&fs_vxi->lock); ++ space->vx_fs = fs; ++ kill = !--fs_vxi->users; ++ spin_unlock(&fs_vxi->lock); ++ } ++ ++ proxy_cur = current->nsproxy; ++ get_nsproxy(proxy_cur); ++ task_unlock(current); ++ ++ if (kill) ++ free_fs_struct(fs_vxi); ++ ++ proxy_new = __vs_merge_nsproxy(proxy_vxi, proxy_cur, mask); ++ if (IS_ERR(proxy_new)) { ++ ret = PTR_ERR(proxy_new); ++ goto out_put; ++ } ++ ++ proxy_new = xchg(&space->vx_nsproxy, proxy_new); ++ space->vx_nsmask |= mask; ++ ++ if (mask & CLONE_NEWUSER) { ++ struct cred *cred; ++ ++ vxdprintk(VXD_CBIT(space, 10), ++ "vx_set_space(%p[#%u],%p) cred (%p,%p)", ++ vxi, vxi->vx_id, space->vx_cred, ++ current->real_cred, current->cred); ++ ++ cred = prepare_creds(); ++ cred = (struct cred *)xchg(&space->vx_cred, cred); ++ if (cred) ++ abort_creds(cred); ++ } ++ ++ ret = 0; ++ ++ if (proxy_new) ++ put_nsproxy(proxy_new); ++out_put: ++ if (proxy_cur) ++ put_nsproxy(proxy_cur); ++ return ret; ++} ++ ++ ++int vc_enter_space_v1(struct vx_info *vxi, void __user *data) ++{ ++ struct vcmd_space_mask_v1 vc_data = { .mask = 0 }; ++ ++ if (data && copy_from_user(&vc_data, data, sizeof(vc_data))) ++ return -EFAULT; ++ ++ return vx_enter_space(vxi, vc_data.mask, 0); ++} ++ ++int vc_enter_space(struct vx_info *vxi, void __user *data) ++{ ++ struct vcmd_space_mask_v2 vc_data = { .mask = 0 }; ++ ++ if (data && copy_from_user(&vc_data, data, sizeof(vc_data))) ++ return -EFAULT; ++ ++ if (vc_data.index >= VX_SPACES) ++ return -EINVAL; ++ ++ return vx_enter_space(vxi, vc_data.mask, vc_data.index); ++} ++ ++int vc_set_space_v1(struct vx_info *vxi, void __user *data) ++{ ++ struct vcmd_space_mask_v1 vc_data = { .mask = 0 }; ++ ++ if (data && copy_from_user(&vc_data, data, sizeof(vc_data))) ++ return -EFAULT; ++ ++ return vx_set_space(vxi, vc_data.mask, 0); ++} ++ ++int vc_set_space(struct vx_info *vxi, void __user *data) ++{ ++ struct vcmd_space_mask_v2 vc_data = { .mask = 0 }; ++ ++ if (data && copy_from_user(&vc_data, data, sizeof(vc_data))) ++ return -EFAULT; ++ ++ if (vc_data.index >= VX_SPACES) ++ return -EINVAL; ++ ++ return vx_set_space(vxi, vc_data.mask, vc_data.index); ++} ++ ++int vc_get_space_mask(void __user *data, int type) ++{ ++ const struct vcmd_space_mask_v1 *mask; ++ ++ if (type == 0) ++ mask = &space_mask_v0; ++ else if (type == 1) ++ mask = &space_mask; ++ else ++ mask = &default_space_mask; ++ ++ vxdprintk(VXD_CBIT(space, 10), ++ "vc_get_space_mask(%d) = %08llx", type, mask->mask); ++ ++ if (copy_to_user(data, mask, sizeof(*mask))) ++ return -EFAULT; ++ return 0; ++} ++ +diff -NurpP --minimal linux-3.0.9/kernel/vserver/switch.c linux-3.0.9-vs2.3.2.1/kernel/vserver/switch.c +--- linux-3.0.9/kernel/vserver/switch.c 1970-01-01 01:00:00.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/kernel/vserver/switch.c 2011-08-01 18:28:26.000000000 +0200 +@@ -0,0 +1,556 @@ ++/* ++ * linux/kernel/vserver/switch.c ++ * ++ * Virtual Server: Syscall Switch ++ * ++ * Copyright (C) 2003-2011 Herbert Pötzl ++ * ++ * V0.01 syscall switch ++ * V0.02 added signal to context ++ * V0.03 added rlimit functions ++ * V0.04 added iattr, task/xid functions ++ * V0.05 added debug/history stuff ++ * V0.06 added compat32 layer ++ * V0.07 vcmd args and perms ++ * V0.08 added status commands ++ * V0.09 added tag commands ++ * V0.10 added oom bias ++ * V0.11 added device commands ++ * V0.12 added warn mask ++ * ++ */ ++ ++#include <linux/vs_context.h> ++#include <linux/vs_network.h> ++#include <linux/vserver/switch.h> ++ ++#include "vci_config.h" ++ ++ ++static inline ++int vc_get_version(uint32_t id) ++{ ++ return VCI_VERSION; ++} ++ ++static inline ++int vc_get_vci(uint32_t id) ++{ ++ return vci_kernel_config(); ++} ++ ++#include <linux/vserver/context_cmd.h> ++#include <linux/vserver/cvirt_cmd.h> ++#include <linux/vserver/cacct_cmd.h> ++#include <linux/vserver/limit_cmd.h> ++#include <linux/vserver/network_cmd.h> ++#include <linux/vserver/sched_cmd.h> ++#include <linux/vserver/debug_cmd.h> ++#include <linux/vserver/inode_cmd.h> ++#include <linux/vserver/dlimit_cmd.h> ++#include <linux/vserver/signal_cmd.h> ++#include <linux/vserver/space_cmd.h> ++#include <linux/vserver/tag_cmd.h> ++#include <linux/vserver/device_cmd.h> ++ ++#include <linux/vserver/inode.h> ++#include <linux/vserver/dlimit.h> ++ ++ ++#ifdef CONFIG_COMPAT ++#define __COMPAT(name, id, data, compat) \ ++ (compat) ? name ## _x32(id, data) : name(id, data) ++#define __COMPAT_NO_ID(name, data, compat) \ ++ (compat) ? name ## _x32(data) : name(data) ++#else ++#define __COMPAT(name, id, data, compat) \ ++ name(id, data) ++#define __COMPAT_NO_ID(name, data, compat) \ ++ name(data) ++#endif ++ ++ ++static inline ++long do_vcmd(uint32_t cmd, uint32_t id, ++ struct vx_info *vxi, struct nx_info *nxi, ++ void __user *data, int compat) ++{ ++ switch (cmd) { ++ ++ case VCMD_get_version: ++ return vc_get_version(id); ++ case VCMD_get_vci: ++ return vc_get_vci(id); ++ ++ case VCMD_task_xid: ++ return vc_task_xid(id); ++ case VCMD_vx_info: ++ return vc_vx_info(vxi, data); ++ ++ case VCMD_task_nid: ++ return vc_task_nid(id); ++ case VCMD_nx_info: ++ return vc_nx_info(nxi, data); ++ ++ case VCMD_task_tag: ++ return vc_task_tag(id); ++ ++ case VCMD_set_space_v1: ++ return vc_set_space_v1(vxi, data); ++ /* this is version 2 */ ++ case VCMD_set_space: ++ return vc_set_space(vxi, data); ++ ++ case VCMD_get_space_mask_v0: ++ return vc_get_space_mask(data, 0); ++ /* this is version 1 */ ++ case VCMD_get_space_mask: ++ return vc_get_space_mask(data, 1); ++ ++ case VCMD_get_space_default: ++ return vc_get_space_mask(data, -1); ++ ++ case VCMD_set_umask: ++ return vc_set_umask(vxi, data); ++ ++ case VCMD_get_umask: ++ return vc_get_umask(vxi, data); ++ ++ case VCMD_set_wmask: ++ return vc_set_wmask(vxi, data); ++ ++ case VCMD_get_wmask: ++ return vc_get_wmask(vxi, data); ++#ifdef CONFIG_IA32_EMULATION ++ case VCMD_get_rlimit: ++ return __COMPAT(vc_get_rlimit, vxi, data, compat); ++ case VCMD_set_rlimit: ++ return __COMPAT(vc_set_rlimit, vxi, data, compat); ++#else ++ case VCMD_get_rlimit: ++ return vc_get_rlimit(vxi, data); ++ case VCMD_set_rlimit: ++ return vc_set_rlimit(vxi, data); ++#endif ++ case VCMD_get_rlimit_mask: ++ return vc_get_rlimit_mask(id, data); ++ case VCMD_reset_hits: ++ return vc_reset_hits(vxi, data); ++ case VCMD_reset_minmax: ++ return vc_reset_minmax(vxi, data); ++ ++ case VCMD_get_vhi_name: ++ return vc_get_vhi_name(vxi, data); ++ case VCMD_set_vhi_name: ++ return vc_set_vhi_name(vxi, data); ++ ++ case VCMD_ctx_stat: ++ return vc_ctx_stat(vxi, data); ++ case VCMD_virt_stat: ++ return vc_virt_stat(vxi, data); ++ case VCMD_sock_stat: ++ return vc_sock_stat(vxi, data); ++ case VCMD_rlimit_stat: ++ return vc_rlimit_stat(vxi, data); ++ ++ case VCMD_set_cflags: ++ return vc_set_cflags(vxi, data); ++ case VCMD_get_cflags: ++ return vc_get_cflags(vxi, data); ++ ++ /* this is version 1 */ ++ case VCMD_set_ccaps: ++ return vc_set_ccaps(vxi, data); ++ /* this is version 1 */ ++ case VCMD_get_ccaps: ++ return vc_get_ccaps(vxi, data); ++ case VCMD_set_bcaps: ++ return vc_set_bcaps(vxi, data); ++ case VCMD_get_bcaps: ++ return vc_get_bcaps(vxi, data); ++ ++ case VCMD_set_badness: ++ return vc_set_badness(vxi, data); ++ case VCMD_get_badness: ++ return vc_get_badness(vxi, data); ++ ++ case VCMD_set_nflags: ++ return vc_set_nflags(nxi, data); ++ case VCMD_get_nflags: ++ return vc_get_nflags(nxi, data); ++ ++ case VCMD_set_ncaps: ++ return vc_set_ncaps(nxi, data); ++ case VCMD_get_ncaps: ++ return vc_get_ncaps(nxi, data); ++ ++ case VCMD_set_prio_bias: ++ return vc_set_prio_bias(vxi, data); ++ case VCMD_get_prio_bias: ++ return vc_get_prio_bias(vxi, data); ++ case VCMD_add_dlimit: ++ return __COMPAT(vc_add_dlimit, id, data, compat); ++ case VCMD_rem_dlimit: ++ return __COMPAT(vc_rem_dlimit, id, data, compat); ++ case VCMD_set_dlimit: ++ return __COMPAT(vc_set_dlimit, id, data, compat); ++ case VCMD_get_dlimit: ++ return __COMPAT(vc_get_dlimit, id, data, compat); ++ ++ case VCMD_ctx_kill: ++ return vc_ctx_kill(vxi, data); ++ ++ case VCMD_wait_exit: ++ return vc_wait_exit(vxi, data); ++ ++ case VCMD_get_iattr: ++ return __COMPAT_NO_ID(vc_get_iattr, data, compat); ++ case VCMD_set_iattr: ++ return __COMPAT_NO_ID(vc_set_iattr, data, compat); ++ ++ case VCMD_fget_iattr: ++ return vc_fget_iattr(id, data); ++ case VCMD_fset_iattr: ++ return vc_fset_iattr(id, data); ++ ++ case VCMD_enter_space_v0: ++ return vc_enter_space_v1(vxi, NULL); ++ case VCMD_enter_space_v1: ++ return vc_enter_space_v1(vxi, data); ++ /* this is version 2 */ ++ case VCMD_enter_space: ++ return vc_enter_space(vxi, data); ++ ++ case VCMD_ctx_create_v0: ++ return vc_ctx_create(id, NULL); ++ case VCMD_ctx_create: ++ return vc_ctx_create(id, data); ++ case VCMD_ctx_migrate_v0: ++ return vc_ctx_migrate(vxi, NULL); ++ case VCMD_ctx_migrate: ++ return vc_ctx_migrate(vxi, data); ++ ++ case VCMD_net_create_v0: ++ return vc_net_create(id, NULL); ++ case VCMD_net_create: ++ return vc_net_create(id, data); ++ case VCMD_net_migrate: ++ return vc_net_migrate(nxi, data); ++ ++ case VCMD_tag_migrate: ++ return vc_tag_migrate(id); ++ ++ case VCMD_net_add: ++ return vc_net_add(nxi, data); ++ case VCMD_net_remove: ++ return vc_net_remove(nxi, data); ++ ++ case VCMD_net_add_ipv4_v1: ++ return vc_net_add_ipv4_v1(nxi, data); ++ /* this is version 2 */ ++ case VCMD_net_add_ipv4: ++ return vc_net_add_ipv4(nxi, data); ++ ++ case VCMD_net_rem_ipv4_v1: ++ return vc_net_rem_ipv4_v1(nxi, data); ++ /* this is version 2 */ ++ case VCMD_net_rem_ipv4: ++ return vc_net_rem_ipv4(nxi, data); ++#ifdef CONFIG_IPV6 ++ case VCMD_net_add_ipv6: ++ return vc_net_add_ipv6(nxi, data); ++ case VCMD_net_remove_ipv6: ++ return vc_net_remove_ipv6(nxi, data); ++#endif ++/* case VCMD_add_match_ipv4: ++ return vc_add_match_ipv4(nxi, data); ++ case VCMD_get_match_ipv4: ++ return vc_get_match_ipv4(nxi, data); ++#ifdef CONFIG_IPV6 ++ case VCMD_add_match_ipv6: ++ return vc_add_match_ipv6(nxi, data); ++ case VCMD_get_match_ipv6: ++ return vc_get_match_ipv6(nxi, data); ++#endif */ ++ ++#ifdef CONFIG_VSERVER_DEVICE ++ case VCMD_set_mapping: ++ return __COMPAT(vc_set_mapping, vxi, data, compat); ++ case VCMD_unset_mapping: ++ return __COMPAT(vc_unset_mapping, vxi, data, compat); ++#endif ++#ifdef CONFIG_VSERVER_HISTORY ++ case VCMD_dump_history: ++ return vc_dump_history(id); ++ case VCMD_read_history: ++ return __COMPAT(vc_read_history, id, data, compat); ++#endif ++ default: ++ vxwprintk_task(1, "unimplemented VCMD_%02d_%d[%d]", ++ VC_CATEGORY(cmd), VC_COMMAND(cmd), VC_VERSION(cmd)); ++ } ++ return -ENOSYS; ++} ++ ++ ++#define __VCMD(vcmd, _perm, _args, _flags) \ ++ case VCMD_ ## vcmd: perm = _perm; \ ++ args = _args; flags = _flags; break ++ ++ ++#define VCA_NONE 0x00 ++#define VCA_VXI 0x01 ++#define VCA_NXI 0x02 ++ ++#define VCF_NONE 0x00 ++#define VCF_INFO 0x01 ++#define VCF_ADMIN 0x02 ++#define VCF_ARES 0x06 /* includes admin */ ++#define VCF_SETUP 0x08 ++ ++#define VCF_ZIDOK 0x10 /* zero id okay */ ++ ++ ++static inline ++long do_vserver(uint32_t cmd, uint32_t id, void __user *data, int compat) ++{ ++ long ret; ++ int permit = -1, state = 0; ++ int perm = -1, args = 0, flags = 0; ++ struct vx_info *vxi = NULL; ++ struct nx_info *nxi = NULL; ++ ++ switch (cmd) { ++ /* unpriviledged commands */ ++ __VCMD(get_version, 0, VCA_NONE, 0); ++ __VCMD(get_vci, 0, VCA_NONE, 0); ++ __VCMD(get_rlimit_mask, 0, VCA_NONE, 0); ++ __VCMD(get_space_mask_v0,0, VCA_NONE, 0); ++ __VCMD(get_space_mask, 0, VCA_NONE, 0); ++ __VCMD(get_space_default,0, VCA_NONE, 0); ++ ++ /* info commands */ ++ __VCMD(task_xid, 2, VCA_NONE, 0); ++ __VCMD(reset_hits, 2, VCA_VXI, 0); ++ __VCMD(reset_minmax, 2, VCA_VXI, 0); ++ __VCMD(vx_info, 3, VCA_VXI, VCF_INFO); ++ __VCMD(get_bcaps, 3, VCA_VXI, VCF_INFO); ++ __VCMD(get_ccaps, 3, VCA_VXI, VCF_INFO); ++ __VCMD(get_cflags, 3, VCA_VXI, VCF_INFO); ++ __VCMD(get_umask, 3, VCA_VXI, VCF_INFO); ++ __VCMD(get_wmask, 3, VCA_VXI, VCF_INFO); ++ __VCMD(get_badness, 3, VCA_VXI, VCF_INFO); ++ __VCMD(get_vhi_name, 3, VCA_VXI, VCF_INFO); ++ __VCMD(get_rlimit, 3, VCA_VXI, VCF_INFO); ++ ++ __VCMD(ctx_stat, 3, VCA_VXI, VCF_INFO); ++ __VCMD(virt_stat, 3, VCA_VXI, VCF_INFO); ++ __VCMD(sock_stat, 3, VCA_VXI, VCF_INFO); ++ __VCMD(rlimit_stat, 3, VCA_VXI, VCF_INFO); ++ ++ __VCMD(task_nid, 2, VCA_NONE, 0); ++ __VCMD(nx_info, 3, VCA_NXI, VCF_INFO); ++ __VCMD(get_ncaps, 3, VCA_NXI, VCF_INFO); ++ __VCMD(get_nflags, 3, VCA_NXI, VCF_INFO); ++ ++ __VCMD(task_tag, 2, VCA_NONE, 0); ++ ++ __VCMD(get_iattr, 2, VCA_NONE, 0); ++ __VCMD(fget_iattr, 2, VCA_NONE, 0); ++ __VCMD(get_dlimit, 3, VCA_NONE, VCF_INFO); ++ __VCMD(get_prio_bias, 3, VCA_VXI, VCF_INFO); ++ ++ /* lower admin commands */ ++ __VCMD(wait_exit, 4, VCA_VXI, VCF_INFO); ++ __VCMD(ctx_create_v0, 5, VCA_NONE, 0); ++ __VCMD(ctx_create, 5, VCA_NONE, 0); ++ __VCMD(ctx_migrate_v0, 5, VCA_VXI, VCF_ADMIN); ++ __VCMD(ctx_migrate, 5, VCA_VXI, VCF_ADMIN); ++ __VCMD(enter_space_v0, 5, VCA_VXI, VCF_ADMIN); ++ __VCMD(enter_space_v1, 5, VCA_VXI, VCF_ADMIN); ++ __VCMD(enter_space, 5, VCA_VXI, VCF_ADMIN); ++ ++ __VCMD(net_create_v0, 5, VCA_NONE, 0); ++ __VCMD(net_create, 5, VCA_NONE, 0); ++ __VCMD(net_migrate, 5, VCA_NXI, VCF_ADMIN); ++ ++ __VCMD(tag_migrate, 5, VCA_NONE, VCF_ADMIN); ++ ++ /* higher admin commands */ ++ __VCMD(ctx_kill, 6, VCA_VXI, VCF_ARES); ++ __VCMD(set_space_v1, 7, VCA_VXI, VCF_ARES | VCF_SETUP); ++ __VCMD(set_space, 7, VCA_VXI, VCF_ARES | VCF_SETUP); ++ ++ __VCMD(set_ccaps, 7, VCA_VXI, VCF_ARES | VCF_SETUP); ++ __VCMD(set_bcaps, 7, VCA_VXI, VCF_ARES | VCF_SETUP); ++ __VCMD(set_cflags, 7, VCA_VXI, VCF_ARES | VCF_SETUP); ++ __VCMD(set_umask, 7, VCA_VXI, VCF_ARES | VCF_SETUP); ++ __VCMD(set_wmask, 7, VCA_VXI, VCF_ARES | VCF_SETUP); ++ __VCMD(set_badness, 7, VCA_VXI, VCF_ARES | VCF_SETUP); ++ ++ __VCMD(set_vhi_name, 7, VCA_VXI, VCF_ARES | VCF_SETUP); ++ __VCMD(set_rlimit, 7, VCA_VXI, VCF_ARES | VCF_SETUP); ++ __VCMD(set_prio_bias, 7, VCA_VXI, VCF_ARES | VCF_SETUP); ++ ++ __VCMD(set_ncaps, 7, VCA_NXI, VCF_ARES | VCF_SETUP); ++ __VCMD(set_nflags, 7, VCA_NXI, VCF_ARES | VCF_SETUP); ++ __VCMD(net_add, 8, VCA_NXI, VCF_ARES | VCF_SETUP); ++ __VCMD(net_remove, 8, VCA_NXI, VCF_ARES | VCF_SETUP); ++ __VCMD(net_add_ipv4_v1, 8, VCA_NXI, VCF_ARES | VCF_SETUP); ++ __VCMD(net_rem_ipv4_v1, 8, VCA_NXI, VCF_ARES | VCF_SETUP); ++ __VCMD(net_add_ipv4, 8, VCA_NXI, VCF_ARES | VCF_SETUP); ++ __VCMD(net_rem_ipv4, 8, VCA_NXI, VCF_ARES | VCF_SETUP); ++#ifdef CONFIG_IPV6 ++ __VCMD(net_add_ipv6, 8, VCA_NXI, VCF_ARES | VCF_SETUP); ++ __VCMD(net_remove_ipv6, 8, VCA_NXI, VCF_ARES | VCF_SETUP); ++#endif ++ __VCMD(set_iattr, 7, VCA_NONE, 0); ++ __VCMD(fset_iattr, 7, VCA_NONE, 0); ++ __VCMD(set_dlimit, 7, VCA_NONE, VCF_ARES); ++ __VCMD(add_dlimit, 8, VCA_NONE, VCF_ARES); ++ __VCMD(rem_dlimit, 8, VCA_NONE, VCF_ARES); ++ ++#ifdef CONFIG_VSERVER_DEVICE ++ __VCMD(set_mapping, 8, VCA_VXI, VCF_ARES|VCF_ZIDOK); ++ __VCMD(unset_mapping, 8, VCA_VXI, VCF_ARES|VCF_ZIDOK); ++#endif ++ /* debug level admin commands */ ++#ifdef CONFIG_VSERVER_HISTORY ++ __VCMD(dump_history, 9, VCA_NONE, 0); ++ __VCMD(read_history, 9, VCA_NONE, 0); ++#endif ++ ++ default: ++ perm = -1; ++ } ++ ++ vxdprintk(VXD_CBIT(switch, 0), ++ "vc: VCMD_%02d_%d[%d], %d,%p [%d,%d,%x,%x]", ++ VC_CATEGORY(cmd), VC_COMMAND(cmd), ++ VC_VERSION(cmd), id, data, compat, ++ perm, args, flags); ++ ++ ret = -ENOSYS; ++ if (perm < 0) ++ goto out; ++ ++ state = 1; ++ if (!capable(CAP_CONTEXT)) ++ goto out; ++ ++ state = 2; ++ /* moved here from the individual commands */ ++ ret = -EPERM; ++ if ((perm > 1) && !capable(CAP_SYS_ADMIN)) ++ goto out; ++ ++ state = 3; ++ /* vcmd involves resource management */ ++ ret = -EPERM; ++ if ((flags & VCF_ARES) && !capable(CAP_SYS_RESOURCE)) ++ goto out; ++ ++ state = 4; ++ /* various legacy exceptions */ ++ switch (cmd) { ++ /* will go away when spectator is a cap */ ++ case VCMD_ctx_migrate_v0: ++ case VCMD_ctx_migrate: ++ if (id == 1) { ++ current->xid = 1; ++ ret = 1; ++ goto out; ++ } ++ break; ++ ++ /* will go away when spectator is a cap */ ++ case VCMD_net_migrate: ++ if (id == 1) { ++ current->nid = 1; ++ ret = 1; ++ goto out; ++ } ++ break; ++ } ++ ++ /* vcmds are fine by default */ ++ permit = 1; ++ ++ /* admin type vcmds require admin ... */ ++ if (flags & VCF_ADMIN) ++ permit = vx_check(0, VS_ADMIN) ? 1 : 0; ++ ++ /* ... but setup type vcmds override that */ ++ if (!permit && (flags & VCF_SETUP)) ++ permit = vx_flags(VXF_STATE_SETUP, 0) ? 2 : 0; ++ ++ state = 5; ++ ret = -EPERM; ++ if (!permit) ++ goto out; ++ ++ state = 6; ++ if (!id && (flags & VCF_ZIDOK)) ++ goto skip_id; ++ ++ ret = -ESRCH; ++ if (args & VCA_VXI) { ++ vxi = lookup_vx_info(id); ++ if (!vxi) ++ goto out; ++ ++ if ((flags & VCF_ADMIN) && ++ /* special case kill for shutdown */ ++ (cmd != VCMD_ctx_kill) && ++ /* can context be administrated? */ ++ !vx_info_flags(vxi, VXF_STATE_ADMIN, 0)) { ++ ret = -EACCES; ++ goto out_vxi; ++ } ++ } ++ state = 7; ++ if (args & VCA_NXI) { ++ nxi = lookup_nx_info(id); ++ if (!nxi) ++ goto out_vxi; ++ ++ if ((flags & VCF_ADMIN) && ++ /* can context be administrated? */ ++ !nx_info_flags(nxi, NXF_STATE_ADMIN, 0)) { ++ ret = -EACCES; ++ goto out_nxi; ++ } ++ } ++skip_id: ++ state = 8; ++ ret = do_vcmd(cmd, id, vxi, nxi, data, compat); ++ ++out_nxi: ++ if ((args & VCA_NXI) && nxi) ++ put_nx_info(nxi); ++out_vxi: ++ if ((args & VCA_VXI) && vxi) ++ put_vx_info(vxi); ++out: ++ vxdprintk(VXD_CBIT(switch, 1), ++ "vc: VCMD_%02d_%d[%d] = %08lx(%ld) [%d,%d]", ++ VC_CATEGORY(cmd), VC_COMMAND(cmd), ++ VC_VERSION(cmd), ret, ret, state, permit); ++ return ret; ++} ++ ++asmlinkage long ++sys_vserver(uint32_t cmd, uint32_t id, void __user *data) ++{ ++ return do_vserver(cmd, id, data, 0); ++} ++ ++#ifdef CONFIG_COMPAT ++ ++asmlinkage long ++sys32_vserver(uint32_t cmd, uint32_t id, void __user *data) ++{ ++ return do_vserver(cmd, id, data, 1); ++} ++ ++#endif /* CONFIG_COMPAT */ +diff -NurpP --minimal linux-3.0.9/kernel/vserver/sysctl.c linux-3.0.9-vs2.3.2.1/kernel/vserver/sysctl.c +--- linux-3.0.9/kernel/vserver/sysctl.c 1970-01-01 01:00:00.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/kernel/vserver/sysctl.c 2011-10-27 15:18:40.000000000 +0200 +@@ -0,0 +1,247 @@ ++/* ++ * kernel/vserver/sysctl.c ++ * ++ * Virtual Context Support ++ * ++ * Copyright (C) 2004-2007 Herbert Pötzl ++ * ++ * V0.01 basic structure ++ * ++ */ ++ ++#include <linux/module.h> ++#include <linux/ctype.h> ++#include <linux/sysctl.h> ++#include <linux/parser.h> ++#include <asm/uaccess.h> ++ ++enum { ++ CTL_DEBUG_ERROR = 0, ++ CTL_DEBUG_SWITCH = 1, ++ CTL_DEBUG_XID, ++ CTL_DEBUG_NID, ++ CTL_DEBUG_TAG, ++ CTL_DEBUG_NET, ++ CTL_DEBUG_LIMIT, ++ CTL_DEBUG_CRES, ++ CTL_DEBUG_DLIM, ++ CTL_DEBUG_QUOTA, ++ CTL_DEBUG_CVIRT, ++ CTL_DEBUG_SPACE, ++ CTL_DEBUG_PERM, ++ CTL_DEBUG_MISC, ++}; ++ ++ ++unsigned int vs_debug_switch = 0; ++unsigned int vs_debug_xid = 0; ++unsigned int vs_debug_nid = 0; ++unsigned int vs_debug_tag = 0; ++unsigned int vs_debug_net = 0; ++unsigned int vs_debug_limit = 0; ++unsigned int vs_debug_cres = 0; ++unsigned int vs_debug_dlim = 0; ++unsigned int vs_debug_quota = 0; ++unsigned int vs_debug_cvirt = 0; ++unsigned int vs_debug_space = 0; ++unsigned int vs_debug_perm = 0; ++unsigned int vs_debug_misc = 0; ++ ++ ++static struct ctl_table_header *vserver_table_header; ++static ctl_table vserver_root_table[]; ++ ++ ++void vserver_register_sysctl(void) ++{ ++ if (!vserver_table_header) { ++ vserver_table_header = register_sysctl_table(vserver_root_table); ++ } ++ ++} ++ ++void vserver_unregister_sysctl(void) ++{ ++ if (vserver_table_header) { ++ unregister_sysctl_table(vserver_table_header); ++ vserver_table_header = NULL; ++ } ++} ++ ++ ++static int proc_dodebug(ctl_table *table, int write, ++ void __user *buffer, size_t *lenp, loff_t *ppos) ++{ ++ char tmpbuf[20], *p, c; ++ unsigned int value; ++ size_t left, len; ++ ++ if ((*ppos && !write) || !*lenp) { ++ *lenp = 0; ++ return 0; ++ } ++ ++ left = *lenp; ++ ++ if (write) { ++ if (!access_ok(VERIFY_READ, buffer, left)) ++ return -EFAULT; ++ p = (char *)buffer; ++ while (left && __get_user(c, p) >= 0 && isspace(c)) ++ left--, p++; ++ if (!left) ++ goto done; ++ ++ if (left > sizeof(tmpbuf) - 1) ++ return -EINVAL; ++ if (copy_from_user(tmpbuf, p, left)) ++ return -EFAULT; ++ tmpbuf[left] = '\0'; ++ ++ for (p = tmpbuf, value = 0; '0' <= *p && *p <= '9'; p++, left--) ++ value = 10 * value + (*p - '0'); ++ if (*p && !isspace(*p)) ++ return -EINVAL; ++ while (left && isspace(*p)) ++ left--, p++; ++ *(unsigned int *)table->data = value; ++ } else { ++ if (!access_ok(VERIFY_WRITE, buffer, left)) ++ return -EFAULT; ++ len = sprintf(tmpbuf, "%d", *(unsigned int *)table->data); ++ if (len > left) ++ len = left; ++ if (__copy_to_user(buffer, tmpbuf, len)) ++ return -EFAULT; ++ if ((left -= len) > 0) { ++ if (put_user('\n', (char *)buffer + len)) ++ return -EFAULT; ++ left--; ++ } ++ } ++ ++done: ++ *lenp -= left; ++ *ppos += *lenp; ++ return 0; ++} ++ ++static int zero; ++ ++#define CTL_ENTRY(ctl, name) \ ++ { \ ++ .procname = #name, \ ++ .data = &vs_ ## name, \ ++ .maxlen = sizeof(int), \ ++ .mode = 0644, \ ++ .proc_handler = &proc_dodebug, \ ++ .extra1 = &zero, \ ++ .extra2 = &zero, \ ++ } ++ ++static ctl_table vserver_debug_table[] = { ++ CTL_ENTRY(CTL_DEBUG_SWITCH, debug_switch), ++ CTL_ENTRY(CTL_DEBUG_XID, debug_xid), ++ CTL_ENTRY(CTL_DEBUG_NID, debug_nid), ++ CTL_ENTRY(CTL_DEBUG_TAG, debug_tag), ++ CTL_ENTRY(CTL_DEBUG_NET, debug_net), ++ CTL_ENTRY(CTL_DEBUG_LIMIT, debug_limit), ++ CTL_ENTRY(CTL_DEBUG_CRES, debug_cres), ++ CTL_ENTRY(CTL_DEBUG_DLIM, debug_dlim), ++ CTL_ENTRY(CTL_DEBUG_QUOTA, debug_quota), ++ CTL_ENTRY(CTL_DEBUG_CVIRT, debug_cvirt), ++ CTL_ENTRY(CTL_DEBUG_SPACE, debug_space), ++ CTL_ENTRY(CTL_DEBUG_PERM, debug_perm), ++ CTL_ENTRY(CTL_DEBUG_MISC, debug_misc), ++ { 0 } ++}; ++ ++static ctl_table vserver_root_table[] = { ++ { ++ .procname = "vserver", ++ .mode = 0555, ++ .child = vserver_debug_table ++ }, ++ { 0 } ++}; ++ ++ ++static match_table_t tokens = { ++ { CTL_DEBUG_SWITCH, "switch=%x" }, ++ { CTL_DEBUG_XID, "xid=%x" }, ++ { CTL_DEBUG_NID, "nid=%x" }, ++ { CTL_DEBUG_TAG, "tag=%x" }, ++ { CTL_DEBUG_NET, "net=%x" }, ++ { CTL_DEBUG_LIMIT, "limit=%x" }, ++ { CTL_DEBUG_CRES, "cres=%x" }, ++ { CTL_DEBUG_DLIM, "dlim=%x" }, ++ { CTL_DEBUG_QUOTA, "quota=%x" }, ++ { CTL_DEBUG_CVIRT, "cvirt=%x" }, ++ { CTL_DEBUG_SPACE, "space=%x" }, ++ { CTL_DEBUG_PERM, "perm=%x" }, ++ { CTL_DEBUG_MISC, "misc=%x" }, ++ { CTL_DEBUG_ERROR, NULL } ++}; ++ ++#define HANDLE_CASE(id, name, val) \ ++ case CTL_DEBUG_ ## id: \ ++ vs_debug_ ## name = val; \ ++ printk("vs_debug_" #name "=0x%x\n", val); \ ++ break ++ ++ ++static int __init vs_debug_setup(char *str) ++{ ++ char *p; ++ int token; ++ ++ printk("vs_debug_setup(%s)\n", str); ++ while ((p = strsep(&str, ",")) != NULL) { ++ substring_t args[MAX_OPT_ARGS]; ++ unsigned int value; ++ ++ if (!*p) ++ continue; ++ ++ token = match_token(p, tokens, args); ++ value = (token > 0) ? simple_strtoul(args[0].from, NULL, 0) : 0; ++ ++ switch (token) { ++ HANDLE_CASE(SWITCH, switch, value); ++ HANDLE_CASE(XID, xid, value); ++ HANDLE_CASE(NID, nid, value); ++ HANDLE_CASE(TAG, tag, value); ++ HANDLE_CASE(NET, net, value); ++ HANDLE_CASE(LIMIT, limit, value); ++ HANDLE_CASE(CRES, cres, value); ++ HANDLE_CASE(DLIM, dlim, value); ++ HANDLE_CASE(QUOTA, quota, value); ++ HANDLE_CASE(CVIRT, cvirt, value); ++ HANDLE_CASE(SPACE, space, value); ++ HANDLE_CASE(PERM, perm, value); ++ HANDLE_CASE(MISC, misc, value); ++ default: ++ return -EINVAL; ++ break; ++ } ++ } ++ return 1; ++} ++ ++__setup("vsdebug=", vs_debug_setup); ++ ++ ++ ++EXPORT_SYMBOL_GPL(vs_debug_switch); ++EXPORT_SYMBOL_GPL(vs_debug_xid); ++EXPORT_SYMBOL_GPL(vs_debug_nid); ++EXPORT_SYMBOL_GPL(vs_debug_net); ++EXPORT_SYMBOL_GPL(vs_debug_limit); ++EXPORT_SYMBOL_GPL(vs_debug_cres); ++EXPORT_SYMBOL_GPL(vs_debug_dlim); ++EXPORT_SYMBOL_GPL(vs_debug_quota); ++EXPORT_SYMBOL_GPL(vs_debug_cvirt); ++EXPORT_SYMBOL_GPL(vs_debug_space); ++EXPORT_SYMBOL_GPL(vs_debug_perm); ++EXPORT_SYMBOL_GPL(vs_debug_misc); ++ +diff -NurpP --minimal linux-3.0.9/kernel/vserver/tag.c linux-3.0.9-vs2.3.2.1/kernel/vserver/tag.c +--- linux-3.0.9/kernel/vserver/tag.c 1970-01-01 01:00:00.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/kernel/vserver/tag.c 2011-06-10 22:11:24.000000000 +0200 +@@ -0,0 +1,63 @@ ++/* ++ * linux/kernel/vserver/tag.c ++ * ++ * Virtual Server: Shallow Tag Space ++ * ++ * Copyright (C) 2007 Herbert Pötzl ++ * ++ * V0.01 basic implementation ++ * ++ */ ++ ++#include <linux/sched.h> ++#include <linux/vserver/debug.h> ++#include <linux/vs_pid.h> ++#include <linux/vs_tag.h> ++ ++#include <linux/vserver/tag_cmd.h> ++ ++ ++int dx_migrate_task(struct task_struct *p, tag_t tag) ++{ ++ if (!p) ++ BUG(); ++ ++ vxdprintk(VXD_CBIT(tag, 5), ++ "dx_migrate_task(%p[#%d],#%d)", p, p->tag, tag); ++ ++ task_lock(p); ++ p->tag = tag; ++ task_unlock(p); ++ ++ vxdprintk(VXD_CBIT(tag, 5), ++ "moved task %p into [#%d]", p, tag); ++ return 0; ++} ++ ++/* vserver syscall commands below here */ ++ ++/* taks xid and vx_info functions */ ++ ++ ++int vc_task_tag(uint32_t id) ++{ ++ tag_t tag; ++ ++ if (id) { ++ struct task_struct *tsk; ++ rcu_read_lock(); ++ tsk = find_task_by_real_pid(id); ++ tag = (tsk) ? tsk->tag : -ESRCH; ++ rcu_read_unlock(); ++ } else ++ tag = dx_current_tag(); ++ return tag; ++} ++ ++ ++int vc_tag_migrate(uint32_t tag) ++{ ++ return dx_migrate_task(current, tag & 0xFFFF); ++} ++ ++ +diff -NurpP --minimal linux-3.0.9/kernel/vserver/vci_config.h linux-3.0.9-vs2.3.2.1/kernel/vserver/vci_config.h +--- linux-3.0.9/kernel/vserver/vci_config.h 1970-01-01 01:00:00.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/kernel/vserver/vci_config.h 2011-06-10 22:11:24.000000000 +0200 +@@ -0,0 +1,76 @@ ++ ++/* interface version */ ++ ++#define VCI_VERSION 0x00020308 ++ ++ ++enum { ++ VCI_KCBIT_NO_DYNAMIC = 0, ++ ++ VCI_KCBIT_PROC_SECURE = 4, ++ /* VCI_KCBIT_HARDCPU = 5, */ ++ /* VCI_KCBIT_IDLELIMIT = 6, */ ++ /* VCI_KCBIT_IDLETIME = 7, */ ++ ++ VCI_KCBIT_COWBL = 8, ++ VCI_KCBIT_FULLCOWBL = 9, ++ VCI_KCBIT_SPACES = 10, ++ VCI_KCBIT_NETV2 = 11, ++ VCI_KCBIT_MEMCG = 12, ++ ++ VCI_KCBIT_DEBUG = 16, ++ VCI_KCBIT_HISTORY = 20, ++ VCI_KCBIT_TAGGED = 24, ++ VCI_KCBIT_PPTAG = 28, ++ ++ VCI_KCBIT_MORE = 31, ++}; ++ ++ ++static inline uint32_t vci_kernel_config(void) ++{ ++ return ++ (1 << VCI_KCBIT_NO_DYNAMIC) | ++ ++ /* configured features */ ++#ifdef CONFIG_VSERVER_PROC_SECURE ++ (1 << VCI_KCBIT_PROC_SECURE) | ++#endif ++#ifdef CONFIG_VSERVER_COWBL ++ (1 << VCI_KCBIT_COWBL) | ++ (1 << VCI_KCBIT_FULLCOWBL) | ++#endif ++ (1 << VCI_KCBIT_SPACES) | ++ (1 << VCI_KCBIT_NETV2) | ++#ifdef CONFIG_CGROUP_MEM_RES_CTLR ++ (1 << VCI_KCBIT_MEMCG) | ++#endif ++ ++ /* debug options */ ++#ifdef CONFIG_VSERVER_DEBUG ++ (1 << VCI_KCBIT_DEBUG) | ++#endif ++#ifdef CONFIG_VSERVER_HISTORY ++ (1 << VCI_KCBIT_HISTORY) | ++#endif ++ ++ /* inode context tagging */ ++#if defined(CONFIG_TAGGING_NONE) ++ (0 << VCI_KCBIT_TAGGED) | ++#elif defined(CONFIG_TAGGING_UID16) ++ (1 << VCI_KCBIT_TAGGED) | ++#elif defined(CONFIG_TAGGING_GID16) ++ (2 << VCI_KCBIT_TAGGED) | ++#elif defined(CONFIG_TAGGING_ID24) ++ (3 << VCI_KCBIT_TAGGED) | ++#elif defined(CONFIG_TAGGING_INTERN) ++ (4 << VCI_KCBIT_TAGGED) | ++#elif defined(CONFIG_TAGGING_RUNTIME) ++ (5 << VCI_KCBIT_TAGGED) | ++#else ++ (7 << VCI_KCBIT_TAGGED) | ++#endif ++ (1 << VCI_KCBIT_PPTAG) | ++ 0; ++} ++ +diff -NurpP --minimal linux-3.0.9/mm/filemap_xip.c linux-3.0.9-vs2.3.2.1/mm/filemap_xip.c +--- linux-3.0.9/mm/filemap_xip.c 2011-07-22 11:18:12.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/mm/filemap_xip.c 2011-06-10 22:11:24.000000000 +0200 +@@ -18,6 +18,7 @@ + #include <linux/seqlock.h> + #include <linux/mutex.h> + #include <linux/gfp.h> ++#include <linux/vs_memory.h> + #include <asm/tlbflush.h> + #include <asm/io.h> + +diff -NurpP --minimal linux-3.0.9/mm/fremap.c linux-3.0.9-vs2.3.2.1/mm/fremap.c +--- linux-3.0.9/mm/fremap.c 2011-07-22 11:18:12.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/mm/fremap.c 2011-06-10 22:11:24.000000000 +0200 +@@ -16,6 +16,7 @@ + #include <linux/module.h> + #include <linux/syscalls.h> + #include <linux/mmu_notifier.h> ++#include <linux/vs_memory.h> + + #include <asm/mmu_context.h> + #include <asm/cacheflush.h> +diff -NurpP --minimal linux-3.0.9/mm/hugetlb.c linux-3.0.9-vs2.3.2.1/mm/hugetlb.c +--- linux-3.0.9/mm/hugetlb.c 2011-07-22 11:18:12.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/mm/hugetlb.c 2011-06-22 12:39:16.000000000 +0200 +@@ -28,6 +28,7 @@ + + #include <linux/hugetlb.h> + #include <linux/node.h> ++#include <linux/vs_memory.h> + #include "internal.h" + + const unsigned long hugetlb_zero = 0, hugetlb_infinity = ~0UL; +diff -NurpP --minimal linux-3.0.9/mm/memcontrol.c linux-3.0.9-vs2.3.2.1/mm/memcontrol.c +--- linux-3.0.9/mm/memcontrol.c 2011-11-15 16:40:47.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/mm/memcontrol.c 2011-08-08 23:04:47.000000000 +0200 +@@ -741,6 +741,31 @@ struct mem_cgroup *mem_cgroup_from_task( + struct mem_cgroup, css); + } + ++u64 mem_cgroup_res_read_u64(struct mem_cgroup *mem, int member) ++{ ++ return res_counter_read_u64(&mem->res, member); ++} ++ ++u64 mem_cgroup_memsw_read_u64(struct mem_cgroup *mem, int member) ++{ ++ return res_counter_read_u64(&mem->memsw, member); ++} ++ ++s64 mem_cgroup_stat_read_cache(struct mem_cgroup *mem) ++{ ++ return mem_cgroup_read_stat(mem, MEM_CGROUP_STAT_CACHE); ++} ++ ++s64 mem_cgroup_stat_read_anon(struct mem_cgroup *mem) ++{ ++ return mem_cgroup_read_stat(mem, MEM_CGROUP_STAT_RSS); ++} ++ ++s64 mem_cgroup_stat_read_mapped(struct mem_cgroup *mem) ++{ ++ return mem_cgroup_read_stat(mem, MEM_CGROUP_STAT_FILE_MAPPED); ++} ++ + struct mem_cgroup *try_get_mem_cgroup_from_mm(struct mm_struct *mm) + { + struct mem_cgroup *mem = NULL; +diff -NurpP --minimal linux-3.0.9/mm/memory.c linux-3.0.9-vs2.3.2.1/mm/memory.c +--- linux-3.0.9/mm/memory.c 2011-11-15 16:40:47.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/mm/memory.c 2011-11-15 17:37:07.000000000 +0100 +@@ -3388,6 +3388,7 @@ int handle_pte_fault(struct mm_struct *m + { + pte_t entry; + spinlock_t *ptl; ++ int ret = 0, type = VXPT_UNKNOWN; + + entry = *pte; + if (!pte_present(entry)) { +@@ -3412,9 +3413,12 @@ int handle_pte_fault(struct mm_struct *m + if (unlikely(!pte_same(*pte, entry))) + goto unlock; + if (flags & FAULT_FLAG_WRITE) { +- if (!pte_write(entry)) +- return do_wp_page(mm, vma, address, ++ if (!pte_write(entry)) { ++ ret = do_wp_page(mm, vma, address, + pte, pmd, ptl, entry); ++ type = VXPT_WRITE; ++ goto out; ++ } + entry = pte_mkdirty(entry); + } + entry = pte_mkyoung(entry); +@@ -3432,7 +3436,10 @@ int handle_pte_fault(struct mm_struct *m + } + unlock: + pte_unmap_unlock(pte, ptl); +- return 0; ++ ret = 0; ++out: ++ vx_page_fault(mm, vma, type, ret); ++ return ret; + } + + /* +diff -NurpP --minimal linux-3.0.9/mm/mremap.c linux-3.0.9-vs2.3.2.1/mm/mremap.c +--- linux-3.0.9/mm/mremap.c 2011-07-22 11:18:12.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/mm/mremap.c 2011-06-10 22:11:24.000000000 +0200 +@@ -19,6 +19,7 @@ + #include <linux/security.h> + #include <linux/syscalls.h> + #include <linux/mmu_notifier.h> ++#include <linux/vs_memory.h> + + #include <asm/uaccess.h> + #include <asm/cacheflush.h> +diff -NurpP --minimal linux-3.0.9/mm/oom_kill.c linux-3.0.9-vs2.3.2.1/mm/oom_kill.c +--- linux-3.0.9/mm/oom_kill.c 2011-11-15 16:40:47.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/mm/oom_kill.c 2011-08-08 23:04:47.000000000 +0200 +@@ -32,6 +32,9 @@ + #include <linux/mempolicy.h> + #include <linux/security.h> + #include <linux/ptrace.h> ++#include <linux/reboot.h> ++#include <linux/vs_memory.h> ++#include <linux/vs_context.h> + + int sysctl_panic_on_oom; + int sysctl_oom_kill_allocating_task; +@@ -134,11 +137,18 @@ struct task_struct *find_lock_task_mm(st + static bool oom_unkillable_task(struct task_struct *p, + const struct mem_cgroup *mem, const nodemask_t *nodemask) + { +- if (is_global_init(p)) ++ unsigned xid = vx_current_xid(); ++ ++ /* skip the init task, global and per guest */ ++ if (task_is_init(p)) + return true; + if (p->flags & PF_KTHREAD) + return true; + ++ /* skip other guest and host processes if oom in guest */ ++ if (xid && vx_task_xid(p) != xid) ++ return true; ++ + /* When mem_cgroup_out_of_memory() and p is not member of the group */ + if (mem && !task_in_mem_cgroup(p, mem)) + return true; +@@ -214,6 +224,18 @@ unsigned int oom_badness(struct task_str + points += p->signal->oom_score_adj; + + /* ++ * add points for context badness and ++ * reduce badness for processes belonging to ++ * a different context ++ */ ++ ++ points += vx_badness(p, p->mm); ++ ++ if ((vx_current_xid() > 1) && ++ vx_current_xid() != vx_task_xid(p)) ++ points /= 16; ++ ++ /* + * Never return 0 for an eligible task that may be killed since it's + * possible that no single user task uses more than 0.1% of memory and + * no single admin tasks uses more than 3.0%. +@@ -429,8 +451,8 @@ static int oom_kill_task(struct task_str + /* mm cannot be safely dereferenced after task_unlock(p) */ + mm = p->mm; + +- pr_err("Killed process %d (%s) total-vm:%lukB, anon-rss:%lukB, file-rss:%lukB\n", +- task_pid_nr(p), p->comm, K(p->mm->total_vm), ++ pr_err("Killed process %d:#%u (%s) total-vm:%lukB, anon-rss:%lukB, file-rss:%lukB\n", ++ task_pid_nr(p), p->xid, p->comm, K(p->mm->total_vm), + K(get_mm_counter(p->mm, MM_ANONPAGES)), + K(get_mm_counter(p->mm, MM_FILEPAGES))); + task_unlock(p); +@@ -484,8 +506,8 @@ static int oom_kill_process(struct task_ + } + + task_lock(p); +- pr_err("%s: Kill process %d (%s) score %d or sacrifice child\n", +- message, task_pid_nr(p), p->comm, points); ++ pr_err("%s: Kill process %d:#%u (%s) score %d or sacrifice child\n", ++ message, task_pid_nr(p), p->xid, p->comm, points); + task_unlock(p); + + /* +@@ -586,6 +608,8 @@ int unregister_oom_notifier(struct notif + } + EXPORT_SYMBOL_GPL(unregister_oom_notifier); + ++long vs_oom_action(unsigned int); ++ + /* + * Try to acquire the OOM killer lock for the zones in zonelist. Returns zero + * if a parallel OOM killing is already taking place that includes a zone in +@@ -744,7 +768,12 @@ retry: + if (!p) { + dump_header(NULL, gfp_mask, order, NULL, mpol_mask); + read_unlock(&tasklist_lock); +- panic("Out of memory and no killable processes...\n"); ++ ++ /* avoid panic for guest OOM */ ++ if (current->xid) ++ vs_oom_action(LINUX_REBOOT_CMD_OOM); ++ else ++ panic("Out of memory and no killable processes...\n"); + } + + if (oom_kill_process(p, gfp_mask, order, points, totalpages, NULL, +diff -NurpP --minimal linux-3.0.9/mm/page_alloc.c linux-3.0.9-vs2.3.2.1/mm/page_alloc.c +--- linux-3.0.9/mm/page_alloc.c 2011-11-15 16:40:47.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/mm/page_alloc.c 2011-10-18 13:51:13.000000000 +0200 +@@ -57,6 +57,8 @@ + #include <linux/ftrace_event.h> + #include <linux/memcontrol.h> + #include <linux/prefetch.h> ++#include <linux/vs_base.h> ++#include <linux/vs_limit.h> + + #include <asm/tlbflush.h> + #include <asm/div64.h> +@@ -2502,6 +2504,9 @@ void si_meminfo(struct sysinfo *val) + val->totalhigh = totalhigh_pages; + val->freehigh = nr_free_highpages(); + val->mem_unit = PAGE_SIZE; ++ ++ if (vx_flags(VXF_VIRT_MEM, 0)) ++ vx_vsi_meminfo(val); + } + + EXPORT_SYMBOL(si_meminfo); +@@ -2522,6 +2527,9 @@ void si_meminfo_node(struct sysinfo *val + val->freehigh = 0; + #endif + val->mem_unit = PAGE_SIZE; ++ ++ if (vx_flags(VXF_VIRT_MEM, 0)) ++ vx_vsi_meminfo(val); + } + #endif + +diff -NurpP --minimal linux-3.0.9/mm/pgtable-generic.c linux-3.0.9-vs2.3.2.1/mm/pgtable-generic.c +--- linux-3.0.9/mm/pgtable-generic.c 2011-03-15 18:07:42.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/mm/pgtable-generic.c 2011-06-10 22:11:24.000000000 +0200 +@@ -6,6 +6,8 @@ + * Copyright (C) 2010 Linus Torvalds + */ + ++#include <linux/mm.h> ++ + #include <linux/pagemap.h> + #include <asm/tlb.h> + #include <asm-generic/pgtable.h> +diff -NurpP --minimal linux-3.0.9/mm/rmap.c linux-3.0.9-vs2.3.2.1/mm/rmap.c +--- linux-3.0.9/mm/rmap.c 2011-07-22 11:18:12.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/mm/rmap.c 2011-07-01 11:35:35.000000000 +0200 +@@ -57,6 +57,7 @@ + #include <linux/mmu_notifier.h> + #include <linux/migrate.h> + #include <linux/hugetlb.h> ++#include <linux/vs_memory.h> + + #include <asm/tlbflush.h> + +diff -NurpP --minimal linux-3.0.9/mm/shmem.c linux-3.0.9-vs2.3.2.1/mm/shmem.c +--- linux-3.0.9/mm/shmem.c 2011-07-22 11:18:12.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/mm/shmem.c 2011-07-01 11:35:35.000000000 +0200 +@@ -1850,7 +1850,7 @@ static int shmem_statfs(struct dentry *d + { + struct shmem_sb_info *sbinfo = SHMEM_SB(dentry->d_sb); + +- buf->f_type = TMPFS_MAGIC; ++ buf->f_type = TMPFS_SUPER_MAGIC; + buf->f_bsize = PAGE_CACHE_SIZE; + buf->f_namelen = NAME_MAX; + if (sbinfo->max_blocks) { +@@ -2605,7 +2605,7 @@ int shmem_fill_super(struct super_block + sb->s_maxbytes = SHMEM_MAX_BYTES; + sb->s_blocksize = PAGE_CACHE_SIZE; + sb->s_blocksize_bits = PAGE_CACHE_SHIFT; +- sb->s_magic = TMPFS_MAGIC; ++ sb->s_magic = TMPFS_SUPER_MAGIC; + sb->s_op = &shmem_ops; + sb->s_time_gran = 1; + #ifdef CONFIG_TMPFS_XATTR +diff -NurpP --minimal linux-3.0.9/mm/slab.c linux-3.0.9-vs2.3.2.1/mm/slab.c +--- linux-3.0.9/mm/slab.c 2011-07-22 11:18:12.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/mm/slab.c 2011-06-15 02:41:23.000000000 +0200 +@@ -411,6 +411,8 @@ static void kmem_list3_init(struct kmem_ + #define STATS_INC_FREEMISS(x) do { } while (0) + #endif + ++#include "slab_vs.h" ++ + #if DEBUG + + /* +@@ -3348,6 +3350,7 @@ retry: + + obj = slab_get_obj(cachep, slabp, nodeid); + check_slabp(cachep, slabp); ++ vx_slab_alloc(cachep, flags); + l3->free_objects--; + /* move slabp to correct slabp list: */ + list_del(&slabp->list); +@@ -3425,6 +3428,7 @@ __cache_alloc_node(struct kmem_cache *ca + /* ___cache_alloc_node can fall back to other nodes */ + ptr = ____cache_alloc_node(cachep, flags, nodeid); + out: ++ vx_slab_alloc(cachep, flags); + local_irq_restore(save_flags); + ptr = cache_alloc_debugcheck_after(cachep, flags, ptr, caller); + kmemleak_alloc_recursive(ptr, obj_size(cachep), 1, cachep->flags, +@@ -3612,6 +3616,7 @@ static inline void __cache_free(struct k + check_irq_off(); + kmemleak_free_recursive(objp, cachep->flags); + objp = cache_free_debugcheck(cachep, objp, caller); ++ vx_slab_free(cachep); + + kmemcheck_slab_free(cachep, objp, obj_size(cachep)); + +diff -NurpP --minimal linux-3.0.9/mm/slab_vs.h linux-3.0.9-vs2.3.2.1/mm/slab_vs.h +--- linux-3.0.9/mm/slab_vs.h 1970-01-01 01:00:00.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/mm/slab_vs.h 2011-06-10 22:11:24.000000000 +0200 +@@ -0,0 +1,29 @@ ++ ++#include <linux/vserver/context.h> ++ ++#include <linux/vs_context.h> ++ ++static inline ++void vx_slab_alloc(struct kmem_cache *cachep, gfp_t flags) ++{ ++ int what = gfp_zone(cachep->gfpflags); ++ struct vx_info *vxi = current_vx_info(); ++ ++ if (!vxi) ++ return; ++ ++ atomic_add(cachep->buffer_size, &vxi->cacct.slab[what]); ++} ++ ++static inline ++void vx_slab_free(struct kmem_cache *cachep) ++{ ++ int what = gfp_zone(cachep->gfpflags); ++ struct vx_info *vxi = current_vx_info(); ++ ++ if (!vxi) ++ return; ++ ++ atomic_sub(cachep->buffer_size, &vxi->cacct.slab[what]); ++} ++ +diff -NurpP --minimal linux-3.0.9/mm/swapfile.c linux-3.0.9-vs2.3.2.1/mm/swapfile.c +--- linux-3.0.9/mm/swapfile.c 2011-07-22 11:18:12.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/mm/swapfile.c 2011-07-01 11:35:35.000000000 +0200 +@@ -37,6 +37,8 @@ + #include <asm/tlbflush.h> + #include <linux/swapops.h> + #include <linux/page_cgroup.h> ++#include <linux/vs_base.h> ++#include <linux/vs_memory.h> + + static bool swap_count_continued(struct swap_info_struct *, pgoff_t, + unsigned char); +@@ -1759,6 +1761,16 @@ static int swap_show(struct seq_file *sw + + if (si == SEQ_START_TOKEN) { + seq_puts(swap,"Filename\t\t\t\tType\t\tSize\tUsed\tPriority\n"); ++ if (vx_flags(VXF_VIRT_MEM, 0)) { ++ struct sysinfo si; ++ ++ vx_vsi_swapinfo(&si); ++ if (si.totalswap < (1 << 10)) ++ return 0; ++ seq_printf(swap, "%s\t\t\t\t\t%s\t%lu\t%lu\t%d\n", ++ "hdv0", "partition", si.totalswap >> 10, ++ (si.totalswap - si.freeswap) >> 10, -1); ++ } + return 0; + } + +@@ -2186,6 +2198,8 @@ void si_swapinfo(struct sysinfo *val) + val->freeswap = nr_swap_pages + nr_to_be_unused; + val->totalswap = total_swap_pages + nr_to_be_unused; + spin_unlock(&swap_lock); ++ if (vx_flags(VXF_VIRT_MEM, 0)) ++ vx_vsi_swapinfo(val); + } + + /* +diff -NurpP --minimal linux-3.0.9/net/bridge/br_multicast.c linux-3.0.9-vs2.3.2.1/net/bridge/br_multicast.c +--- linux-3.0.9/net/bridge/br_multicast.c 2011-11-15 16:40:47.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/net/bridge/br_multicast.c 2011-10-18 13:51:13.000000000 +0200 +@@ -447,7 +447,7 @@ static struct sk_buff *br_ip6_multicast_ + ip6h->hop_limit = 1; + ipv6_addr_set(&ip6h->daddr, htonl(0xff020000), 0, 0, htonl(1)); + ipv6_dev_get_saddr(dev_net(br->dev), br->dev, &ip6h->daddr, 0, +- &ip6h->saddr); ++ &ip6h->saddr, NULL); + ipv6_eth_mc_map(&ip6h->daddr, eth->h_dest); + + hopopt = (u8 *)(ip6h + 1); +diff -NurpP --minimal linux-3.0.9/net/core/dev.c linux-3.0.9-vs2.3.2.1/net/core/dev.c +--- linux-3.0.9/net/core/dev.c 2011-11-15 16:40:47.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/net/core/dev.c 2011-11-15 17:37:07.000000000 +0100 +@@ -127,6 +127,7 @@ + #include <linux/in.h> + #include <linux/jhash.h> + #include <linux/random.h> ++#include <linux/vs_inet.h> + #include <trace/events/napi.h> + #include <trace/events/net.h> + #include <trace/events/skb.h> +@@ -609,7 +610,8 @@ struct net_device *__dev_get_by_name(str + struct hlist_head *head = dev_name_hash(net, name); + + hlist_for_each_entry(dev, p, head, name_hlist) +- if (!strncmp(dev->name, name, IFNAMSIZ)) ++ if (!strncmp(dev->name, name, IFNAMSIZ) && ++ nx_dev_visible(current_nx_info(), dev)) + return dev; + + return NULL; +@@ -635,7 +637,8 @@ struct net_device *dev_get_by_name_rcu(s + struct hlist_head *head = dev_name_hash(net, name); + + hlist_for_each_entry_rcu(dev, p, head, name_hlist) +- if (!strncmp(dev->name, name, IFNAMSIZ)) ++ if (!strncmp(dev->name, name, IFNAMSIZ) && ++ nx_dev_visible(current_nx_info(), dev)) + return dev; + + return NULL; +@@ -686,7 +689,8 @@ struct net_device *__dev_get_by_index(st + struct hlist_head *head = dev_index_hash(net, ifindex); + + hlist_for_each_entry(dev, p, head, index_hlist) +- if (dev->ifindex == ifindex) ++ if ((dev->ifindex == ifindex) && ++ nx_dev_visible(current_nx_info(), dev)) + return dev; + + return NULL; +@@ -704,7 +708,7 @@ EXPORT_SYMBOL(__dev_get_by_index); + * about locking. The caller must hold RCU lock. + */ + +-struct net_device *dev_get_by_index_rcu(struct net *net, int ifindex) ++struct net_device *dev_get_by_index_real_rcu(struct net *net, int ifindex) + { + struct hlist_node *p; + struct net_device *dev; +@@ -716,6 +720,16 @@ struct net_device *dev_get_by_index_rcu( + + return NULL; + } ++EXPORT_SYMBOL(dev_get_by_index_real_rcu); ++ ++struct net_device *dev_get_by_index_rcu(struct net *net, int ifindex) ++{ ++ struct net_device *dev = dev_get_by_index_real_rcu(net, ifindex); ++ ++ if (nx_dev_visible(current_nx_info(), dev)) ++ return dev; ++ return NULL; ++} + EXPORT_SYMBOL(dev_get_by_index_rcu); + + +@@ -764,7 +778,8 @@ struct net_device *dev_getbyhwaddr_rcu(s + + for_each_netdev_rcu(net, dev) + if (dev->type == type && +- !memcmp(dev->dev_addr, ha, dev->addr_len)) ++ !memcmp(dev->dev_addr, ha, dev->addr_len) && ++ nx_dev_visible(current_nx_info(), dev)) + return dev; + + return NULL; +@@ -776,9 +791,11 @@ struct net_device *__dev_getfirstbyhwtyp + struct net_device *dev; + + ASSERT_RTNL(); +- for_each_netdev(net, dev) +- if (dev->type == type) ++ for_each_netdev(net, dev) { ++ if ((dev->type == type) && ++ nx_dev_visible(current_nx_info(), dev)) + return dev; ++ } + + return NULL; + } +@@ -896,6 +913,8 @@ static int __dev_alloc_name(struct net * + continue; + if (i < 0 || i >= max_netdevices) + continue; ++ if (!nx_dev_visible(current_nx_info(), d)) ++ continue; + + /* avoid cases where sscanf is not exact inverse of printf */ + snprintf(buf, IFNAMSIZ, name, i); +@@ -3937,6 +3956,8 @@ static int dev_ifconf(struct net *net, c + + total = 0; + for_each_netdev(net, dev) { ++ if (!nx_dev_visible(current_nx_info(), dev)) ++ continue; + for (i = 0; i < NPROTO; i++) { + if (gifconf_list[i]) { + int done; +@@ -4011,6 +4032,10 @@ static void dev_seq_printf_stats(struct + struct rtnl_link_stats64 temp; + const struct rtnl_link_stats64 *stats = dev_get_stats(dev, &temp); + ++ /* device visible inside network context? */ ++ if (!nx_dev_visible(current_nx_info(), dev)) ++ return; ++ + seq_printf(seq, "%6s: %7llu %7llu %4llu %4llu %4llu %5llu %10llu %9llu " + "%8llu %7llu %4llu %4llu %4llu %5llu %7llu %10llu\n", + dev->name, stats->rx_bytes, stats->rx_packets, +diff -NurpP --minimal linux-3.0.9/net/core/rtnetlink.c linux-3.0.9-vs2.3.2.1/net/core/rtnetlink.c +--- linux-3.0.9/net/core/rtnetlink.c 2011-07-22 11:18:13.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/net/core/rtnetlink.c 2011-06-10 22:11:24.000000000 +0200 +@@ -1015,6 +1015,8 @@ static int rtnl_dump_ifinfo(struct sk_bu + hlist_for_each_entry_rcu(dev, node, head, index_hlist) { + if (idx < s_idx) + goto cont; ++ if (!nx_dev_visible(skb->sk->sk_nx_info, dev)) ++ continue; + if (rtnl_fill_ifinfo(skb, dev, RTM_NEWLINK, + NETLINK_CB(cb->skb).pid, + cb->nlh->nlmsg_seq, 0, +@@ -1848,6 +1850,9 @@ void rtmsg_ifinfo(int type, struct net_d + struct sk_buff *skb; + int err = -ENOBUFS; + ++ if (!nx_dev_visible(current_nx_info(), dev)) ++ return; ++ + skb = nlmsg_new(if_nlmsg_size(dev), GFP_KERNEL); + if (skb == NULL) + goto errout; +diff -NurpP --minimal linux-3.0.9/net/core/sock.c linux-3.0.9-vs2.3.2.1/net/core/sock.c +--- linux-3.0.9/net/core/sock.c 2011-11-15 16:40:47.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/net/core/sock.c 2011-11-15 17:37:07.000000000 +0100 +@@ -127,6 +127,10 @@ + #include <net/cls_cgroup.h> + + #include <linux/filter.h> ++#include <linux/vs_socket.h> ++#include <linux/vs_limit.h> ++#include <linux/vs_context.h> ++#include <linux/vs_network.h> + + #ifdef CONFIG_INET + #include <net/tcp.h> +@@ -1070,6 +1074,8 @@ static struct sock *sk_prot_alloc(struct + goto out_free_sec; + sk_tx_queue_clear(sk); + } ++ sock_vx_init(sk); ++ sock_nx_init(sk); + + return sk; + +@@ -1169,6 +1175,11 @@ static void __sk_free(struct sock *sk) + put_cred(sk->sk_peer_cred); + put_pid(sk->sk_peer_pid); + put_net(sock_net(sk)); ++ vx_sock_dec(sk); ++ clr_vx_info(&sk->sk_vx_info); ++ sk->sk_xid = -1; ++ clr_nx_info(&sk->sk_nx_info); ++ sk->sk_nid = -1; + sk_prot_free(sk->sk_prot_creator, sk); + } + +@@ -1216,6 +1227,8 @@ struct sock *sk_clone(const struct sock + + /* SANITY */ + get_net(sock_net(newsk)); ++ sock_vx_init(newsk); ++ sock_nx_init(newsk); + sk_node_init(&newsk->sk_node); + sock_lock_init(newsk); + bh_lock_sock(newsk); +@@ -1272,6 +1285,12 @@ struct sock *sk_clone(const struct sock + smp_wmb(); + atomic_set(&newsk->sk_refcnt, 2); + ++ set_vx_info(&newsk->sk_vx_info, sk->sk_vx_info); ++ newsk->sk_xid = sk->sk_xid; ++ vx_sock_inc(newsk); ++ set_nx_info(&newsk->sk_nx_info, sk->sk_nx_info); ++ newsk->sk_nid = sk->sk_nid; ++ + /* + * Increment the counter in the same struct proto as the master + * sock (sk_refcnt_debug_inc uses newsk->sk_prot->socks, that +@@ -2018,6 +2037,12 @@ void sock_init_data(struct socket *sock, + + sk->sk_stamp = ktime_set(-1L, 0); + ++ set_vx_info(&sk->sk_vx_info, current_vx_info()); ++ sk->sk_xid = vx_current_xid(); ++ vx_sock_inc(sk); ++ set_nx_info(&sk->sk_nx_info, current_nx_info()); ++ sk->sk_nid = nx_current_nid(); ++ + /* + * Before updating sk_refcnt, we must commit prior changes to memory + * (Documentation/RCU/rculist_nulls.txt for details) +diff -NurpP --minimal linux-3.0.9/net/ipv4/af_inet.c linux-3.0.9-vs2.3.2.1/net/ipv4/af_inet.c +--- linux-3.0.9/net/ipv4/af_inet.c 2011-07-22 11:18:13.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/net/ipv4/af_inet.c 2011-07-19 00:44:39.000000000 +0200 +@@ -117,6 +117,7 @@ + #ifdef CONFIG_IP_MROUTE + #include <linux/mroute.h> + #endif ++#include <linux/vs_limit.h> + + + /* The inetsw table contains everything that inet_create needs to +@@ -326,9 +327,13 @@ lookup_protocol: + } + + err = -EPERM; ++ if ((protocol == IPPROTO_ICMP) && ++ nx_capable(CAP_NET_RAW, NXC_RAW_ICMP)) ++ goto override; ++ + if (sock->type == SOCK_RAW && !kern && !capable(CAP_NET_RAW)) + goto out_rcu_unlock; +- ++override: + err = -EAFNOSUPPORT; + if (!inet_netns_ok(net, protocol)) + goto out_rcu_unlock; +@@ -452,6 +457,7 @@ int inet_bind(struct socket *sock, struc + struct sockaddr_in *addr = (struct sockaddr_in *)uaddr; + struct sock *sk = sock->sk; + struct inet_sock *inet = inet_sk(sk); ++ struct nx_v4_sock_addr nsa; + unsigned short snum; + int chk_addr_ret; + int err; +@@ -470,7 +476,11 @@ int inet_bind(struct socket *sock, struc + goto out; + } + +- chk_addr_ret = inet_addr_type(sock_net(sk), addr->sin_addr.s_addr); ++ err = v4_map_sock_addr(inet, addr, &nsa); ++ if (err) ++ goto out; ++ ++ chk_addr_ret = inet_addr_type(sock_net(sk), nsa.saddr); + + /* Not specified by any standard per-se, however it breaks too + * many applications when removed. It is unfortunate since +@@ -482,7 +492,7 @@ int inet_bind(struct socket *sock, struc + err = -EADDRNOTAVAIL; + if (!sysctl_ip_nonlocal_bind && + !(inet->freebind || inet->transparent) && +- addr->sin_addr.s_addr != htonl(INADDR_ANY) && ++ nsa.saddr != htonl(INADDR_ANY) && + chk_addr_ret != RTN_LOCAL && + chk_addr_ret != RTN_MULTICAST && + chk_addr_ret != RTN_BROADCAST) +@@ -507,7 +517,7 @@ int inet_bind(struct socket *sock, struc + if (sk->sk_state != TCP_CLOSE || inet->inet_num) + goto out_release_sock; + +- inet->inet_rcv_saddr = inet->inet_saddr = addr->sin_addr.s_addr; ++ v4_set_sock_addr(inet, &nsa); + if (chk_addr_ret == RTN_MULTICAST || chk_addr_ret == RTN_BROADCAST) + inet->inet_saddr = 0; /* Use device */ + +@@ -710,11 +720,13 @@ int inet_getname(struct socket *sock, st + peer == 1)) + return -ENOTCONN; + sin->sin_port = inet->inet_dport; +- sin->sin_addr.s_addr = inet->inet_daddr; ++ sin->sin_addr.s_addr = ++ nx_map_sock_lback(sk->sk_nx_info, inet->inet_daddr); + } else { + __be32 addr = inet->inet_rcv_saddr; + if (!addr) + addr = inet->inet_saddr; ++ addr = nx_map_sock_lback(sk->sk_nx_info, addr); + sin->sin_port = inet->inet_sport; + sin->sin_addr.s_addr = addr; + } +diff -NurpP --minimal linux-3.0.9/net/ipv4/arp.c linux-3.0.9-vs2.3.2.1/net/ipv4/arp.c +--- linux-3.0.9/net/ipv4/arp.c 2011-05-22 16:18:00.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/net/ipv4/arp.c 2011-08-29 05:14:51.000000000 +0200 +@@ -1365,6 +1365,7 @@ static void arp_format_neigh_entry(struc + struct net_device *dev = n->dev; + int hatype = dev->type; + ++ /* FIXME: check for network context */ + read_lock(&n->lock); + /* Convert hardware address to XX:XX:XX:XX ... form. */ + #if defined(CONFIG_AX25) || defined(CONFIG_AX25_MODULE) +@@ -1396,6 +1397,7 @@ static void arp_format_pneigh_entry(stru + int hatype = dev ? dev->type : 0; + char tbuf[16]; + ++ /* FIXME: check for network context */ + sprintf(tbuf, "%pI4", n->key); + seq_printf(seq, "%-16s 0x%-10x0x%-10x%s * %s\n", + tbuf, hatype, ATF_PUBL | ATF_PERM, "00:00:00:00:00:00", +diff -NurpP --minimal linux-3.0.9/net/ipv4/devinet.c linux-3.0.9-vs2.3.2.1/net/ipv4/devinet.c +--- linux-3.0.9/net/ipv4/devinet.c 2011-11-15 16:40:47.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/net/ipv4/devinet.c 2011-08-29 03:45:09.000000000 +0200 +@@ -518,6 +518,7 @@ struct in_device *inetdev_by_index(struc + } + EXPORT_SYMBOL(inetdev_by_index); + ++ + /* Called only from RTNL semaphored context. No locks. */ + + struct in_ifaddr *inet_ifa_byprefix(struct in_device *in_dev, __be32 prefix, +@@ -759,6 +760,8 @@ int devinet_ioctl(struct net *net, unsig + + in_dev = __in_dev_get_rtnl(dev); + if (in_dev) { ++ struct nx_info *nxi = current_nx_info(); ++ + if (tryaddrmatch) { + /* Matthias Andree */ + /* compare label and address (4.4BSD style) */ +@@ -767,6 +770,8 @@ int devinet_ioctl(struct net *net, unsig + This is checked above. */ + for (ifap = &in_dev->ifa_list; (ifa = *ifap) != NULL; + ifap = &ifa->ifa_next) { ++ if (!nx_v4_ifa_visible(nxi, ifa)) ++ continue; + if (!strcmp(ifr.ifr_name, ifa->ifa_label) && + sin_orig.sin_addr.s_addr == + ifa->ifa_local) { +@@ -779,9 +784,12 @@ int devinet_ioctl(struct net *net, unsig + comparing just the label */ + if (!ifa) { + for (ifap = &in_dev->ifa_list; (ifa = *ifap) != NULL; +- ifap = &ifa->ifa_next) ++ ifap = &ifa->ifa_next) { ++ if (!nx_v4_ifa_visible(nxi, ifa)) ++ continue; + if (!strcmp(ifr.ifr_name, ifa->ifa_label)) + break; ++ } + } + } + +@@ -934,6 +942,8 @@ static int inet_gifconf(struct net_devic + goto out; + + for (ifa = in_dev->ifa_list; ifa; ifa = ifa->ifa_next) { ++ if (!nx_v4_ifa_visible(current_nx_info(), ifa)) ++ continue; + if (!buf) { + done += sizeof(ifr); + continue; +@@ -1294,6 +1304,7 @@ static int inet_dump_ifaddr(struct sk_bu + struct net_device *dev; + struct in_device *in_dev; + struct in_ifaddr *ifa; ++ struct sock *sk = skb->sk; + struct hlist_head *head; + struct hlist_node *node; + +@@ -1316,6 +1327,8 @@ static int inet_dump_ifaddr(struct sk_bu + + for (ifa = in_dev->ifa_list, ip_idx = 0; ifa; + ifa = ifa->ifa_next, ip_idx++) { ++ if (sk && !nx_v4_ifa_visible(sk->sk_nx_info, ifa)) ++ continue; + if (ip_idx < s_ip_idx) + continue; + if (inet_fill_ifaddr(skb, ifa, +diff -NurpP --minimal linux-3.0.9/net/ipv4/fib_trie.c linux-3.0.9-vs2.3.2.1/net/ipv4/fib_trie.c +--- linux-3.0.9/net/ipv4/fib_trie.c 2011-07-22 11:18:13.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/net/ipv4/fib_trie.c 2011-08-29 04:56:47.000000000 +0200 +@@ -2554,6 +2554,7 @@ static int fib_route_seq_show(struct seq + || fa->fa_type == RTN_MULTICAST) + continue; + ++ /* FIXME: check for network context? */ + if (fi) + seq_printf(seq, + "%s\t%08X\t%08X\t%04X\t%d\t%u\t" +diff -NurpP --minimal linux-3.0.9/net/ipv4/inet_connection_sock.c linux-3.0.9-vs2.3.2.1/net/ipv4/inet_connection_sock.c +--- linux-3.0.9/net/ipv4/inet_connection_sock.c 2011-07-22 11:18:13.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/net/ipv4/inet_connection_sock.c 2011-06-10 22:11:24.000000000 +0200 +@@ -52,6 +52,37 @@ void inet_get_local_port_range(int *low, + } + EXPORT_SYMBOL(inet_get_local_port_range); + ++int ipv4_rcv_saddr_equal(const struct sock *sk1, const struct sock *sk2) ++{ ++ __be32 sk1_rcv_saddr = sk_rcv_saddr(sk1), ++ sk2_rcv_saddr = sk_rcv_saddr(sk2); ++ ++ if (inet_v6_ipv6only(sk2)) ++ return 0; ++ ++ if (sk1_rcv_saddr && ++ sk2_rcv_saddr && ++ sk1_rcv_saddr == sk2_rcv_saddr) ++ return 1; ++ ++ if (sk1_rcv_saddr && ++ !sk2_rcv_saddr && ++ v4_addr_in_nx_info(sk2->sk_nx_info, sk1_rcv_saddr, NXA_MASK_BIND)) ++ return 1; ++ ++ if (sk2_rcv_saddr && ++ !sk1_rcv_saddr && ++ v4_addr_in_nx_info(sk1->sk_nx_info, sk2_rcv_saddr, NXA_MASK_BIND)) ++ return 1; ++ ++ if (!sk1_rcv_saddr && ++ !sk2_rcv_saddr && ++ nx_v4_addr_conflict(sk1->sk_nx_info, sk2->sk_nx_info)) ++ return 1; ++ ++ return 0; ++} ++ + int inet_csk_bind_conflict(const struct sock *sk, + const struct inet_bind_bucket *tb) + { +@@ -74,9 +105,7 @@ int inet_csk_bind_conflict(const struct + sk->sk_bound_dev_if == sk2->sk_bound_dev_if)) { + if (!reuse || !sk2->sk_reuse || + sk2->sk_state == TCP_LISTEN) { +- const __be32 sk2_rcv_saddr = sk_rcv_saddr(sk2); +- if (!sk2_rcv_saddr || !sk_rcv_saddr(sk) || +- sk2_rcv_saddr == sk_rcv_saddr(sk)) ++ if (ipv4_rcv_saddr_equal(sk, sk2)) + break; + } + } +diff -NurpP --minimal linux-3.0.9/net/ipv4/inet_diag.c linux-3.0.9-vs2.3.2.1/net/ipv4/inet_diag.c +--- linux-3.0.9/net/ipv4/inet_diag.c 2011-07-22 11:18:13.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/net/ipv4/inet_diag.c 2011-06-22 12:39:16.000000000 +0200 +@@ -33,6 +33,8 @@ + #include <linux/stddef.h> + + #include <linux/inet_diag.h> ++#include <linux/vs_network.h> ++#include <linux/vs_inet.h> + + static const struct inet_diag_handler **inet_diag_table; + +@@ -119,8 +121,10 @@ static int inet_csk_diag_fill(struct soc + + r->id.idiag_sport = inet->inet_sport; + r->id.idiag_dport = inet->inet_dport; +- r->id.idiag_src[0] = inet->inet_rcv_saddr; +- r->id.idiag_dst[0] = inet->inet_daddr; ++ r->id.idiag_src[0] = nx_map_sock_lback(sk->sk_nx_info, ++ inet->inet_rcv_saddr); ++ r->id.idiag_dst[0] = nx_map_sock_lback(sk->sk_nx_info, ++ inet->inet_daddr); + + #if defined(CONFIG_IPV6) || defined (CONFIG_IPV6_MODULE) + if (r->idiag_family == AF_INET6) { +@@ -205,8 +209,8 @@ static int inet_twsk_diag_fill(struct in + r->id.idiag_cookie[1] = (u32)(((unsigned long)tw >> 31) >> 1); + r->id.idiag_sport = tw->tw_sport; + r->id.idiag_dport = tw->tw_dport; +- r->id.idiag_src[0] = tw->tw_rcv_saddr; +- r->id.idiag_dst[0] = tw->tw_daddr; ++ r->id.idiag_src[0] = nx_map_sock_lback(tw->tw_nx_info, tw->tw_rcv_saddr); ++ r->id.idiag_dst[0] = nx_map_sock_lback(tw->tw_nx_info, tw->tw_daddr); + r->idiag_state = tw->tw_substate; + r->idiag_timer = 3; + r->idiag_expires = DIV_ROUND_UP(tmo * 1000, HZ); +@@ -263,6 +267,7 @@ static int inet_diag_get_exact(struct sk + err = -EINVAL; + + if (req->idiag_family == AF_INET) { ++ /* TODO: lback */ + sk = inet_lookup(&init_net, hashinfo, req->id.idiag_dst[0], + req->id.idiag_dport, req->id.idiag_src[0], + req->id.idiag_sport, req->id.idiag_if); +@@ -505,6 +510,7 @@ static int inet_csk_diag_dump(struct soc + } else + #endif + { ++ /* TODO: lback */ + entry.saddr = &inet->inet_rcv_saddr; + entry.daddr = &inet->inet_daddr; + } +@@ -543,6 +549,7 @@ static int inet_twsk_diag_dump(struct in + } else + #endif + { ++ /* TODO: lback */ + entry.saddr = &tw->tw_rcv_saddr; + entry.daddr = &tw->tw_daddr; + } +@@ -589,8 +596,8 @@ static int inet_diag_fill_req(struct sk_ + + r->id.idiag_sport = inet->inet_sport; + r->id.idiag_dport = ireq->rmt_port; +- r->id.idiag_src[0] = ireq->loc_addr; +- r->id.idiag_dst[0] = ireq->rmt_addr; ++ r->id.idiag_src[0] = nx_map_sock_lback(sk->sk_nx_info, ireq->loc_addr); ++ r->id.idiag_dst[0] = nx_map_sock_lback(sk->sk_nx_info, ireq->rmt_addr); + r->idiag_expires = jiffies_to_msecs(tmo); + r->idiag_rqueue = 0; + r->idiag_wqueue = 0; +@@ -661,6 +668,7 @@ static int inet_diag_dump_reqs(struct sk + continue; + + if (bc) { ++ /* TODO: lback */ + entry.saddr = + #if defined(CONFIG_IPV6) || defined (CONFIG_IPV6_MODULE) + (entry.family == AF_INET6) ? +@@ -731,6 +739,8 @@ static int inet_diag_dump(struct sk_buff + sk_nulls_for_each(sk, node, &ilb->head) { + struct inet_sock *inet = inet_sk(sk); + ++ if (!nx_check(sk->sk_nid, VS_WATCH_P | VS_IDENT)) ++ continue; + if (num < s_num) { + num++; + continue; +@@ -797,6 +807,8 @@ skip_listen_ht: + sk_nulls_for_each(sk, node, &head->chain) { + struct inet_sock *inet = inet_sk(sk); + ++ if (!nx_check(sk->sk_nid, VS_WATCH_P | VS_IDENT)) ++ continue; + if (num < s_num) + goto next_normal; + if (!(r->idiag_states & (1 << sk->sk_state))) +@@ -821,6 +833,8 @@ next_normal: + inet_twsk_for_each(tw, node, + &head->twchain) { + ++ if (!nx_check(tw->tw_nid, VS_WATCH_P | VS_IDENT)) ++ continue; + if (num < s_num) + goto next_dying; + if (r->id.idiag_sport != tw->tw_sport && +diff -NurpP --minimal linux-3.0.9/net/ipv4/inet_hashtables.c linux-3.0.9-vs2.3.2.1/net/ipv4/inet_hashtables.c +--- linux-3.0.9/net/ipv4/inet_hashtables.c 2011-11-15 16:40:47.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/net/ipv4/inet_hashtables.c 2011-08-29 03:45:56.000000000 +0200 +@@ -22,6 +22,7 @@ + #include <net/inet_connection_sock.h> + #include <net/inet_hashtables.h> + #include <net/secure_seq.h> ++#include <net/route.h> + #include <net/ip.h> + + /* +@@ -156,6 +157,11 @@ static inline int compute_score(struct s + if (rcv_saddr != daddr) + return -1; + score += 2; ++ } else { ++ /* block non nx_info ips */ ++ if (!v4_addr_in_nx_info(sk->sk_nx_info, ++ daddr, NXA_MASK_BIND)) ++ return -1; + } + if (sk->sk_bound_dev_if) { + if (sk->sk_bound_dev_if != dif) +@@ -173,7 +179,6 @@ static inline int compute_score(struct s + * wildcarded during the search since they can never be otherwise. + */ + +- + struct sock *__inet_lookup_listener(struct net *net, + struct inet_hashinfo *hashinfo, + const __be32 daddr, const unsigned short hnum, +@@ -196,6 +201,7 @@ begin: + hiscore = score; + } + } ++ + /* + * if the nulls value we got at the end of this lookup is + * not the expected one, we must restart lookup. +diff -NurpP --minimal linux-3.0.9/net/ipv4/netfilter/nf_nat_helper.c linux-3.0.9-vs2.3.2.1/net/ipv4/netfilter/nf_nat_helper.c +--- linux-3.0.9/net/ipv4/netfilter/nf_nat_helper.c 2011-07-22 11:18:13.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/net/ipv4/netfilter/nf_nat_helper.c 2011-06-15 02:40:14.000000000 +0200 +@@ -20,6 +20,7 @@ + #include <net/route.h> + + #include <linux/netfilter_ipv4.h> ++#include <net/route.h> + #include <net/netfilter/nf_conntrack.h> + #include <net/netfilter/nf_conntrack_helper.h> + #include <net/netfilter/nf_conntrack_ecache.h> +diff -NurpP --minimal linux-3.0.9/net/ipv4/netfilter.c linux-3.0.9-vs2.3.2.1/net/ipv4/netfilter.c +--- linux-3.0.9/net/ipv4/netfilter.c 2011-11-15 16:40:47.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/net/ipv4/netfilter.c 2011-10-18 13:51:13.000000000 +0200 +@@ -5,7 +5,7 @@ + #include <linux/ip.h> + #include <linux/skbuff.h> + #include <linux/gfp.h> +-#include <net/route.h> ++// #include <net/route.h> + #include <net/xfrm.h> + #include <net/ip.h> + #include <net/netfilter/nf_queue.h> +diff -NurpP --minimal linux-3.0.9/net/ipv4/raw.c linux-3.0.9-vs2.3.2.1/net/ipv4/raw.c +--- linux-3.0.9/net/ipv4/raw.c 2011-07-22 11:18:13.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/net/ipv4/raw.c 2011-07-27 20:26:04.000000000 +0200 +@@ -117,7 +117,7 @@ static struct sock *__raw_v4_lookup(stru + + if (net_eq(sock_net(sk), net) && inet->inet_num == num && + !(inet->inet_daddr && inet->inet_daddr != raddr) && +- !(inet->inet_rcv_saddr && inet->inet_rcv_saddr != laddr) && ++ v4_sock_addr_match(sk->sk_nx_info, inet, laddr) && + !(sk->sk_bound_dev_if && sk->sk_bound_dev_if != dif)) + goto found; /* gotcha */ + } +@@ -385,6 +385,12 @@ static int raw_send_hdrinc(struct sock * + icmp_out_count(net, ((struct icmphdr *) + skb_transport_header(skb))->type); + ++ err = -EPERM; ++ if (!nx_check(0, VS_ADMIN) && !capable(CAP_NET_RAW) && ++ sk->sk_nx_info && ++ !v4_addr_in_nx_info(sk->sk_nx_info, iph->saddr, NXA_MASK_BIND)) ++ goto error_free; ++ + err = NF_HOOK(NFPROTO_IPV4, NF_INET_LOCAL_OUT, skb, NULL, + rt->dst.dev, dst_output); + if (err > 0) +@@ -571,6 +577,16 @@ static int raw_sendmsg(struct kiocb *ioc + goto done; + } + ++ if (sk->sk_nx_info) { ++ rt = ip_v4_find_src(sock_net(sk), sk->sk_nx_info, &fl4); ++ if (IS_ERR(rt)) { ++ err = PTR_ERR(rt); ++ rt = NULL; ++ goto done; ++ } ++ ip_rt_put(rt); ++ } ++ + security_sk_classify_flow(sk, flowi4_to_flowi(&fl4)); + rt = ip_route_output_flow(sock_net(sk), &fl4, sk); + if (IS_ERR(rt)) { +@@ -647,17 +663,19 @@ static int raw_bind(struct sock *sk, str + { + struct inet_sock *inet = inet_sk(sk); + struct sockaddr_in *addr = (struct sockaddr_in *) uaddr; ++ struct nx_v4_sock_addr nsa = { 0 }; + int ret = -EINVAL; + int chk_addr_ret; + + if (sk->sk_state != TCP_CLOSE || addr_len < sizeof(struct sockaddr_in)) + goto out; +- chk_addr_ret = inet_addr_type(sock_net(sk), addr->sin_addr.s_addr); ++ v4_map_sock_addr(inet, addr, &nsa); ++ chk_addr_ret = inet_addr_type(sock_net(sk), nsa.saddr); + ret = -EADDRNOTAVAIL; +- if (addr->sin_addr.s_addr && chk_addr_ret != RTN_LOCAL && ++ if (nsa.saddr && chk_addr_ret != RTN_LOCAL && + chk_addr_ret != RTN_MULTICAST && chk_addr_ret != RTN_BROADCAST) + goto out; +- inet->inet_rcv_saddr = inet->inet_saddr = addr->sin_addr.s_addr; ++ v4_set_sock_addr(inet, &nsa); + if (chk_addr_ret == RTN_MULTICAST || chk_addr_ret == RTN_BROADCAST) + inet->inet_saddr = 0; /* Use device */ + sk_dst_reset(sk); +@@ -709,7 +727,8 @@ static int raw_recvmsg(struct kiocb *ioc + /* Copy the address. */ + if (sin) { + sin->sin_family = AF_INET; +- sin->sin_addr.s_addr = ip_hdr(skb)->saddr; ++ sin->sin_addr.s_addr = ++ nx_map_sock_lback(sk->sk_nx_info, ip_hdr(skb)->saddr); + sin->sin_port = 0; + memset(&sin->sin_zero, 0, sizeof(sin->sin_zero)); + } +@@ -905,7 +924,8 @@ static struct sock *raw_get_first(struct + struct hlist_node *node; + + sk_for_each(sk, node, &state->h->ht[state->bucket]) +- if (sock_net(sk) == seq_file_net(seq)) ++ if ((sock_net(sk) == seq_file_net(seq)) && ++ nx_check(sk->sk_nid, VS_WATCH_P | VS_IDENT)) + goto found; + } + sk = NULL; +@@ -921,7 +941,8 @@ static struct sock *raw_get_next(struct + sk = sk_next(sk); + try_again: + ; +- } while (sk && sock_net(sk) != seq_file_net(seq)); ++ } while (sk && ((sock_net(sk) != seq_file_net(seq)) || ++ !nx_check(sk->sk_nid, VS_WATCH_P | VS_IDENT))); + + if (!sk && ++state->bucket < RAW_HTABLE_SIZE) { + sk = sk_head(&state->h->ht[state->bucket]); +diff -NurpP --minimal linux-3.0.9/net/ipv4/route.c linux-3.0.9-vs2.3.2.1/net/ipv4/route.c +--- linux-3.0.9/net/ipv4/route.c 2011-11-15 16:40:47.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/net/ipv4/route.c 2011-10-18 13:51:13.000000000 +0200 +@@ -2523,7 +2523,7 @@ static struct rtable *ip_route_output_sl + + + if (fl4->flowi4_oif) { +- dev_out = dev_get_by_index_rcu(net, fl4->flowi4_oif); ++ dev_out = dev_get_by_index_real_rcu(net, fl4->flowi4_oif); + rth = ERR_PTR(-ENODEV); + if (dev_out == NULL) + goto out; +diff -NurpP --minimal linux-3.0.9/net/ipv4/tcp.c linux-3.0.9-vs2.3.2.1/net/ipv4/tcp.c +--- linux-3.0.9/net/ipv4/tcp.c 2011-07-22 11:18:13.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/net/ipv4/tcp.c 2011-07-19 00:44:39.000000000 +0200 +@@ -266,6 +266,7 @@ + #include <linux/crypto.h> + #include <linux/time.h> + #include <linux/slab.h> ++#include <linux/in.h> + + #include <net/icmp.h> + #include <net/tcp.h> +diff -NurpP --minimal linux-3.0.9/net/ipv4/tcp_ipv4.c linux-3.0.9-vs2.3.2.1/net/ipv4/tcp_ipv4.c +--- linux-3.0.9/net/ipv4/tcp_ipv4.c 2011-11-15 16:40:47.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/net/ipv4/tcp_ipv4.c 2011-11-15 17:37:07.000000000 +0100 +@@ -2006,6 +2006,12 @@ static void *listening_get_next(struct s + req = req->dl_next; + while (1) { + while (req) { ++ vxdprintk(VXD_CBIT(net, 6), ++ "sk,req: %p [#%d] (from %d)", req->sk, ++ (req->sk)?req->sk->sk_nid:0, nx_current_nid()); ++ if (req->sk && ++ !nx_check(req->sk->sk_nid, VS_WATCH_P | VS_IDENT)) ++ continue; + if (req->rsk_ops->family == st->family) { + cur = req; + goto out; +@@ -2030,6 +2036,10 @@ get_req: + } + get_sk: + sk_nulls_for_each_from(sk, node) { ++ vxdprintk(VXD_CBIT(net, 6), "sk: %p [#%d] (from %d)", ++ sk, sk->sk_nid, nx_current_nid()); ++ if (!nx_check(sk->sk_nid, VS_WATCH_P | VS_IDENT)) ++ continue; + if (!net_eq(sock_net(sk), net)) + continue; + if (sk->sk_family == st->family) { +@@ -2106,6 +2116,11 @@ static void *established_get_first(struc + + spin_lock_bh(lock); + sk_nulls_for_each(sk, node, &tcp_hashinfo.ehash[st->bucket].chain) { ++ vxdprintk(VXD_CBIT(net, 6), ++ "sk,egf: %p [#%d] (from %d)", ++ sk, sk->sk_nid, nx_current_nid()); ++ if (!nx_check(sk->sk_nid, VS_WATCH_P | VS_IDENT)) ++ continue; + if (sk->sk_family != st->family || + !net_eq(sock_net(sk), net)) { + continue; +@@ -2116,6 +2131,11 @@ static void *established_get_first(struc + st->state = TCP_SEQ_STATE_TIME_WAIT; + inet_twsk_for_each(tw, node, + &tcp_hashinfo.ehash[st->bucket].twchain) { ++ vxdprintk(VXD_CBIT(net, 6), ++ "tw: %p [#%d] (from %d)", ++ tw, tw->tw_nid, nx_current_nid()); ++ if (!nx_check(tw->tw_nid, VS_WATCH_P | VS_IDENT)) ++ continue; + if (tw->tw_family != st->family || + !net_eq(twsk_net(tw), net)) { + continue; +@@ -2145,7 +2165,9 @@ static void *established_get_next(struct + tw = cur; + tw = tw_next(tw); + get_tw: +- while (tw && (tw->tw_family != st->family || !net_eq(twsk_net(tw), net))) { ++ while (tw && (tw->tw_family != st->family || ++ !net_eq(twsk_net(tw), net) || ++ !nx_check(tw->tw_nid, VS_WATCH_P | VS_IDENT))) { + tw = tw_next(tw); + } + if (tw) { +@@ -2169,6 +2191,11 @@ get_tw: + sk = sk_nulls_next(sk); + + sk_nulls_for_each_from(sk, node) { ++ vxdprintk(VXD_CBIT(net, 6), ++ "sk,egn: %p [#%d] (from %d)", ++ sk, sk->sk_nid, nx_current_nid()); ++ if (!nx_check(sk->sk_nid, VS_WATCH_P | VS_IDENT)) ++ continue; + if (sk->sk_family == st->family && net_eq(sock_net(sk), net)) + goto found; + } +@@ -2378,9 +2405,9 @@ static void get_openreq4(struct sock *sk + seq_printf(f, "%4d: %08X:%04X %08X:%04X" + " %02X %08X:%08X %02X:%08lX %08X %5d %8d %u %d %pK%n", + i, +- ireq->loc_addr, ++ nx_map_sock_lback(current_nx_info(), ireq->loc_addr), + ntohs(inet_sk(sk)->inet_sport), +- ireq->rmt_addr, ++ nx_map_sock_lback(current_nx_info(), ireq->rmt_addr), + ntohs(ireq->rmt_port), + TCP_SYN_RECV, + 0, 0, /* could print option size, but that is af dependent. */ +@@ -2402,8 +2429,8 @@ static void get_tcp4_sock(struct sock *s + struct tcp_sock *tp = tcp_sk(sk); + const struct inet_connection_sock *icsk = inet_csk(sk); + struct inet_sock *inet = inet_sk(sk); +- __be32 dest = inet->inet_daddr; +- __be32 src = inet->inet_rcv_saddr; ++ __be32 dest = nx_map_sock_lback(current_nx_info(), inet->inet_daddr); ++ __be32 src = nx_map_sock_lback(current_nx_info(), inet->inet_rcv_saddr); + __u16 destp = ntohs(inet->inet_dport); + __u16 srcp = ntohs(inet->inet_sport); + int rx_queue; +@@ -2460,8 +2487,8 @@ static void get_timewait4_sock(struct in + if (ttd < 0) + ttd = 0; + +- dest = tw->tw_daddr; +- src = tw->tw_rcv_saddr; ++ dest = nx_map_sock_lback(current_nx_info(), tw->tw_daddr); ++ src = nx_map_sock_lback(current_nx_info(), tw->tw_rcv_saddr); + destp = ntohs(tw->tw_dport); + srcp = ntohs(tw->tw_sport); + +diff -NurpP --minimal linux-3.0.9/net/ipv4/tcp_minisocks.c linux-3.0.9-vs2.3.2.1/net/ipv4/tcp_minisocks.c +--- linux-3.0.9/net/ipv4/tcp_minisocks.c 2011-03-15 18:07:45.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/net/ipv4/tcp_minisocks.c 2011-06-10 22:11:24.000000000 +0200 +@@ -23,6 +23,9 @@ + #include <linux/slab.h> + #include <linux/sysctl.h> + #include <linux/workqueue.h> ++#include <linux/vs_limit.h> ++#include <linux/vs_socket.h> ++#include <linux/vs_context.h> + #include <net/tcp.h> + #include <net/inet_common.h> + #include <net/xfrm.h> +@@ -335,6 +338,11 @@ void tcp_time_wait(struct sock *sk, int + tcptw->tw_ts_recent = tp->rx_opt.ts_recent; + tcptw->tw_ts_recent_stamp = tp->rx_opt.ts_recent_stamp; + ++ tw->tw_xid = sk->sk_xid; ++ tw->tw_vx_info = NULL; ++ tw->tw_nid = sk->sk_nid; ++ tw->tw_nx_info = NULL; ++ + #if defined(CONFIG_IPV6) || defined(CONFIG_IPV6_MODULE) + if (tw->tw_family == PF_INET6) { + struct ipv6_pinfo *np = inet6_sk(sk); +diff -NurpP --minimal linux-3.0.9/net/ipv4/udp.c linux-3.0.9-vs2.3.2.1/net/ipv4/udp.c +--- linux-3.0.9/net/ipv4/udp.c 2011-07-22 11:18:13.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/net/ipv4/udp.c 2011-07-27 20:26:14.000000000 +0200 +@@ -296,14 +296,7 @@ fail: + } + EXPORT_SYMBOL(udp_lib_get_port); + +-static int ipv4_rcv_saddr_equal(const struct sock *sk1, const struct sock *sk2) +-{ +- struct inet_sock *inet1 = inet_sk(sk1), *inet2 = inet_sk(sk2); +- +- return (!ipv6_only_sock(sk2) && +- (!inet1->inet_rcv_saddr || !inet2->inet_rcv_saddr || +- inet1->inet_rcv_saddr == inet2->inet_rcv_saddr)); +-} ++extern int ipv4_rcv_saddr_equal(const struct sock *, const struct sock *); + + static unsigned int udp4_portaddr_hash(struct net *net, __be32 saddr, + unsigned int port) +@@ -338,6 +331,11 @@ static inline int compute_score(struct s + if (inet->inet_rcv_saddr != daddr) + return -1; + score += 2; ++ } else { ++ /* block non nx_info ips */ ++ if (!v4_addr_in_nx_info(sk->sk_nx_info, ++ daddr, NXA_MASK_BIND)) ++ return -1; + } + if (inet->inet_daddr) { + if (inet->inet_daddr != saddr) +@@ -441,6 +439,7 @@ exact_match: + return result; + } + ++ + /* UDP is nearly always wildcards out the wazoo, it makes no sense to try + * harder than this. -DaveM + */ +@@ -486,6 +485,11 @@ begin: + sk_nulls_for_each_rcu(sk, node, &hslot->head) { + score = compute_score(sk, net, saddr, hnum, sport, + daddr, dport, dif); ++ /* FIXME: disabled? ++ if (score == 9) { ++ result = sk; ++ break; ++ } else */ + if (score > badness) { + result = sk; + badness = score; +@@ -499,6 +503,7 @@ begin: + if (get_nulls_value(node) != slot) + goto begin; + ++ + if (result) { + if (unlikely(!atomic_inc_not_zero_hint(&result->sk_refcnt, 2))) + result = NULL; +@@ -508,6 +513,7 @@ begin: + goto begin; + } + } ++ + rcu_read_unlock(); + return result; + } +@@ -550,8 +556,7 @@ static inline struct sock *udp_v4_mcast_ + udp_sk(s)->udp_port_hash != hnum || + (inet->inet_daddr && inet->inet_daddr != rmt_addr) || + (inet->inet_dport != rmt_port && inet->inet_dport) || +- (inet->inet_rcv_saddr && +- inet->inet_rcv_saddr != loc_addr) || ++ !v4_sock_addr_match(sk->sk_nx_info, inet, loc_addr) || + ipv6_only_sock(s) || + (s->sk_bound_dev_if && s->sk_bound_dev_if != dif)) + continue; +@@ -929,6 +934,16 @@ int udp_sendmsg(struct kiocb *iocb, stru + inet_sk_flowi_flags(sk)|FLOWI_FLAG_CAN_SLEEP, + faddr, saddr, dport, inet->inet_sport); + ++ if (sk->sk_nx_info) { ++ rt = ip_v4_find_src(net, sk->sk_nx_info, fl4); ++ if (IS_ERR(rt)) { ++ err = PTR_ERR(rt); ++ rt = NULL; ++ goto out; ++ } ++ ip_rt_put(rt); ++ } ++ + security_sk_classify_flow(sk, flowi4_to_flowi(fl4)); + rt = ip_route_output_flow(net, fl4, sk); + if (IS_ERR(rt)) { +@@ -1226,7 +1241,8 @@ try_again: + if (sin) { + sin->sin_family = AF_INET; + sin->sin_port = udp_hdr(skb)->source; +- sin->sin_addr.s_addr = ip_hdr(skb)->saddr; ++ sin->sin_addr.s_addr = nx_map_sock_lback( ++ skb->sk->sk_nx_info, ip_hdr(skb)->saddr); + memset(sin->sin_zero, 0, sizeof(sin->sin_zero)); + } + if (inet->cmsg_flags) +@@ -1970,6 +1986,8 @@ static struct sock *udp_get_first(struct + sk_nulls_for_each(sk, node, &hslot->head) { + if (!net_eq(sock_net(sk), net)) + continue; ++ if (!nx_check(sk->sk_nid, VS_WATCH_P | VS_IDENT)) ++ continue; + if (sk->sk_family == state->family) + goto found; + } +@@ -1987,7 +2005,9 @@ static struct sock *udp_get_next(struct + + do { + sk = sk_nulls_next(sk); +- } while (sk && (!net_eq(sock_net(sk), net) || sk->sk_family != state->family)); ++ } while (sk && (!net_eq(sock_net(sk), net) || ++ sk->sk_family != state->family || ++ !nx_check(sk->sk_nid, VS_WATCH_P | VS_IDENT))); + + if (!sk) { + if (state->bucket <= state->udp_table->mask) +diff -NurpP --minimal linux-3.0.9/net/ipv6/Kconfig linux-3.0.9-vs2.3.2.1/net/ipv6/Kconfig +--- linux-3.0.9/net/ipv6/Kconfig 2010-08-02 16:52:59.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/net/ipv6/Kconfig 2011-06-10 22:11:24.000000000 +0200 +@@ -4,8 +4,8 @@ + + # IPv6 as module will cause a CRASH if you try to unload it + menuconfig IPV6 +- tristate "The IPv6 protocol" +- default m ++ bool "The IPv6 protocol" ++ default n + ---help--- + This is complemental support for the IP version 6. + You will still be able to do traditional IPv4 networking as well. +diff -NurpP --minimal linux-3.0.9/net/ipv6/addrconf.c linux-3.0.9-vs2.3.2.1/net/ipv6/addrconf.c +--- linux-3.0.9/net/ipv6/addrconf.c 2011-07-22 11:18:13.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/net/ipv6/addrconf.c 2011-06-10 22:11:24.000000000 +0200 +@@ -87,6 +87,8 @@ + + #include <linux/proc_fs.h> + #include <linux/seq_file.h> ++#include <linux/vs_network.h> ++#include <linux/vs_inet6.h> + + /* Set to 3 to get tracing... */ + #define ACONF_DEBUG 2 +@@ -1108,7 +1110,7 @@ out: + + int ipv6_dev_get_saddr(struct net *net, struct net_device *dst_dev, + const struct in6_addr *daddr, unsigned int prefs, +- struct in6_addr *saddr) ++ struct in6_addr *saddr, struct nx_info *nxi) + { + struct ipv6_saddr_score scores[2], + *score = &scores[0], *hiscore = &scores[1]; +@@ -1180,6 +1182,8 @@ int ipv6_dev_get_saddr(struct net *net, + dev->name); + continue; + } ++ if (!v6_addr_in_nx_info(nxi, &score->ifa->addr, -1)) ++ continue; + + score->rule = -1; + bitmap_zero(score->scorebits, IPV6_SADDR_RULE_MAX); +@@ -3048,7 +3052,10 @@ static void if6_seq_stop(struct seq_file + static int if6_seq_show(struct seq_file *seq, void *v) + { + struct inet6_ifaddr *ifp = (struct inet6_ifaddr *)v; +- seq_printf(seq, "%pi6 %02x %02x %02x %02x %8s\n", ++ ++ if (nx_check(0, VS_ADMIN|VS_WATCH) || ++ v6_addr_in_nx_info(current_nx_info(), &ifp->addr, -1)) ++ seq_printf(seq, "%pi6 %02x %02x %02x %02x %8s\n", + &ifp->addr, + ifp->idev->dev->ifindex, + ifp->prefix_len, +@@ -3554,6 +3561,11 @@ static int in6_dump_addrs(struct inet6_d + struct ifacaddr6 *ifaca; + int err = 1; + int ip_idx = *p_ip_idx; ++ struct nx_info *nxi = skb->sk ? skb->sk->sk_nx_info : NULL; ++ ++ /* disable ipv6 on non v6 guests */ ++ if (nxi && !nx_info_has_v6(nxi)) ++ return skb->len; + + read_lock_bh(&idev->lock); + switch (type) { +@@ -3564,6 +3576,8 @@ static int in6_dump_addrs(struct inet6_d + list_for_each_entry(ifa, &idev->addr_list, if_list) { + if (++ip_idx < s_ip_idx) + continue; ++ if (!v6_addr_in_nx_info(nxi, &ifa->addr, -1)) ++ continue; + err = inet6_fill_ifaddr(skb, ifa, + NETLINK_CB(cb->skb).pid, + cb->nlh->nlmsg_seq, +@@ -3580,6 +3594,8 @@ static int in6_dump_addrs(struct inet6_d + ifmca = ifmca->next, ip_idx++) { + if (ip_idx < s_ip_idx) + continue; ++ if (!v6_addr_in_nx_info(nxi, &ifmca->mca_addr, -1)) ++ continue; + err = inet6_fill_ifmcaddr(skb, ifmca, + NETLINK_CB(cb->skb).pid, + cb->nlh->nlmsg_seq, +@@ -3595,6 +3611,8 @@ static int in6_dump_addrs(struct inet6_d + ifaca = ifaca->aca_next, ip_idx++) { + if (ip_idx < s_ip_idx) + continue; ++ if (!v6_addr_in_nx_info(nxi, &ifaca->aca_addr, -1)) ++ continue; + err = inet6_fill_ifacaddr(skb, ifaca, + NETLINK_CB(cb->skb).pid, + cb->nlh->nlmsg_seq, +@@ -3980,6 +3998,11 @@ static int inet6_dump_ifinfo(struct sk_b + struct inet6_dev *idev; + struct hlist_head *head; + struct hlist_node *node; ++ struct nx_info *nxi = skb->sk ? skb->sk->sk_nx_info : NULL; ++ ++ /* FIXME: maybe disable ipv6 on non v6 guests? ++ if (skb->sk && skb->sk->sk_vx_info) ++ return skb->len; */ + + s_h = cb->args[0]; + s_idx = cb->args[1]; +@@ -3991,6 +4014,8 @@ static int inet6_dump_ifinfo(struct sk_b + hlist_for_each_entry_rcu(dev, node, head, index_hlist) { + if (idx < s_idx) + goto cont; ++ if (!v6_dev_in_nx_info(dev, nxi)) ++ goto cont; + idev = __in6_dev_get(dev); + if (!idev) + goto cont; +diff -NurpP --minimal linux-3.0.9/net/ipv6/af_inet6.c linux-3.0.9-vs2.3.2.1/net/ipv6/af_inet6.c +--- linux-3.0.9/net/ipv6/af_inet6.c 2011-11-15 16:40:47.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/net/ipv6/af_inet6.c 2011-08-29 03:45:10.000000000 +0200 +@@ -42,6 +42,8 @@ + #include <linux/netdevice.h> + #include <linux/icmpv6.h> + #include <linux/netfilter_ipv6.h> ++#include <linux/vs_inet.h> ++#include <linux/vs_inet6.h> + + #include <net/ip.h> + #include <net/ipv6.h> +@@ -160,9 +162,12 @@ lookup_protocol: + } + + err = -EPERM; ++ if ((protocol == IPPROTO_ICMPV6) && ++ nx_capable(CAP_NET_RAW, NXC_RAW_ICMP)) ++ goto override; + if (sock->type == SOCK_RAW && !kern && !capable(CAP_NET_RAW)) + goto out_rcu_unlock; +- ++override: + sock->ops = answer->ops; + answer_prot = answer->prot; + answer_no_check = answer->no_check; +@@ -261,6 +266,7 @@ int inet6_bind(struct socket *sock, stru + struct inet_sock *inet = inet_sk(sk); + struct ipv6_pinfo *np = inet6_sk(sk); + struct net *net = sock_net(sk); ++ struct nx_v6_sock_addr nsa; + __be32 v4addr = 0; + unsigned short snum; + int addr_type = 0; +@@ -276,6 +282,10 @@ int inet6_bind(struct socket *sock, stru + if (addr->sin6_family != AF_INET6) + return -EAFNOSUPPORT; + ++ err = v6_map_sock_addr(inet, addr, &nsa); ++ if (err) ++ return err; ++ + addr_type = ipv6_addr_type(&addr->sin6_addr); + if ((addr_type & IPV6_ADDR_MULTICAST) && sock->type == SOCK_STREAM) + return -EINVAL; +@@ -307,6 +317,7 @@ int inet6_bind(struct socket *sock, stru + /* Reproduce AF_INET checks to make the bindings consistent */ + v4addr = addr->sin6_addr.s6_addr32[3]; + chk_addr_ret = inet_addr_type(net, v4addr); ++ + if (!sysctl_ip_nonlocal_bind && + !(inet->freebind || inet->transparent) && + v4addr != htonl(INADDR_ANY) && +@@ -316,6 +327,10 @@ int inet6_bind(struct socket *sock, stru + err = -EADDRNOTAVAIL; + goto out; + } ++ if (!v4_addr_in_nx_info(sk->sk_nx_info, v4addr, NXA_MASK_BIND)) { ++ err = -EADDRNOTAVAIL; ++ goto out; ++ } + } else { + if (addr_type != IPV6_ADDR_ANY) { + struct net_device *dev = NULL; +@@ -342,6 +357,11 @@ int inet6_bind(struct socket *sock, stru + } + } + ++ if (!v6_addr_in_nx_info(sk->sk_nx_info, &addr->sin6_addr, -1)) { ++ err = -EADDRNOTAVAIL; ++ goto out; ++ } ++ + /* ipv4 addr of the socket is invalid. Only the + * unspecified and mapped address have a v4 equivalent. + */ +@@ -358,6 +378,9 @@ int inet6_bind(struct socket *sock, stru + } + } + ++ /* what's that for? */ ++ v6_set_sock_addr(inet, &nsa); ++ + inet->inet_rcv_saddr = v4addr; + inet->inet_saddr = v4addr; + +@@ -459,9 +482,11 @@ int inet6_getname(struct socket *sock, s + return -ENOTCONN; + sin->sin6_port = inet->inet_dport; + ipv6_addr_copy(&sin->sin6_addr, &np->daddr); ++ /* FIXME: remap lback? */ + if (np->sndflow) + sin->sin6_flowinfo = np->flow_label; + } else { ++ /* FIXME: remap lback? */ + if (ipv6_addr_any(&np->rcv_saddr)) + ipv6_addr_copy(&sin->sin6_addr, &np->saddr); + else +diff -NurpP --minimal linux-3.0.9/net/ipv6/datagram.c linux-3.0.9-vs2.3.2.1/net/ipv6/datagram.c +--- linux-3.0.9/net/ipv6/datagram.c 2011-05-22 16:18:00.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/net/ipv6/datagram.c 2011-09-17 19:24:15.000000000 +0200 +@@ -639,7 +639,7 @@ int datagram_send_ctl(struct net *net, + + rcu_read_lock(); + if (fl6->flowi6_oif) { +- dev = dev_get_by_index_rcu(net, fl6->flowi6_oif); ++ dev = dev_get_by_index_real_rcu(net, fl6->flowi6_oif); + if (!dev) { + rcu_read_unlock(); + return -ENODEV; +diff -NurpP --minimal linux-3.0.9/net/ipv6/fib6_rules.c linux-3.0.9-vs2.3.2.1/net/ipv6/fib6_rules.c +--- linux-3.0.9/net/ipv6/fib6_rules.c 2011-05-22 16:18:00.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/net/ipv6/fib6_rules.c 2011-06-10 22:11:24.000000000 +0200 +@@ -90,7 +90,7 @@ static int fib6_rule_action(struct fib_r + ip6_dst_idev(&rt->dst)->dev, + &flp6->daddr, + rt6_flags2srcprefs(flags), +- &saddr)) ++ &saddr, NULL)) + goto again; + if (!ipv6_prefix_equal(&saddr, &r->src.addr, + r->src.plen)) +diff -NurpP --minimal linux-3.0.9/net/ipv6/inet6_hashtables.c linux-3.0.9-vs2.3.2.1/net/ipv6/inet6_hashtables.c +--- linux-3.0.9/net/ipv6/inet6_hashtables.c 2011-11-15 16:40:47.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/net/ipv6/inet6_hashtables.c 2011-08-29 03:45:10.000000000 +0200 +@@ -16,6 +16,7 @@ + + #include <linux/module.h> + #include <linux/random.h> ++#include <linux/vs_inet6.h> + + #include <net/inet_connection_sock.h> + #include <net/inet_hashtables.h> +@@ -83,7 +84,6 @@ struct sock *__inet6_lookup_established( + unsigned int slot = hash & hashinfo->ehash_mask; + struct inet_ehash_bucket *head = &hashinfo->ehash[slot]; + +- + rcu_read_lock(); + begin: + sk_nulls_for_each_rcu(sk, node, &head->chain) { +@@ -95,7 +95,7 @@ begin: + sock_put(sk); + goto begin; + } +- goto out; ++ goto out; + } + } + if (get_nulls_value(node) != slot) +@@ -141,6 +141,9 @@ static inline int compute_score(struct s + if (!ipv6_addr_equal(&np->rcv_saddr, daddr)) + return -1; + score++; ++ } else { ++ if (!v6_addr_in_nx_info(sk->sk_nx_info, daddr, -1)) ++ return -1; + } + if (sk->sk_bound_dev_if) { + if (sk->sk_bound_dev_if != dif) +diff -NurpP --minimal linux-3.0.9/net/ipv6/ip6_output.c linux-3.0.9-vs2.3.2.1/net/ipv6/ip6_output.c +--- linux-3.0.9/net/ipv6/ip6_output.c 2011-11-15 16:40:47.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/net/ipv6/ip6_output.c 2011-10-18 13:51:13.000000000 +0200 +@@ -962,7 +962,8 @@ static int ip6_dst_lookup_tail(struct so + struct rt6_info *rt = (struct rt6_info *) *dst; + err = ip6_route_get_saddr(net, rt, &fl6->daddr, + sk ? inet6_sk(sk)->srcprefs : 0, +- &fl6->saddr); ++ &fl6->saddr, ++ sk ? sk->sk_nx_info : NULL); + if (err) + goto out_err_release; + } +diff -NurpP --minimal linux-3.0.9/net/ipv6/ndisc.c linux-3.0.9-vs2.3.2.1/net/ipv6/ndisc.c +--- linux-3.0.9/net/ipv6/ndisc.c 2011-07-22 11:18:13.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/net/ipv6/ndisc.c 2011-06-10 22:11:24.000000000 +0200 +@@ -597,7 +597,7 @@ static void ndisc_send_na(struct net_dev + } else { + if (ipv6_dev_get_saddr(dev_net(dev), dev, daddr, + inet6_sk(dev_net(dev)->ipv6.ndisc_sk)->srcprefs, +- &tmpaddr)) ++ &tmpaddr, NULL)) + return; + src_addr = &tmpaddr; + } +diff -NurpP --minimal linux-3.0.9/net/ipv6/raw.c linux-3.0.9-vs2.3.2.1/net/ipv6/raw.c +--- linux-3.0.9/net/ipv6/raw.c 2011-07-22 11:18:13.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/net/ipv6/raw.c 2011-06-10 22:11:24.000000000 +0200 +@@ -30,6 +30,7 @@ + #include <linux/icmpv6.h> + #include <linux/netfilter.h> + #include <linux/netfilter_ipv6.h> ++#include <linux/vs_inet6.h> + #include <linux/skbuff.h> + #include <linux/compat.h> + #include <asm/uaccess.h> +@@ -284,6 +285,13 @@ static int rawv6_bind(struct sock *sk, s + goto out_unlock; + } + ++ if (!v6_addr_in_nx_info(sk->sk_nx_info, &addr->sin6_addr, -1)) { ++ err = -EADDRNOTAVAIL; ++ if (dev) ++ dev_put(dev); ++ goto out; ++ } ++ + /* ipv4 addr of the socket is invalid. Only the + * unspecified and mapped address have a v4 equivalent. + */ +diff -NurpP --minimal linux-3.0.9/net/ipv6/route.c linux-3.0.9-vs2.3.2.1/net/ipv6/route.c +--- linux-3.0.9/net/ipv6/route.c 2011-07-22 11:18:13.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/net/ipv6/route.c 2011-08-29 05:05:08.000000000 +0200 +@@ -54,6 +54,7 @@ + #include <net/xfrm.h> + #include <net/netevent.h> + #include <net/netlink.h> ++#include <linux/vs_inet6.h> + + #include <asm/uaccess.h> + +@@ -2050,15 +2051,17 @@ int ip6_route_get_saddr(struct net *net, + struct rt6_info *rt, + const struct in6_addr *daddr, + unsigned int prefs, +- struct in6_addr *saddr) ++ struct in6_addr *saddr, ++ struct nx_info *nxi) + { + struct inet6_dev *idev = ip6_dst_idev((struct dst_entry*)rt); + int err = 0; +- if (rt->rt6i_prefsrc.plen) ++ if (rt->rt6i_prefsrc.plen && (!nxi || ++ v6_addr_in_nx_info(nxi, &rt->rt6i_prefsrc.addr, NXA_TYPE_ADDR))) + ipv6_addr_copy(saddr, &rt->rt6i_prefsrc.addr); + else + err = ipv6_dev_get_saddr(net, idev ? idev->dev : NULL, +- daddr, prefs, saddr); ++ daddr, prefs, saddr, nxi); + return err; + } + +@@ -2387,7 +2390,8 @@ static int rt6_fill_node(struct net *net + NLA_PUT_U32(skb, RTA_IIF, iif); + } else if (dst) { + struct in6_addr saddr_buf; +- if (ip6_route_get_saddr(net, rt, dst, 0, &saddr_buf) == 0) ++ if (ip6_route_get_saddr(net, rt, dst, 0, &saddr_buf, ++ (skb->sk ? skb->sk->sk_nx_info : NULL)) == 0) + NLA_PUT(skb, RTA_PREFSRC, 16, &saddr_buf); + } + +@@ -2586,6 +2590,7 @@ static int rt6_info_route(struct rt6_inf + { + struct seq_file *m = p_arg; + ++ /* FIXME: check for network context? */ + seq_printf(m, "%pi6 %02x ", &rt->rt6i_dst.addr, rt->rt6i_dst.plen); + + #ifdef CONFIG_IPV6_SUBTREES +diff -NurpP --minimal linux-3.0.9/net/ipv6/tcp_ipv6.c linux-3.0.9-vs2.3.2.1/net/ipv6/tcp_ipv6.c +--- linux-3.0.9/net/ipv6/tcp_ipv6.c 2011-11-15 16:40:47.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/net/ipv6/tcp_ipv6.c 2011-11-15 17:37:07.000000000 +0100 +@@ -70,6 +70,7 @@ + + #include <linux/crypto.h> + #include <linux/scatterlist.h> ++#include <linux/vs_inet6.h> + + static void tcp_v6_send_reset(struct sock *sk, struct sk_buff *skb); + static void tcp_v6_reqsk_send_ack(struct sock *sk, struct sk_buff *skb, +@@ -162,8 +163,15 @@ static int tcp_v6_connect(struct sock *s + * connect() to INADDR_ANY means loopback (BSD'ism). + */ + +- if(ipv6_addr_any(&usin->sin6_addr)) +- usin->sin6_addr.s6_addr[15] = 0x1; ++ if(ipv6_addr_any(&usin->sin6_addr)) { ++ struct nx_info *nxi = sk->sk_nx_info; ++ ++ if (nxi && nx_info_has_v6(nxi)) ++ /* FIXME: remap lback? */ ++ usin->sin6_addr = nxi->v6.ip; ++ else ++ usin->sin6_addr.s6_addr[15] = 0x1; ++ } + + addr_type = ipv6_addr_type(&usin->sin6_addr); + +diff -NurpP --minimal linux-3.0.9/net/ipv6/udp.c linux-3.0.9-vs2.3.2.1/net/ipv6/udp.c +--- linux-3.0.9/net/ipv6/udp.c 2011-11-15 16:40:47.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/net/ipv6/udp.c 2011-10-18 13:51:13.000000000 +0200 +@@ -45,41 +45,67 @@ + #include <net/tcp_states.h> + #include <net/ip6_checksum.h> + #include <net/xfrm.h> ++#include <linux/vs_inet6.h> + + #include <linux/proc_fs.h> + #include <linux/seq_file.h> + #include "udp_impl.h" + +-int ipv6_rcv_saddr_equal(const struct sock *sk, const struct sock *sk2) ++int ipv6_rcv_saddr_equal(const struct sock *sk1, const struct sock *sk2) + { +- const struct in6_addr *sk_rcv_saddr6 = &inet6_sk(sk)->rcv_saddr; ++ const struct in6_addr *sk1_rcv_saddr6 = &inet6_sk(sk1)->rcv_saddr; + const struct in6_addr *sk2_rcv_saddr6 = inet6_rcv_saddr(sk2); +- __be32 sk1_rcv_saddr = sk_rcv_saddr(sk); ++ __be32 sk1_rcv_saddr = sk_rcv_saddr(sk1); + __be32 sk2_rcv_saddr = sk_rcv_saddr(sk2); +- int sk_ipv6only = ipv6_only_sock(sk); ++ int sk1_ipv6only = ipv6_only_sock(sk1); + int sk2_ipv6only = inet_v6_ipv6only(sk2); +- int addr_type = ipv6_addr_type(sk_rcv_saddr6); ++ int addr_type = ipv6_addr_type(sk1_rcv_saddr6); + int addr_type2 = sk2_rcv_saddr6 ? ipv6_addr_type(sk2_rcv_saddr6) : IPV6_ADDR_MAPPED; + + /* if both are mapped, treat as IPv4 */ +- if (addr_type == IPV6_ADDR_MAPPED && addr_type2 == IPV6_ADDR_MAPPED) +- return (!sk2_ipv6only && ++ if (addr_type == IPV6_ADDR_MAPPED && addr_type2 == IPV6_ADDR_MAPPED) { ++ if (!sk2_ipv6only && + (!sk1_rcv_saddr || !sk2_rcv_saddr || +- sk1_rcv_saddr == sk2_rcv_saddr)); ++ sk1_rcv_saddr == sk2_rcv_saddr)) ++ goto vs_v4; ++ else ++ return 0; ++ } + + if (addr_type2 == IPV6_ADDR_ANY && + !(sk2_ipv6only && addr_type == IPV6_ADDR_MAPPED)) +- return 1; ++ goto vs; + + if (addr_type == IPV6_ADDR_ANY && +- !(sk_ipv6only && addr_type2 == IPV6_ADDR_MAPPED)) +- return 1; ++ !(sk1_ipv6only && addr_type2 == IPV6_ADDR_MAPPED)) ++ goto vs; + + if (sk2_rcv_saddr6 && +- ipv6_addr_equal(sk_rcv_saddr6, sk2_rcv_saddr6)) +- return 1; ++ ipv6_addr_equal(sk1_rcv_saddr6, sk2_rcv_saddr6)) ++ goto vs; + + return 0; ++ ++vs_v4: ++ if (!sk1_rcv_saddr && !sk2_rcv_saddr) ++ return nx_v4_addr_conflict(sk1->sk_nx_info, sk2->sk_nx_info); ++ if (!sk2_rcv_saddr) ++ return v4_addr_in_nx_info(sk1->sk_nx_info, sk2_rcv_saddr, -1); ++ if (!sk1_rcv_saddr) ++ return v4_addr_in_nx_info(sk2->sk_nx_info, sk1_rcv_saddr, -1); ++ return 1; ++vs: ++ if (addr_type2 == IPV6_ADDR_ANY && addr_type == IPV6_ADDR_ANY) ++ return nx_v6_addr_conflict(sk1->sk_nx_info, sk2->sk_nx_info); ++ else if (addr_type2 == IPV6_ADDR_ANY) ++ return v6_addr_in_nx_info(sk2->sk_nx_info, sk1_rcv_saddr6, -1); ++ else if (addr_type == IPV6_ADDR_ANY) { ++ if (addr_type2 == IPV6_ADDR_MAPPED) ++ return nx_v4_addr_conflict(sk1->sk_nx_info, sk2->sk_nx_info); ++ else ++ return v6_addr_in_nx_info(sk1->sk_nx_info, sk2_rcv_saddr6, -1); ++ } ++ return 1; + } + + static unsigned int udp6_portaddr_hash(struct net *net, +@@ -143,6 +169,10 @@ static inline int compute_score(struct s + if (!ipv6_addr_equal(&np->rcv_saddr, daddr)) + return -1; + score++; ++ } else { ++ /* block non nx_info ips */ ++ if (!v6_addr_in_nx_info(sk->sk_nx_info, daddr, -1)) ++ return -1; + } + if (!ipv6_addr_any(&np->daddr)) { + if (!ipv6_addr_equal(&np->daddr, saddr)) +diff -NurpP --minimal linux-3.0.9/net/ipv6/xfrm6_policy.c linux-3.0.9-vs2.3.2.1/net/ipv6/xfrm6_policy.c +--- linux-3.0.9/net/ipv6/xfrm6_policy.c 2011-07-22 11:18:13.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/net/ipv6/xfrm6_policy.c 2011-06-10 22:11:24.000000000 +0200 +@@ -63,7 +63,7 @@ static int xfrm6_get_saddr(struct net *n + dev = ip6_dst_idev(dst)->dev; + ipv6_dev_get_saddr(dev_net(dev), dev, + (struct in6_addr *)&daddr->a6, 0, +- (struct in6_addr *)&saddr->a6); ++ (struct in6_addr *)&saddr->a6, NULL); + dst_release(dst); + return 0; + } +diff -NurpP --minimal linux-3.0.9/net/netfilter/ipvs/ip_vs_xmit.c linux-3.0.9-vs2.3.2.1/net/netfilter/ipvs/ip_vs_xmit.c +--- linux-3.0.9/net/netfilter/ipvs/ip_vs_xmit.c 2011-07-22 11:18:13.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/net/netfilter/ipvs/ip_vs_xmit.c 2011-07-22 15:27:52.000000000 +0200 +@@ -226,7 +226,7 @@ __ip_vs_route_output_v6(struct net *net, + return dst; + if (ipv6_addr_any(&fl6.saddr) && + ipv6_dev_get_saddr(net, ip6_dst_idev(dst)->dev, +- &fl6.daddr, 0, &fl6.saddr) < 0) ++ &fl6.daddr, 0, &fl6.saddr, NULL) < 0) + goto out_err; + if (do_xfrm) { + dst = xfrm_lookup(net, dst, flowi6_to_flowi(&fl6), NULL, 0); +diff -NurpP --minimal linux-3.0.9/net/netlink/af_netlink.c linux-3.0.9-vs2.3.2.1/net/netlink/af_netlink.c +--- linux-3.0.9/net/netlink/af_netlink.c 2011-07-22 11:18:13.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/net/netlink/af_netlink.c 2011-06-10 22:11:24.000000000 +0200 +@@ -55,6 +55,9 @@ + #include <linux/types.h> + #include <linux/audit.h> + #include <linux/mutex.h> ++#include <linux/vs_context.h> ++#include <linux/vs_network.h> ++#include <linux/vs_limit.h> + + #include <net/net_namespace.h> + #include <net/sock.h> +@@ -1907,6 +1910,8 @@ static struct sock *netlink_seq_socket_i + sk_for_each(s, node, &hash->table[j]) { + if (sock_net(s) != seq_file_net(seq)) + continue; ++ if (!nx_check(s->sk_nid, VS_WATCH_P | VS_IDENT)) ++ continue; + if (off == pos) { + iter->link = i; + iter->hash_idx = j; +@@ -1941,7 +1946,8 @@ static void *netlink_seq_next(struct seq + s = v; + do { + s = sk_next(s); +- } while (s && sock_net(s) != seq_file_net(seq)); ++ } while (s && (sock_net(s) != seq_file_net(seq) || ++ !nx_check(s->sk_nid, VS_WATCH_P | VS_IDENT))); + if (s) + return s; + +@@ -1953,7 +1959,8 @@ static void *netlink_seq_next(struct seq + + for (; j <= hash->mask; j++) { + s = sk_head(&hash->table[j]); +- while (s && sock_net(s) != seq_file_net(seq)) ++ while (s && (sock_net(s) != seq_file_net(seq) || ++ !nx_check(s->sk_nid, VS_WATCH_P | VS_IDENT))) + s = sk_next(s); + if (s) { + iter->link = i; +diff -NurpP --minimal linux-3.0.9/net/socket.c linux-3.0.9-vs2.3.2.1/net/socket.c +--- linux-3.0.9/net/socket.c 2011-11-15 16:40:48.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/net/socket.c 2011-10-18 13:51:13.000000000 +0200 +@@ -98,6 +98,10 @@ + + #include <net/sock.h> + #include <linux/netfilter.h> ++#include <linux/vs_base.h> ++#include <linux/vs_socket.h> ++#include <linux/vs_inet.h> ++#include <linux/vs_inet6.h> + + #include <linux/if_tun.h> + #include <linux/ipv6_route.h> +@@ -546,6 +550,7 @@ static inline int __sock_sendmsg_nosec(s + struct msghdr *msg, size_t size) + { + struct sock_iocb *si = kiocb_to_siocb(iocb); ++ size_t len; + + sock_update_classid(sock->sk); + +@@ -554,7 +559,22 @@ static inline int __sock_sendmsg_nosec(s + si->msg = msg; + si->size = size; + +- return sock->ops->sendmsg(iocb, sock, msg, size); ++ len = sock->ops->sendmsg(iocb, sock, msg, size); ++ if (sock->sk) { ++ if (len == size) ++ vx_sock_send(sock->sk, size); ++ else ++ vx_sock_fail(sock->sk, size); ++ } ++ vxdprintk(VXD_CBIT(net, 7), ++ "__sock_sendmsg: %p[%p,%p,%p;%d/%d]:%d/%zu", ++ sock, sock->sk, ++ (sock->sk)?sock->sk->sk_nx_info:0, ++ (sock->sk)?sock->sk->sk_vx_info:0, ++ (sock->sk)?sock->sk->sk_xid:0, ++ (sock->sk)?sock->sk->sk_nid:0, ++ (unsigned int)size, len); ++ return len; + } + + static inline int __sock_sendmsg(struct kiocb *iocb, struct socket *sock, +@@ -694,6 +714,7 @@ static inline int __sock_recvmsg_nosec(s + struct msghdr *msg, size_t size, int flags) + { + struct sock_iocb *si = kiocb_to_siocb(iocb); ++ int len; + + sock_update_classid(sock->sk); + +@@ -703,7 +724,18 @@ static inline int __sock_recvmsg_nosec(s + si->size = size; + si->flags = flags; + +- return sock->ops->recvmsg(iocb, sock, msg, size, flags); ++ len = sock->ops->recvmsg(iocb, sock, msg, size, flags); ++ if ((len >= 0) && sock->sk) ++ vx_sock_recv(sock->sk, len); ++ vxdprintk(VXD_CBIT(net, 7), ++ "__sock_recvmsg: %p[%p,%p,%p;%d/%d]:%d/%d", ++ sock, sock->sk, ++ (sock->sk)?sock->sk->sk_nx_info:0, ++ (sock->sk)?sock->sk->sk_vx_info:0, ++ (sock->sk)?sock->sk->sk_xid:0, ++ (sock->sk)?sock->sk->sk_nid:0, ++ (unsigned int)size, len); ++ return len; + } + + static inline int __sock_recvmsg(struct kiocb *iocb, struct socket *sock, +@@ -1188,6 +1220,13 @@ int __sock_create(struct net *net, int f + if (type < 0 || type >= SOCK_MAX) + return -EINVAL; + ++ if (!nx_check(0, VS_ADMIN)) { ++ if (family == PF_INET && !current_nx_info_has_v4()) ++ return -EAFNOSUPPORT; ++ if (family == PF_INET6 && !current_nx_info_has_v6()) ++ return -EAFNOSUPPORT; ++ } ++ + /* Compatibility. + + This uglymoron is moved from INET layer to here to avoid +@@ -1323,6 +1362,7 @@ SYSCALL_DEFINE3(socket, int, family, int + if (retval < 0) + goto out; + ++ set_bit(SOCK_USER_SOCKET, &sock->flags); + retval = sock_map_fd(sock, flags & (O_CLOEXEC | O_NONBLOCK)); + if (retval < 0) + goto out_release; +@@ -1364,10 +1404,12 @@ SYSCALL_DEFINE4(socketpair, int, family, + err = sock_create(family, type, protocol, &sock1); + if (err < 0) + goto out; ++ set_bit(SOCK_USER_SOCKET, &sock1->flags); + + err = sock_create(family, type, protocol, &sock2); + if (err < 0) + goto out_release_1; ++ set_bit(SOCK_USER_SOCKET, &sock2->flags); + + err = sock1->ops->socketpair(sock1, sock2); + if (err < 0) +diff -NurpP --minimal linux-3.0.9/net/sunrpc/auth.c linux-3.0.9-vs2.3.2.1/net/sunrpc/auth.c +--- linux-3.0.9/net/sunrpc/auth.c 2011-07-22 11:18:13.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/net/sunrpc/auth.c 2011-06-10 22:11:24.000000000 +0200 +@@ -14,6 +14,7 @@ + #include <linux/hash.h> + #include <linux/sunrpc/clnt.h> + #include <linux/spinlock.h> ++#include <linux/vs_tag.h> + + #ifdef RPC_DEBUG + # define RPCDBG_FACILITY RPCDBG_AUTH +@@ -427,6 +428,7 @@ rpcauth_lookupcred(struct rpc_auth *auth + memset(&acred, 0, sizeof(acred)); + acred.uid = cred->fsuid; + acred.gid = cred->fsgid; ++ acred.tag = dx_current_tag(); + acred.group_info = get_group_info(((struct cred *)cred)->group_info); + + ret = auth->au_ops->lookup_cred(auth, &acred, flags); +@@ -467,6 +469,7 @@ rpcauth_bind_root_cred(struct rpc_task * + struct auth_cred acred = { + .uid = 0, + .gid = 0, ++ .tag = dx_current_tag(), + }; + + dprintk("RPC: %5u looking up %s cred\n", +diff -NurpP --minimal linux-3.0.9/net/sunrpc/auth_unix.c linux-3.0.9-vs2.3.2.1/net/sunrpc/auth_unix.c +--- linux-3.0.9/net/sunrpc/auth_unix.c 2011-11-15 16:40:48.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/net/sunrpc/auth_unix.c 2011-11-15 17:37:07.000000000 +0100 +@@ -12,12 +12,14 @@ + #include <linux/module.h> + #include <linux/sunrpc/clnt.h> + #include <linux/sunrpc/auth.h> ++#include <linux/vs_tag.h> + + #define NFS_NGROUPS 16 + + struct unx_cred { + struct rpc_cred uc_base; + gid_t uc_gid; ++ tag_t uc_tag; + gid_t uc_gids[NFS_NGROUPS]; + }; + #define uc_uid uc_base.cr_uid +@@ -78,6 +80,7 @@ unx_create_cred(struct rpc_auth *auth, s + groups = NFS_NGROUPS; + + cred->uc_gid = acred->gid; ++ cred->uc_tag = acred->tag; + for (i = 0; i < groups; i++) + cred->uc_gids[i] = GROUP_AT(acred->group_info, i); + if (i < NFS_NGROUPS) +@@ -119,7 +122,9 @@ unx_match(struct auth_cred *acred, struc + unsigned int i; + + +- if (cred->uc_uid != acred->uid || cred->uc_gid != acred->gid) ++ if (cred->uc_uid != acred->uid || ++ cred->uc_gid != acred->gid || ++ cred->uc_tag != acred->tag) + return 0; + + if (acred->group_info != NULL) +@@ -145,7 +150,7 @@ unx_marshal(struct rpc_task *task, __be3 + struct rpc_clnt *clnt = task->tk_client; + struct unx_cred *cred = container_of(task->tk_rqstp->rq_cred, struct unx_cred, uc_base); + __be32 *base, *hold; +- int i; ++ int i, tag; + + *p++ = htonl(RPC_AUTH_UNIX); + base = p++; +@@ -155,9 +160,12 @@ unx_marshal(struct rpc_task *task, __be3 + * Copy the UTS nodename captured when the client was created. + */ + p = xdr_encode_array(p, clnt->cl_nodename, clnt->cl_nodelen); ++ tag = task->tk_client->cl_tag; + +- *p++ = htonl((u32) cred->uc_uid); +- *p++ = htonl((u32) cred->uc_gid); ++ *p++ = htonl((u32) TAGINO_UID(tag, ++ cred->uc_uid, cred->uc_tag)); ++ *p++ = htonl((u32) TAGINO_GID(tag, ++ cred->uc_gid, cred->uc_tag)); + hold = p++; + for (i = 0; i < 16 && cred->uc_gids[i] != (gid_t) NOGROUP; i++) + *p++ = htonl((u32) cred->uc_gids[i]); +diff -NurpP --minimal linux-3.0.9/net/sunrpc/clnt.c linux-3.0.9-vs2.3.2.1/net/sunrpc/clnt.c +--- linux-3.0.9/net/sunrpc/clnt.c 2011-07-22 11:18:13.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/net/sunrpc/clnt.c 2011-07-01 11:35:35.000000000 +0200 +@@ -31,6 +31,7 @@ + #include <linux/in.h> + #include <linux/in6.h> + #include <linux/un.h> ++#include <linux/vs_cvirt.h> + + #include <linux/sunrpc/clnt.h> + #include <linux/sunrpc/rpc_pipe_fs.h> +@@ -362,6 +363,9 @@ struct rpc_clnt *rpc_create(struct rpc_c + if (!(args->flags & RPC_CLNT_CREATE_QUIET)) + clnt->cl_chatty = 1; + ++ /* TODO: handle RPC_CLNT_CREATE_TAGGED ++ if (args->flags & RPC_CLNT_CREATE_TAGGED) ++ clnt->cl_tag = 1; */ + return clnt; + } + EXPORT_SYMBOL_GPL(rpc_create); +diff -NurpP --minimal linux-3.0.9/net/unix/af_unix.c linux-3.0.9-vs2.3.2.1/net/unix/af_unix.c +--- linux-3.0.9/net/unix/af_unix.c 2011-07-22 11:18:13.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/net/unix/af_unix.c 2011-06-10 22:11:24.000000000 +0200 +@@ -114,6 +114,8 @@ + #include <linux/mount.h> + #include <net/checksum.h> + #include <linux/security.h> ++#include <linux/vs_context.h> ++#include <linux/vs_limit.h> + + static struct hlist_head unix_socket_table[UNIX_HASH_SIZE + 1]; + static DEFINE_SPINLOCK(unix_table_lock); +@@ -258,6 +260,8 @@ static struct sock *__unix_find_socket_b + if (!net_eq(sock_net(s), net)) + continue; + ++ if (!nx_check(s->sk_nid, VS_WATCH_P | VS_IDENT)) ++ continue; + if (u->addr->len == len && + !memcmp(u->addr->name, sunname, len)) + goto found; +@@ -2208,6 +2212,8 @@ static struct sock *unix_seq_idx(struct + for (s = first_unix_socket(&iter->i); s; s = next_unix_socket(&iter->i, s)) { + if (sock_net(s) != seq_file_net(seq)) + continue; ++ if (!nx_check(s->sk_nid, VS_WATCH_P | VS_IDENT)) ++ continue; + if (off == pos) + return s; + ++off; +@@ -2232,7 +2238,8 @@ static void *unix_seq_next(struct seq_fi + sk = first_unix_socket(&iter->i); + else + sk = next_unix_socket(&iter->i, sk); +- while (sk && (sock_net(sk) != seq_file_net(seq))) ++ while (sk && (sock_net(sk) != seq_file_net(seq) || ++ !nx_check(sk->sk_nid, VS_WATCH_P | VS_IDENT))) + sk = next_unix_socket(&iter->i, sk); + return sk; + } +diff -NurpP --minimal linux-3.0.9/scripts/checksyscalls.sh linux-3.0.9-vs2.3.2.1/scripts/checksyscalls.sh +--- linux-3.0.9/scripts/checksyscalls.sh 2011-03-15 18:07:46.000000000 +0100 ++++ linux-3.0.9-vs2.3.2.1/scripts/checksyscalls.sh 2011-06-10 22:11:24.000000000 +0200 +@@ -193,7 +193,6 @@ cat << EOF + #define __IGNORE_afs_syscall + #define __IGNORE_getpmsg + #define __IGNORE_putpmsg +-#define __IGNORE_vserver + EOF + } + +diff -NurpP --minimal linux-3.0.9/security/commoncap.c linux-3.0.9-vs2.3.2.1/security/commoncap.c +--- linux-3.0.9/security/commoncap.c 2011-07-22 11:18:14.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/security/commoncap.c 2011-08-30 15:42:43.000000000 +0200 +@@ -62,6 +62,7 @@ int cap_netlink_recv(struct sk_buff *skb + return -EPERM; + return 0; + } ++ + EXPORT_SYMBOL(cap_netlink_recv); + + /** +@@ -83,14 +84,20 @@ EXPORT_SYMBOL(cap_netlink_recv); + int cap_capable(struct task_struct *tsk, const struct cred *cred, + struct user_namespace *targ_ns, int cap, int audit) + { ++ struct vx_info *vxi = tsk->vx_info; ++ + for (;;) { + /* The creator of the user namespace has all caps. */ + if (targ_ns != &init_user_ns && targ_ns->creator == cred->user) + return 0; + + /* Do we have the necessary capabilities? */ +- if (targ_ns == cred->user->user_ns) +- return cap_raised(cred->cap_effective, cap) ? 0 : -EPERM; ++ if (targ_ns == cred->user->user_ns) { ++ if (vx_info_flags(vxi, VXF_STATE_SETUP, 0) && ++ cap_raised(cred->cap_effective, cap)) ++ return 0; ++ return vx_cap_raised(vxi, cred->cap_effective, cap) ? 0 : -EPERM; ++ } + + /* Have we tried all of the parent namespaces? */ + if (targ_ns == &init_user_ns) +@@ -611,7 +618,7 @@ int cap_inode_setxattr(struct dentry *de + + if (!strncmp(name, XATTR_SECURITY_PREFIX, + sizeof(XATTR_SECURITY_PREFIX) - 1) && +- !capable(CAP_SYS_ADMIN)) ++ !vx_capable(CAP_SYS_ADMIN, VXC_FS_SECURITY)) + return -EPERM; + return 0; + } +@@ -637,7 +644,7 @@ int cap_inode_removexattr(struct dentry + + if (!strncmp(name, XATTR_SECURITY_PREFIX, + sizeof(XATTR_SECURITY_PREFIX) - 1) && +- !capable(CAP_SYS_ADMIN)) ++ !vx_capable(CAP_SYS_ADMIN, VXC_FS_SECURITY)) + return -EPERM; + return 0; + } +diff -NurpP --minimal linux-3.0.9/security/selinux/hooks.c linux-3.0.9-vs2.3.2.1/security/selinux/hooks.c +--- linux-3.0.9/security/selinux/hooks.c 2011-07-22 11:18:14.000000000 +0200 ++++ linux-3.0.9-vs2.3.2.1/security/selinux/hooks.c 2011-06-15 02:40:14.000000000 +0200 +@@ -67,7 +67,6 @@ + #include <linux/dccp.h> + #include <linux/quota.h> + #include <linux/un.h> /* for Unix socket types */ +-#include <net/af_unix.h> /* for Unix socket types */ + #include <linux/parser.h> + #include <linux/nfs_mount.h> + #include <net/ipv6.h> |